Abstract

Designing a convolution neural network from scratch is one of the biggest challenges facing the creation of reproducible models. Despite feeding the model with an adequate amount of labeled data to mitigate fluctuations during training, the model still suffers from high variance in the final overall accuracy and loss among identical training runs. Many of the reasons behind this are the randomness in data shuffling and augmentation, the behavior of gradient decent function, and the non-determinism in convnet layers and floating-point computation of the GPU.The method used to address the aforementioned issues, specifically in the case of negative transfer learning, is divided into three steps: First, designing an efficient lightweight convnet architecture with respect to available resources. Second, mitigating oscillations during training. Third, after setting the random seed across the training, select the appropriate weight initialization. Our extensive experiments in the use case of binary slum localization and detection show that our method improves the reproducibility of our model from scratch with an accuracy of 98.88±1.15%, a loss of 0.03±0.05 for a confidence level of 99.73%. These results make this model a strong competitor to the pre-trained models using transfer learning.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.