Abstract
Training a neural network is a process of finding a suitable set of weights that results in high performance on training and validation data. This chapter provides some of the most significant techniques that used in training network models, including backpropagation, weight initialization, and batch normalization. Then, we examine the effectiveness of each weight initialization technique and verify batch normalization by applying it to train the neural network model on the CIFAR-10 dataset. Finally, we summarize, compare, and visualize the performance of all the trained neural networks in Chapter 4 and this chapter on TensorBoard to see the impact of each technique in the convergence of the model during the training process.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.