Abstract
Hyperparameters and architecture greatly influence the performance of convolutional neural networks (CNNs); therefore, their optimization is important to obtain the desired results. One of the state-of-the-art methods to achieve this is the use of neuroevolution that utilizes a genetic algorithm (GA) to optimize a CNN. However, the GA is often trapped into a local optimum resulting in premature convergence. In this study, we propose an approach called the “diversity-guided genetic algorithm-convolutional neural network (DGGA-CNN)” that uses adaptive parameter control and random injection to facilitate the search process by exploration and exploitation while preserving the population diversity. The alternation between exploration and exploitation is guided by using an average pairwise Hamming distance. Moreover, the DGGA fully handles the architecture of the CNN by using a novel finite state machine (FSM) combined with three novel mutation mechanisms that are specifically created for architecture chromosomes. Tests conducted on suggestion mining and twitter airline datasets reveal that the DGGA-CNN performs well with valid architectures and a comparison with other methods demonstrates its capability and efficiency.
Highlights
Despite the noteworthy achievements of convolutional neural networks (CNNs) in recent years (e.g., [1]–[5]), the selection of its hyperparameters and architecture remains a challenging task
The source code of our approach is available on GitHub
(a) Suggestion mining believe that expanding the range is necessary, and it would be interesting if the algorithm could automatically determine the ranges based on the present situation
Summary
Despite the noteworthy achievements of convolutional neural networks (CNNs) in recent years (e.g., [1]–[5]), the selection of its hyperparameters and architecture remains a challenging task. Traditional approaches usually use uninformed searches such as grid search [6], [7] and random search [8]. These optimization methods are popular as they are simple and easy to use [8]. An alternative method, called neuroevolution, utilizes an evolutionary algorithm (EA) to optimize a neural network. GA is resource-friendly as it effectively finds better solutions faster than the other EAs [24], [25] We adopted this algorithm as the base of our approach in this study
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.