Genetic Structural NAS: A Neural Network Architecture Search with Flexible Slot Connections Jakub Sadel KP Labs Gliwice, Poland jsadel@kplabs.pl Michal Kawulok Silesian University of Technology Gliwice, Poland michal.kawulok@polsl.pl Mateusz Przeliorz KP Labs Gliwice, Poland mprzeliorz@kplabs.pl Jakub Nalepa Silesian University of Technology Gliwice, Poland jakub.nalepa@polsl.pl Daniel Kostrzewa Silesian University of Technology Gliwice, Poland daniel.kostrzewa@polsl.pl ABSTRACT Selecting an appropriate neural network architecture and hyper- parameters to optimize performance for a given application is a difcult task. To overcome this challenge, Neural Architecture Search (NAS) and Hyperparameter Optimization (HPO) have been introduced. However, these techniques are often computationally- expensive and require signifcant amounts of execution time. To address this issue, we propose a semi-automated NAS approach that optimizes pre-existing architectural structures using genetic algo- rithms and eliminates unsuccessful combinations through shallow network training. The efectiveness of our technique was verifed through an experiment that produces a family of AlexNet-like neu- ral networks comprising 1296 models for image classifcation tasks. The computational study was conducted with fve runs of a genetic algorithm, resulting in the deep models with mean loss of 0.7044 and accuracy of 0.7334, both with low standard deviations, outper- forming the original AlexNet model with a signifcant margin. CCS CONCEPTS · Computing methodologies Neural networks; Genetic algorithmsComputer systems organization Neural net- works. KEYWORDS neural networks, genetic algorithm, neural architecture search ACM Reference Format: Jakub Sadel, Michal Kawulok, Mateusz Przeliorz, Jakub Nalepa, and Daniel Kostrzewa. 2023. Genetic Structural NAS: A Neural Network Architecture Search with Flexible Slot Connections. In Genetic and Evolutionary Com- putation Conference Companion (GECCO ’23 Companion), July 15ś19, 2023, Lisbon, Portugal. ACM, New York, NY, USA, 2 pages. https://doi.org/10.1145/ 3583133.3596435 Corresponding author. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for proft or commercial advantage and that copies bear this notice and the full citation on the frst page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s). GECCO ’23 Companion, July 15ś19, 2023, Lisbon, Portugal © 2023 Copyright held by the owner/author(s). ACM ISBN 979-8-4007-0120-7/23/07. https://doi.org/10.1145/3583133.3596435 1 INTRODUCTION Deep neural networks have been widely recognized for their re- markable success in various felds, particularly in computer vision and natural language processing, as they have become state-of- the-art solutions for tackling complex tasks. The ability of neural networks to learn intricate features and patterns from large datasets has led to signifcant breakthroughs in many domains. Despite their practical utility, selecting an appropriate network architecture and associated hyperparameters that can optimize their performance for a given task remains an open challenge. This is where Neural Architecture Search (NAS) [4] and Hyperparameter Optimization (HPO) [6] come into play by providing automated techniques for identifying the best-performing network architec- ture and hyperparameters for a task at hand. NAS automates the design process of a network’s architecture, while HPO fne-tunes the hyperparameters of a given deep network architecture to boost its performance. While these operations are currently being carried out in an automated manner in the majority of cases, due to large search spaces, they require signifcant amounts of time and com- putational power. To overcome this challenge, various techniques are employed, including shallow training, exploiting pre-trained modules, as well as utilizing (meta)heuristics for faster convergence. In this paper, we introduce a semi-automated NAS approach that can enhance the performance of networks deep within a short time frame. Specifcally, our method optimizes pre-existing structures using genetic algorithms and eliminates unsuccessful combinations through shallow network training, providing a practical and ef- cient solution for boosting the deep network’s performance. 2 THE METHOD The approach proposed in this work is based on creating a main network structure consisting of multiple slots that can be fexibly connected to one another. A similar partitioning can already be observed in the feld of object detection, where feature extraction and classifcation components were identifed as part of the model- building process. However, in our solution, there is a great deal of fexibility aforded in the slot connections. In contrast to the approach introduced in [5], a single slot in this methodology can, but does not necessarily have to be, a complex structure. Instead, it can equally be a single layer within the network. In the context of optimizing a (deep) neural network, it is im- portant to consider three main components: the main network 79