Abstract
The rapid development of deep neural networks has highlighted the importance of research in this domain. Neural Architecture Search (NAS) has emerged as a pivotal technique for automating and optimizing neural network designs. However, due to the complex and evolving nature of this field, staying up to date with the latest research, trends, and best practices is challenging. This article addresses the need for practical considerations, best practices, and open frameworks to guide practitioners in NAS endeavors. It discusses key considerations, challenges, opportunities, and open problems, along with a compilation of best practices and open frameworks. Readers will gain a practical guide for developing, testing, or applying NAS techniques.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
About vertex AI neural architecture search. https://cloud.google.com/vertex-ai/docs/training/neural-architecture-search/overview. Accessed 21 May 2023
Benmeziane, H., Maghraoui, K.E., Ouarnoughi, H., Niar, S., Wistuba, M., Wang, N.: Hardware-aware neural architecture search: survey and taxonomy. In: International Joint Conference on Artificial Intelligence (2021)
Brock, A., Lim, T., Ritchie, J.M., Weston, N.: SMASH: one-shot model architecture search through hypernetworks (2017)
Cai, H., Zhu, L., Han, S.: ProxylessNAS: direct neural architecture search on target task and hardware (2019)
Cai, S., Li, L., Deng, J., Zhang, B., Zha, Z.J., Su, L., Huang, Q.: Rethinking graph neural architecture search from message-passing (2021)
Cha, S., Kim, T., Lee, H., Yun, S.Y.: SuperNet in neural architecture search: a taxonomic survey. ArXiv abs/2204.03916 (2022)
Chen, D., Chen, L., Shang, Z., Zhang, Y., Wen, B., Yang, C.: Scale-aware neural architecture search for multivariate time series forecasting (2021)
Chitnis, S., Hosseini, R., Xie, P.: Brain tumor classification based on neural architecture search. Sci. Rep. 12(1), 19206 (2022). https://doi.org/10.1038/s41598-022-22172-6
Chu, J., Yu, X., Yang, S., Qiu, J., Wang, Q.: Architecture entropy sampling-based evolutionary neural architecture search and its application in osteoporosis diagnosis. Complex Intell. Syst. 9(1), 213–231 (2023). https://doi.org/10.1007/s40747-022-00794-7
Elsken, T., Metzen, J.H., Hutter, F.: Neural architecture search: a survey (2019)
Elsken, T., Staffler, B., Zela, A., Metzen, J.H., Hutter, F.: Bag of tricks for neural architecture search (2021)
Gao, Y., Yang, H., Zhang, P., Zhou, C., Hu, Y.: GraphNAS: graph neural architecture search with reinforcement learning (2019)
Guan, C., Wang, X., Chen, H., Zhang, Z., Zhu, W.: Large-scale graph neural architecture search. In: Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., Sabato, S. (eds.) Proceedings of the 39th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 162, pp. 7968–7981. PMLR (2022). https://proceedings.mlr.press/v162/guan22d.html
Gupta, A., Sheth, P., Xie, P.: Neural architecture search for pneumonia diagnosis from chest X-rays. Sci. Rep. 12(1), 11309 (2022). https://doi.org/10.1038/s41598-022-15341-0
He, C., Ye, H., Shen, L., Zhang, T.: MileNAS: efficient neural architecture search via mixed-level reformulation (2020)
Hu, S., Xie, X., Liu, S., Geng, M., Liu, X., Meng, H.: Neural architecture search for speech recognition (2020)
Jiang, Y., Hu, C., Xiao, T., Zhang, C., Zhu, J.: Improved differentiable architecture search for language modeling and named entity recognition. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3585–3590. Association for Computational Linguistics, Hong Kong (2019). https://doi.org/10.18653/v1/D19-1367, https://aclanthology.org/D19-1367
Jin, H., Song, Q., Hu, X.: Auto-keras: an efficient neural architecture search system (2019)
Kim, Y., Yun, W.J., Lee, Y.K., Jung, S., Kim, J.: Trends in neural architecture search: Towards the acceleration of search (2021)
Klyuchnikov, N., Trofimov, I., Artemova, E., Salnikov, M., Fedorov, M., Burnaev, E.: NAS-bench-NLP: neural architecture search benchmark for natural language processing. IEEE Access 1 (2020)
Li, C., et al.: BossNAS: exploring hybrid CNN-transformers with block-wisely self-supervised neural architecture search (2021)
Li, L., Talwalkar, A.: Random search and reproducibility for neural architecture search (2019)
Li, Y., Hao, C., Li, P., Xiong, J., Chen, D.: Generic neural architecture search via regression (2021)
Lindauer, M., Hutter, F.: Best practices for scientific research on neural architecture search. J. Mach. Learn. Res. 21(1), 9820–9837 (2020)
Liu, H., Simonyan, K., Vinyals, O., Fernando, C., Kavukcuoglu, K.: Hierarchical representations for efficient architecture search (2018)
Liu, H., Simonyan, K., Yang, Y.: DARTS: differentiable architecture search (2019)
Luo, R., Tian, F., Qin, T., Chen, E., Liu, T.Y.: Neural architecture optimization. In: Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 31. Curran Associates, Inc. (2018)
Mehrotra, A., et al.: NAS-bench-ASR: reproducible neural architecture search for speech recognition. In: International Conference on Learning Representations (2021)
Microsoft: Neural network intelligence (2021). https://github.com/microsoft/nni. If you use this software, please cite it as above
Miikkulainen, R., et al.: Evolving deep neural networks (2017)
Moser, B.B., Raue, F., Hees, J., Dengel, A.: DartsReNet: exploring New RNN cells in ReNet architectures. In: Farkaš, I., Masulli, P., Wermter, S. (eds.) ICANN 2020. LNCS, vol. 12396, pp. 850–861. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-61609-0_67
Ning, X., et al.: Evaluating efficient performance estimators of neural architectures. In: Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems (2021). https://openreview.net/forum?id=Esd7tGH3Spl
Odema, M., Rashid, N., Faruque, M.A.A.: EExNAS: early-exit neural architecture search solutions for low-power wearable devices. In: 2021 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED), pp. 1–6 (2021)
Oyelade, O.N., Ezugwu, A.E.: A bioinspired neural architecture search based convolutional neural network for breast cancer detection using histopathology images. Sci. Rep. 11(1), 19940 (2021). https://doi.org/10.1038/s41598-021-98978-7
Pham, H., Guan, M.Y., Zoph, B., Le, Q.V., Dean, J.: Efficient neural architecture search via parameter sharing (2018)
Qian, G., et al.: When NAS meets trees: an efficient algorithm for neural architecture search (2022)
Rakhshani, H., et al.: Neural architecture search for time series classification. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8 (2020). https://doi.org/10.1109/IJCNN48605.2020.9206721
Ren, P., et al.: A comprehensive survey of neural architecture search: Challenges and solutions (2021)
Robles, J.G., Vanschoren, J.: Learning to reinforcement learn for neural architecture search (2019)
Ru, B., Wan, X., Dong, X., Osborne, M.: Interpretable neural architecture search via Bayesian optimisation with Weisfeiler-Lehman kernels (2021)
Ruan, D., Han, J., Yan, J., Gühmann, C.: Light convolutional neural network by neural architecture search and model pruning for bearing fault diagnosis and remaining useful life prediction. Sci. Rep. 13(1), 5484 (2023). https://doi.org/10.1038/s41598-023-31532-9
Stanley, K.O., Miikkulainen, R.: Evolving neural networks through augmenting topologies. Evol. Comput. 10(2), 99–127 (2002). https://doi.org/10.1162/106365602320169811
Tao, T.M., Kim, H., Youn, C.H.: A compact neural architecture search for accelerating image classification models. In: 2021 International Conference on Information and Communication Technology Convergence (ICTC), pp. 1713–1718 (2021). https://doi.org/10.1109/ICTC52510.2021.9620797
Vo-Ho, V.K., Yamazaki, K., Hoang, H., Tran, M.T., Le, N.: Chapter 19 - neural architecture search for medical image applications. In: Nguyen, H.V., Summers, R., Chellappa, R. (eds.) Meta Learning With Medical Imaging and Health Informatics Applications, The MICCAI Society book Series, pp. 369–384. Academic Press (2023). https://doi.org/10.1016/B978-0-32-399851-2.00029-6, https://www.sciencedirect.com/science/article/pii/B9780323998512000296
Wang, D., Gong, C., Li, M., Liu, Q., Chandra, V.: AlphaNet: improved training of supernets with alpha-divergence (2021)
Wang, W., Zhang, X., Cui, H., Yin, H., Zhang, Y.: FP-DARTS: fast parallel differentiable neural architecture search for image classification. Pattern Recognit. 136, 109193 (2023). https://doi.org/10.1016/j.patcog.2022.109193, https://www.sciencedirect.com/science/article/pii/S0031320322006720
Wang, Y., et al.: TextNAS: a neural architecture search space tailored for text representation (2019)
White, C., Neiswanger, W., Savani, Y.: BANANAS: Bayesian optimization with neural architectures for neural architecture search (2020)
White, C., Nolen, S., Savani, Y.: Exploring the loss landscape in neural architecture search (2021)
White, C., Zela, A., Ru, R., Liu, Y., Hutter, F.: How powerful are performance predictors in neural architecture search? Adv. Neural Inf. Process. Syst. 34 (2021)
Wistuba, M., Rawat, A., Pedapati, T.: A survey on neural architecture search (2019)
Wu, X., Hu, S., Wu, Z., Liu, X., Meng, H.: Neural architecture search for speech emotion recognition (2022)
Xie, L., Yuille, A.: Genetic CNN (2017)
Xie, S., Zheng, H., Liu, C., Lin, L.: SNAS: stochastic neural architecture search (2020)
Xie, X., Song, X., Lv, Z., Yen, G.G., Ding, W., Sun, Y.: Efficient evaluation methods for neural architecture search: a survey (2023)
Yao, X., Liu, Y.: A new evolutionary system for evolving artificial neural networks. IEEE Trans. Neural Netw. 8(3), 694–713 (1997). https://doi.org/10.1109/72.572107
Yu, Q., Yang, D., Roth, H., Bai, Y., Zhang, Y., Yuille, A.L., Xu, D.: C2FNAS: coarse-to-fine neural architecture search for 3D medical image segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4126–4135 (2020)
Zhou, H., Yang, M., Wang, J., Pan, W.: BayesNAS: a Bayesian approach for neural architecture search (2019)
Zimmer, L., Lindauer, M., Hutter, F.: Auto-pytorch: multi-fidelity metalearning for efficient and robust AutoDL. IEEE Trans. Pattern Anal. Mach. Intell. 43(9), 3079–3090 (2021). https://doi.org/10.1109/TPAMI.2021.3067763
Zoph, B., Le, Q.V.: Neural architecture search with reinforcement learning (2017)
Zoph, B., Vasudevan, V., Shlens, J., Le, Q.V.: Learning transferable architectures for scalable image recognition (2018)
Acknowledgements
This research has been funded through the call of the public business entity red.es, for 2021 grants for research and development projects in artificial intelligence and other digital technologies and their integration into value chains with Code: C005/21-ED, Funded by the European Union NextGenerationEU.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Alonso-García, M., Corchado, J.M. (2023). Neural Architecture Search: Practical Key Considerations. In: Ossowski, S., Sitek, P., Analide, C., Marreiros, G., Chamoso, P., Rodríguez, S. (eds) Distributed Computing and Artificial Intelligence, 20th International Conference. DCAI 2023. Lecture Notes in Networks and Systems, vol 740. Springer, Cham. https://doi.org/10.1007/978-3-031-38333-5_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-38333-5_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-38332-8
Online ISBN: 978-3-031-38333-5
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)