Quantum Dynamic Optimization Algorithm for Neural Architecture Search on Image Classification
Abstract
:1. Introduction
- In addition to conventional evolutionary algorithms, for the first time, this paper applies a quantum heuristic optimization algorithm as a search algorithm for a neural network architecture search problem. We transform the applicability of quantum dynamics optimization algorithms from traditional optimization problems to neural network architecture search problems. The designed algorithm does not depend on specific data and is a general neural network architecture search algorithm.
- Reduce the problem search space by defining reasonable discretization encoding methods, and quantum heuristic rules. The use of the quantum tunneling effect and barrier estimation principle makes the proposed algorithm more competitive with general evolutionary methods.
- Conduct extensive experiments on NAS-Benchmark to demonstrate the effectiveness of the proposed models.
2. Quantum Dynamic Optimization
Algorithm 1: Pseudocode of QDO. |
- (1)
- Generate k sampled individuals in the domain [,].
- (2)
- The probability evolution of the location distribution of k sampled individuals can be considered as the evolution of the particle wave function modulus. The larger the value of k, the closer to the probability distribution of the wave function modulus. The initial mean square error takes the length of the domain. When the initial mean square error is large, the algorithm is not sensitive to the initial position of the sampled individual.
- (3)
- Generate new solutions with a normal distribution , if the new solution ; that is, the new solution is better than the old solution, then the new solution is directly accepted; if the new solution is worse than the old solution, it can be considered from the physical image that the particle is blocked by the potential barrier, and the difference solution is accepted according to the probability that the barrier penetrates the transmission coefficient T.
- (4)
- This iterative process is repeated until the mean square error of the positions of the k sampled individuals is less than or equal to the mean square error of the current normal sampling.
- (5)
- Replacing the worst position with the mean of the sampled individuals , reduces the mean square error of normal sampling and enters a smaller scale to perform the same iterative process.
- (6)
- If the algorithm meets the set maximum function evolution times , the entire iterative process ends, and the optimal solution among the current k sampled individuals is output.
3. Proposed Method
3.1. NAS Problem Black Box Modeling
3.2. QDNAS
- (1)
- Initialize the population, specifying the dataset D to use.
- (2)
- Randomly sample the architecture in the search space and assign it to a queue .
- (3)
- The particles are discretized according to Equation (2).
- (4)
- Generate new particle according to .
- (5)
- If , then is assigned to . Otherwise, the poor solution is accepted with a certain probability. In this part, the probability is 0.1. This probability is selected on the basis of many trials.
- (6)
- Replace the worst position with the mean of the sampled individuals , and discretize the sampled individuals again.
- (7)
- Keep repeating lines 2 to 12 in QDNAS until the maximum number of iterations is reached.
Algorithm 2: Pseudocode of QDNAS. |
4. Experiments
4.1. Nas-Bench-101
4.2. Nas-Bench-201
4.3. Nas-Bench-1shot1
4.4. NATs-Bench
4.5. Results Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Yin, D.; Gontijo Lopes, R.; Shlens, J.; Cubuk, E.D.; Gilmer, J. A fourier perspective on model robustness in computer vision. Adv. Neural Inf. Process. Syst. 2019, 32, 13276–13286. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
- Guo, Y.; Luo, Y.; He, Z.; Huang, J.; Chen, J. Hierarchical neural architecture search for single image super-resolution. IEEE Signal Process. Lett. 2020, 27, 1255–1259. [Google Scholar] [CrossRef]
- Wang, Y.; Liu, Y.; Dai, W.; Li, C.; Zou, J.; Xiong, H. Learning Latent Architectural Distribution in Differentiable Neural Architecture Search via Variational Information Maximization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12312–12321. [Google Scholar]
- Elsken, T.; Metzen, J.H.; Hutter, F. Neural architecture search: A survey. J. Mach. Learn. Res. 2019, 20, 1997–2017. [Google Scholar]
- Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G. Evolving deep convolutional neural networks for image classification. IEEE Trans. Evol. Comput. 2019, 24, 394–407. [Google Scholar] [CrossRef] [Green Version]
- Stanley, K.O.; Miikkulainen, R. Evolving neural networks through augmenting topologies. Evol. Comput. 2002, 10, 99–127. [Google Scholar] [CrossRef] [PubMed]
- Sun, J.D.; Yao, C.; Liu, J.; Liu, W.; Yu, Z.K. GNAS-U 2 Net: A New Optic Cup and Optic Disc Segmentation Architecture With Genetic Neural Architecture Search. IEEE Signal Process. Lett. 2022, 29, 697–701. [Google Scholar] [CrossRef]
- Gong, M.; Liu, J.; Qin, A.K.; Zhao, K.; Tan, K.C. Evolving deep neural networks via cooperative coevolution with backpropagation. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 420–434. [Google Scholar] [CrossRef]
- Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.L.; Tan, J.; Le, Q.V.; Kurakin, A. Large-scale evolution of image classifiers. In Proceedings of the International Conference on Machine Learning. PMLR, Sydney, NSW, Australia, 6–11 August 2017; pp. 2902–2911. [Google Scholar]
- Niu, R.; Li, H.; Zhang, Y.; Kang, Y. Neural Architecture Search Based on Particle Swarm Optimization. In Proceedings of the 2019 3rd International Conference on Data Science and Business Analytics (ICDSBA), Istanbul, Turkey, 11–12 October 2019; pp. 319–324. [Google Scholar]
- Xie, L.; Yuille, A. Genetic cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1379–1388. [Google Scholar]
- Junior, F.E.F.; Yen, G.G. Particle swarm optimization of deep neural networks architectures for image classification. Swarm Evol. Comput. 2019, 49, 62–74. [Google Scholar] [CrossRef]
- Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G.; Lv, J. Automatically designing CNN architectures using the genetic algorithm for image classification. IEEE Trans. Cybern. 2020, 50, 3840–3854. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Xue, Y.; Wang, Y.; Liang, J.; Slowik, A. A self-adaptive mutation neural architecture search algorithm based on blocks. IEEE Comput. Intell. Mag. 2021, 16, 67–78. [Google Scholar] [CrossRef]
- Wang, P.; Xin, G.; Jiao, Y. Quantum Dynamics Interpretation of Black-box Optimization. arXiv 2021, arXiv:2106.13927. [Google Scholar]
- Zhang, M.; Li, H.; Pan, S.; Liu, T.; Su, S.W. One-Shot Neural Architecture Search via Novelty Driven Sampling. In Proceedings of the IJCAI, Yokohama, Japan, 11–17 July 2020; pp. 3188–3194. [Google Scholar]
- Jin, J.; Wang, P. Multiscale Quantum Harmonic Oscillator Algorithm with Guiding Information for Single Objective Optimization. Swarm Evol. Comput. 2021, 65, 100916. [Google Scholar] [CrossRef]
- Ying, C.; Klein, A.; Christiansen, E.; Real, E.; Murphy, K.; Hutter, F. Nas-bench-101: Towards reproducible neural architecture search. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 7105–7114. [Google Scholar]
- Zela, A.; Siems, J.; Hutter, F. Nas-bench-1shot1: Benchmarking and dissecting one-shot neural architecture search. arXiv 2020, arXiv:2001.10422. [Google Scholar]
- Dong, X.; Yang, Y. Nas-bench-201: Extending the scope of reproducible neural architecture search. arXiv 2020, arXiv:2001.00326. [Google Scholar]
- Dong, X.; Liu, L.; Musial, K.; Gabrys, B. Nats-bench: Benchmarking nas algorithms for architecture topology and size. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3634–3646. [Google Scholar] [CrossRef] [PubMed]
- Liu, Y.; Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G.; Tan, K.C. A survey on evolutionary neural architecture search. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–21. [Google Scholar] [CrossRef] [PubMed]
- Li, L.; Talwalkar, A. Random search and reproducibility for neural architecture search. In Proceedings of the Uncertainty in Artificial Intelligence. PMLR, virtual online, 3–6 August 2020; pp. 367–377. [Google Scholar]
- Kandasamy, K.; Neiswanger, W.; Schneider, J.; Poczos, B.; Xing, E.P. Neural architecture search with bayesian optimisation and optimal transport. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, Canada, 3–8 December 2018; pp. 2020–2029. [Google Scholar]
- Chen, Y.; Meng, G.; Zhang, Q.; Xiang, S.; Huang, C.; Mu, L.; Wang, X. Renas: Reinforced evolutionary neural architecture search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4787–4796. [Google Scholar]
- Santra, S.; Hsieh, J.W.; Lin, C.F. Gradient descent effects on differential neural architecture search: A survey. IEEE Access 2021, 9, 89602–89618. [Google Scholar] [CrossRef]
- Zoph, B.; Le, Q.V. Neural architecture search with reinforcement learning. arXiv 2016, arXiv:1611.01578. [Google Scholar]
- Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4780–4789. [Google Scholar]
NAS-101 | NAS-201 | NAS-1Shot1 | |
---|---|---|---|
RS | 0.940 | 0.939 | 0.946 |
TPE | 0.933 | 0.936 | 0.947 |
RE | 0.940 | 0.940 | 0.947 |
Ours | 0.943 | 0.942 | 0.947 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jin, J.; Zhang, Q.; He, J.; Yu, H. Quantum Dynamic Optimization Algorithm for Neural Architecture Search on Image Classification. Electronics 2022, 11, 3969. https://doi.org/10.3390/electronics11233969
Jin J, Zhang Q, He J, Yu H. Quantum Dynamic Optimization Algorithm for Neural Architecture Search on Image Classification. Electronics. 2022; 11(23):3969. https://doi.org/10.3390/electronics11233969
Chicago/Turabian StyleJin, Jin, Qian Zhang, Jia He, and Hongnian Yu. 2022. "Quantum Dynamic Optimization Algorithm for Neural Architecture Search on Image Classification" Electronics 11, no. 23: 3969. https://doi.org/10.3390/electronics11233969
APA StyleJin, J., Zhang, Q., He, J., & Yu, H. (2022). Quantum Dynamic Optimization Algorithm for Neural Architecture Search on Image Classification. Electronics, 11(23), 3969. https://doi.org/10.3390/electronics11233969