Conference paper Open Access

POPNASv2: An Efficient Multi-Objective Neural Architecture Search Technique

Andrea Falanti; Eugenio Lomurno; Stefano Samele; Danilo Ardagna; Matteo Matteucci

Automating the research for the best neural network model is a task that has gained more and more relevance in the last few years. In this context, Neural Architecture Search (NAS) represents the most effective technique whose results rival
the state of the art hand-crafted architectures.
However, this approach requires a lot of computational capabilities as well as research time, which make prohibitive its usage in many real-world scenarios.
With its sequential model-based optimization strategy, Progressive Neural Architecture Search (PNAS) represents a possible step forward to face this resources issue. Despite the quality of the found network architectures, this technique is still limited in research time.
A significant step in this direction has been done by Pareto-Optimal Progressive Neural Architecture Search (POPNAS), which expand PNAS with a time predictor to enable a trade-off between search time and accuracy, considering a multi-objective optimization problem.
This paper proposes a new version of the Pareto-Optimal Progressive Neural Architecture Search, called POPNASv2.
Our approach enhances its first version and improves its performance. 
We expanded the search space by adding new operators and improved the quality of both predictors to build more accurate Pareto fronts.
Moreover, we introduced cell equivalence checks and enriched the search strategy with an adaptive greedy exploration step.
Our efforts allow POPNASv2 to achieve PNAS-like performance with an average 4x factor search time speed-up.


The official version of this tool is located in the following link:  AndreaFalanti/popnas-v2 (

Files (725.7 kB)
Name Size
725.7 kB Download
All versions This version
Views 6464
Downloads 3232
Data volume 23.2 MB23.2 MB
Unique views 6060
Unique downloads 3232


Cite as