Score: 1

HyperNAS: Enhancing Architecture Representation for NAS Predictor via Hypernetwork

Published: September 16, 2025 | arXiv ID: 2509.18151v1

By: Jindi Lv , Yuhao Zhou , Yuxin Tian and more

Potential Business Impact:

Finds best computer brain designs faster.

Business Areas:
Predictive Analytics Artificial Intelligence, Data and Analytics, Software

Time-intensive performance evaluations significantly impede progress in Neural Architecture Search (NAS). To address this, neural predictors leverage surrogate models trained on proxy datasets, allowing for direct performance predictions for new architectures. However, these predictors often exhibit poor generalization due to their limited ability to capture intricate relationships among various architectures. In this paper, we propose HyperNAS, a novel neural predictor paradigm for enhancing architecture representation learning. HyperNAS consists of two primary components: a global encoding scheme and a shared hypernetwork. The global encoding scheme is devised to capture the comprehensive macro-structure information, while the shared hypernetwork serves as an auxiliary task to enhance the investigation of inter-architecture patterns. To ensure training stability, we further develop a dynamic adaptive multi-task loss to facilitate personalized exploration on the Pareto front. Extensive experiments across five representative search spaces, including ViTs, demonstrate the advantages of HyperNAS, particularly in few-shot scenarios. For instance, HyperNAS strikes new state-of-the-art results, with 97.60\% top-1 accuracy on CIFAR-10 and 82.4\% top-1 accuracy on ImageNet, using at least 5.0$\times$ fewer samples.

Page Count
17 pages

Category
Computer Science:
Machine Learning (CS)