1Northeastern University 2Swinburne University of Technology ✉Corresponding author
In this paper, we take a close look at the efficient differentiable Neural Architecture Search (NAS) problem through the lens of data pruning. A key insight is that the core samples in the pruned dataset can be interpreted as support vectors for networks to make informed decisions. Therefore, we introduce ProtoNAS, an innovative prototypical data pruning method that dynamically and explicitly measures the prototypical distance of each data sample to its class-specific clustering centroid (i.e., prototype). These prototypes are dynamically updated on-the-fly throughout the search phase, enabling flexible data pruning that adjusts with the architectures and network parameters. Compared to conventional cross-entropy-based or Euclidean-based distances, the prototypical distance not only considers the distance between a sample and its class-specific prototype, but also the relative distances to the prototypes of other classes in the embedding space, facilitating more precise score assignment. Furthermore, beyond sample pruning, we introduce a class pruning policy aimed at masking the classes with the lowest average scores. This additional policy is designed to further expedite the NAS without compromising the performance of the final searched architecture. Extensive experiments conducted on NAS-Bench-201 and NAS-Bench-1Shot1 search spaces demonstrate that ProtoNAS can achieve a significant 90\% pruning rate with minimal performance trade-offs or even improved performance when applied to DARTS and its variants, such as PC-DARTS and $\beta$-DARTS. Compared to State-Of-The-Art (SOTA) data pruning techniques, ProtoNAS achieves a superior balance between search performance and search cost.
Overview of ProtoNAS for accelerating neural architecture search: (a) In the prototypical data pruning, we introduce a novel scoring metric that is based on the distance between a sample's feature vectors output by a specific feature layer in the supernet $\mathcal{F}^*_{\alpha^t,\omega^t}$ and its corresponding class-specific prototype. The prototypes are dynamically updated throughout the search phase. The score of a sample is formulated as the prototypical distance and its running variance. We then propose an innovative prototypical class pruning policy $\mathcal{P}_\mathcal{C}$ that efficiently accelerate the search phase without compromising the overall performance of the searched architecture. In addition, a class-aware sample pruning policy $\mathcal{P}_\mathcal{D}$ is introduced according to the score of each sample. The sample pruning policy is able to maintain the sample distribution of the dataset, circumventing class bias after the pruning. (b) The proposed prototypical data pruning promises lossless searching acceleration with significant pruning rate.
Illustration of prototypical distance in the embedding space. Samples $\textbf{x}_b$ can be assigned with a larger prototypical distance than samples $\textbf{x}_a$, despite having a smaller Euclidean distance. Therefore, the employed prototypical distance facilitates more precise score assignments compared to alternative scoring metrics.
NAS-Bench-201: Comparison of the comprehensive performance of different data pruning methods using PC-DARTS on NAS-Bench-201 (CIFAR10).
NAS-Bench-1shot1: Comparison of the comprehensive performance of different data pruning methods using PC-DARTS on NAS-Bench-1shot1 (Space 3).
Data visualization: Distribution of samples at different stages when using the prototypical data pruning policy. (a) and (b) are the distributions of all samples and selected samples before class pruning, respectively. (c) and (d) are the distributions of all samples and selected samples after class pruning, respectively. The sample features are derived from the search phase of ProtoNAS (PC-DARTS) on NAS-Bench-1shot1-Space 1.
@ARTICLE{11267098,
author={Shi, Liming and Deng, Lan and Wu, Gaochang and Zhang, Jingxin and Ding, Jinliang},
journal={IEEE Transactions on Emerging Topics in Computational Intelligence},
title={ProtoNAS: Accelerating Differentiable Neural Architecture Search via Prototypical Data Pruning},
year={2025},
volume={},
number={},
pages={1-15}}