TY - JOUR
T1 - BNAS-v2: memory-efficient and performance-collapse-prevented broad neural architecture search
AU - Ding, Zixiang
AU - Chen, Yaran
AU - Li, Nannan
AU - Zhao, Dongbin
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2022/10/1
Y1 - 2022/10/1
N2 - In this article, we propose BNAS-v2 to further improve the efficiency of broad neural architecture search (BNAS), which employs a broad convolutional neural network (BCNN) as the search space. In BNAS, the single-path sampling-updating strategy of an overparameterized BCNN leads to terrible unfair training issue, which restricts the efficiency improvement. To mitigate the unfair training issue, we employ a continuous relaxation strategy to optimize all paths of the overparameterized BCNN simultaneously. However, continuous relaxation leads to a performance collapse issue that leads to the unsatisfactory performance of the learned BCNN. For that, we propose the confident learning rate (CLR) and introduce the combination of partial channel connections and edge normalization. Experimental results show that 1) BNAS-v2 delivers state-of-the-art search efficiency on both CIFAR-10 (0.05 GPU days, which is 4× faster than BNAS) and ImageNet (0.19 GPU days) with better or competitive performance; 2) the above two solutions are effectively alleviating the performance collapse issue; and 3) BNAS-v2 achieves powerful generalization ability on multiple transfer tasks, e.g., MNIST, FashionMNIST, NORB, and SVHN. The code is available at https://github.com/zixiangding/BNASv2.
AB - In this article, we propose BNAS-v2 to further improve the efficiency of broad neural architecture search (BNAS), which employs a broad convolutional neural network (BCNN) as the search space. In BNAS, the single-path sampling-updating strategy of an overparameterized BCNN leads to terrible unfair training issue, which restricts the efficiency improvement. To mitigate the unfair training issue, we employ a continuous relaxation strategy to optimize all paths of the overparameterized BCNN simultaneously. However, continuous relaxation leads to a performance collapse issue that leads to the unsatisfactory performance of the learned BCNN. For that, we propose the confident learning rate (CLR) and introduce the combination of partial channel connections and edge normalization. Experimental results show that 1) BNAS-v2 delivers state-of-the-art search efficiency on both CIFAR-10 (0.05 GPU days, which is 4× faster than BNAS) and ImageNet (0.19 GPU days) with better or competitive performance; 2) the above two solutions are effectively alleviating the performance collapse issue; and 3) BNAS-v2 achieves powerful generalization ability on multiple transfer tasks, e.g., MNIST, FashionMNIST, NORB, and SVHN. The code is available at https://github.com/zixiangding/BNASv2.
KW - Broad neural architecture search (BNAS)
KW - confident learning rate (CLR)
KW - continuous relaxation
KW - image classification
KW - partial channel connections (PC)
UR - http://www.scopus.com/inward/record.url?scp=85123774305&partnerID=8YFLogxK
U2 - 10.1109/TSMC.2022.3143201
DO - 10.1109/TSMC.2022.3143201
M3 - Article
AN - SCOPUS:85123774305
SN - 2168-2216
VL - 52
SP - 6259
EP - 6272
JO - IEEE Transactions on Systems, Man, and Cybernetics: Systems
JF - IEEE Transactions on Systems, Man, and Cybernetics: Systems
IS - 10
ER -