TY - JOUR
T1 - ModuleNet
T2 - Knowledge-Inherited Neural Architecture Search
AU - Chen, Yaran
AU - Gao, Ruiyuan
AU - Liu, Fenggang
AU - Zhao, Dongbin
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2022/11/1
Y1 - 2022/11/1
N2 - Although neural the architecture search (NAS) can bring improvement to deep models, it always neglects precious knowledge of existing models. The computation and time costing property in NAS also means that we should not start from scratch to search, but make every attempt to reuse the existing knowledge. In this article, we discuss what kind of knowledge in a model can and should be used for a new architecture design. Then, we propose a new NAS algorithm, namely, ModuleNet, which can fully inherit knowledge from the existing convolutional neural networks. To make full use of the existing models, we decompose existing models into different modules, which also keep their weights, consisting of a knowledge base. Then, we sample and search for a new architecture according to the knowledge base. Unlike previous search algorithms, and benefiting from inherited knowledge, our method is able to directly search for architectures in the macrospace by the NSGA-II algorithm without tuning parameters in these modules. Experiments show that our strategy can efficiently evaluate the performance of a new architecture even without tuning weights in convolutional layers. With the help of knowledge we inherited, our search results can always achieve better performance on various datasets (CIFAR10, CIFAR100, and ImageNet) over original architectures.
AB - Although neural the architecture search (NAS) can bring improvement to deep models, it always neglects precious knowledge of existing models. The computation and time costing property in NAS also means that we should not start from scratch to search, but make every attempt to reuse the existing knowledge. In this article, we discuss what kind of knowledge in a model can and should be used for a new architecture design. Then, we propose a new NAS algorithm, namely, ModuleNet, which can fully inherit knowledge from the existing convolutional neural networks. To make full use of the existing models, we decompose existing models into different modules, which also keep their weights, consisting of a knowledge base. Then, we sample and search for a new architecture according to the knowledge base. Unlike previous search algorithms, and benefiting from inherited knowledge, our method is able to directly search for architectures in the macrospace by the NSGA-II algorithm without tuning parameters in these modules. Experiments show that our strategy can efficiently evaluate the performance of a new architecture even without tuning weights in convolutional layers. With the help of knowledge we inherited, our search results can always achieve better performance on various datasets (CIFAR10, CIFAR100, and ImageNet) over original architectures.
KW - Evaluation algorithm
KW - knowledge inherited
KW - neural architecture search (NAS)
UR - http://www.scopus.com/inward/record.url?scp=85111043389&partnerID=8YFLogxK
U2 - 10.1109/TCYB.2021.3078573
DO - 10.1109/TCYB.2021.3078573
M3 - Article
C2 - 34097629
AN - SCOPUS:85111043389
SN - 2168-2267
VL - 52
SP - 11661
EP - 11671
JO - IEEE Transactions on Cybernetics
JF - IEEE Transactions on Cybernetics
IS - 11
ER -