TY - GEN
T1 - Device Placement Optimization for Deep Neural Networks via One-shot Model and Reinforcement Learning
AU - Ding, Zixiang
AU - Chen, Yaran
AU - Li, Nannan
AU - Zhao, Dongbin
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/12/1
Y1 - 2020/12/1
N2 - With the development of deep learning that employs deep neural networks (DNN) as powerful tool, its computational requirement grows rapid together with the increasing size (e.g. depth and parameter) of DNN. Currently, model and data parallelism are employed for accelerating the training and inference process of DNN. However, the above techniques make placement decision on devices for DNN based on heuristics and intuitions by machine learning experts. In this paper, we propose an novel approach for designing device placement of DNN in an automatic way. For a DNN, we employ a sequence-to-sequence model as controller to sample device placement from a one-shot model, which contains all possible device placements with respect to a specific hardware environment (e.g. CPU and GPU). Then, reinforcement learning treats the execution time of sampled device placement as reward to guide the sequence-to-sequence model for finding better one. The proposed approach is employed to optimize the device placement for both model and data parallelism of Inception-V3 on ImageNet. Experimental results show that the optimal placements discovered by our method can outperform hand-crafted one.
AB - With the development of deep learning that employs deep neural networks (DNN) as powerful tool, its computational requirement grows rapid together with the increasing size (e.g. depth and parameter) of DNN. Currently, model and data parallelism are employed for accelerating the training and inference process of DNN. However, the above techniques make placement decision on devices for DNN based on heuristics and intuitions by machine learning experts. In this paper, we propose an novel approach for designing device placement of DNN in an automatic way. For a DNN, we employ a sequence-to-sequence model as controller to sample device placement from a one-shot model, which contains all possible device placements with respect to a specific hardware environment (e.g. CPU and GPU). Then, reinforcement learning treats the execution time of sampled device placement as reward to guide the sequence-to-sequence model for finding better one. The proposed approach is employed to optimize the device placement for both model and data parallelism of Inception-V3 on ImageNet. Experimental results show that the optimal placements discovered by our method can outperform hand-crafted one.
KW - controller
KW - deep neural networks
KW - device placement
KW - one-shot model
KW - reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85099704921&partnerID=8YFLogxK
U2 - 10.1109/SSCI47803.2020.9308141
DO - 10.1109/SSCI47803.2020.9308141
M3 - Conference Proceeding
AN - SCOPUS:85099704921
T3 - 2020 IEEE Symposium Series on Computational Intelligence, SSCI 2020
SP - 1478
EP - 1484
BT - 2020 IEEE Symposium Series on Computational Intelligence, SSCI 2020
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2020 IEEE Symposium Series on Computational Intelligence, SSCI 2020
Y2 - 1 December 2020 through 4 December 2020
ER -