TY - GEN
T1 - Causal Unstructured Pruning in Linear Networks Using Effective Information
AU - Zeng, Changyu
AU - Liu, Li
AU - Zhao, Haocheng
AU - Zhang, Yu
AU - Wang, Wei
AU - Cai, Ning
AU - Yue, Yutao
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Excessive number of parameters in today's (deep) neural networks demands tremendous computational resources and slows down training speed. The problem also makes it difficult to deploy these neural network models on capability constrained devices such as mobile devices. To address this challenge, we propose an unstructured pruning method that measures the causal structure of neural networks based on effective information (EI). It introduces an intervention to the input and computes the mutual information between the interference and its corresponding output, within a single linear layer measuring the importance of each weight. In the experiments, we found that the sparsity of EI pruning can reach more than 90%. Only 10% of non-zero parameters in the linear layers were needed compared to the benchmark methods without pruning, while ensuring similar level of accuracy and stable training performance in iterative pruning. In addition, as the invariance of the causal structure of the network is exploited, the network after pruning using EI is highly generalizable and interpretable than other methods.
AB - Excessive number of parameters in today's (deep) neural networks demands tremendous computational resources and slows down training speed. The problem also makes it difficult to deploy these neural network models on capability constrained devices such as mobile devices. To address this challenge, we propose an unstructured pruning method that measures the causal structure of neural networks based on effective information (EI). It introduces an intervention to the input and computes the mutual information between the interference and its corresponding output, within a single linear layer measuring the importance of each weight. In the experiments, we found that the sparsity of EI pruning can reach more than 90%. Only 10% of non-zero parameters in the linear layers were needed compared to the benchmark methods without pruning, while ensuring similar level of accuracy and stable training performance in iterative pruning. In addition, as the invariance of the causal structure of the network is exploited, the network after pruning using EI is highly generalizable and interpretable than other methods.
KW - Causal Inference
KW - Deep Learning
KW - Effective Information
KW - Unstructured Pruning
UR - http://www.scopus.com/inward/record.url?scp=85153686412&partnerID=8YFLogxK
U2 - 10.1109/CyberC55534.2022.00056
DO - 10.1109/CyberC55534.2022.00056
M3 - Conference Proceeding
AN - SCOPUS:85153686412
T3 - Proceedings - 2022 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, CyberC 2022
SP - 294
EP - 302
BT - Proceedings - 2022 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, CyberC 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 12th International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, CyberC 2022
Y2 - 15 December 2022 through 16 December 2022
ER -