TY - GEN
T1 - Evaluating and Selecting Deep Reinforcement Learning Models for OptimalDynamic Pricing
T2 - 8th International Conference on Control Engineering and Artificial Intelligence, CCEAI 2024
AU - Liu, Yuchen
AU - Man, Ka Lok
AU - Li, Gangmin
AU - Payne, Terry R.
AU - Yue, Yong
N1 - Publisher Copyright:
© 2024 ACM.
PY - 2024/1/26
Y1 - 2024/1/26
N2 - Given the plethora of available solutions, choosing an appropriate Deep Reinforcement Learning (DRL) model for dynamic pricing poses a significant challenge for practitioners. While many DRL solutions claim superior performance, there lacks a standardized framework for their evaluation. Addressing this gap, we introduce a novel framework and a set of metrics to select and assess DRL models systematically. To validate the utility of our framework, we critically compared three representative DRL models, emphasizing their performance in dynamic pricing tasks. Further ensuring the robustness of our assessment, we benchmarked these models against a well-established human agent policy. The DRL model that emerged as the most effective was rigorously tested on an Amazon dataset, demonstrating a notable performance boost of 5.64%. Our findings underscore the value of our proposed metrics and framework in guiding practitioners towards the most suitable DRL solution for dynamic pricing.
AB - Given the plethora of available solutions, choosing an appropriate Deep Reinforcement Learning (DRL) model for dynamic pricing poses a significant challenge for practitioners. While many DRL solutions claim superior performance, there lacks a standardized framework for their evaluation. Addressing this gap, we introduce a novel framework and a set of metrics to select and assess DRL models systematically. To validate the utility of our framework, we critically compared three representative DRL models, emphasizing their performance in dynamic pricing tasks. Further ensuring the robustness of our assessment, we benchmarked these models against a well-established human agent policy. The DRL model that emerged as the most effective was rigorously tested on an Amazon dataset, demonstrating a notable performance boost of 5.64%. Our findings underscore the value of our proposed metrics and framework in guiding practitioners towards the most suitable DRL solution for dynamic pricing.
KW - DDPG (Deep Deterministic Policy Gradient)
KW - Deep Reinforcement Learning (DRL)
KW - Dynamic Pricing
KW - E-commerce
KW - Inventory Management
KW - Markov Decision Process
KW - Model Evaluation
KW - PPO (Proximal Policy Optimization)
KW - Price Elasticity of Demand
KW - SAC (Soft Actor-Critic)
UR - http://www.scopus.com/inward/record.url?scp=85188251454&partnerID=8YFLogxK
U2 - 10.1145/3640824.3640871
DO - 10.1145/3640824.3640871
M3 - Conference Proceeding
AN - SCOPUS:85188251454
T3 - ACM International Conference Proceeding Series
SP - 215
EP - 219
BT - Proceedings - 2024 8th International Conference on Control Engineering and Artificial Intelligence, CCEAI 2024
A2 - Zhang, Wenqiang
A2 - Yue, Yong
A2 - Ogiela, Marek
PB - Association for Computing Machinery
Y2 - 26 January 2024 through 28 January 2024
ER -