Abstract
Self-learning agents can be used in numerous ways for dynamic pricing nowadays. It has been shown, that reinforcement learning can serve as a toolkit to efficiently develop pricing strategies in dynamic environments. In many real-world situations, it can be expected that multiple market participants rely on such self-learning agents to implement pricing decisions. From the view of one agent, this violates the fundamental Markov property, which leads to instability in the learning process. Past publications proposed to rely on asymmetric information to achieve equilibria and usually focused on tabular solutions or solvers. We use multi-agent learning and asymmetric information with function approximation tools for high-dimensional state spaces by exchanging policy information between multiple actors. We discuss possible problems and their solutions and propose a simulation environment for further evaluation of the developed system.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Goodfellow, I. J., Bengio, Y., & Courville, A. C. (2016). Deep learning. In Adaptive computation and machine learning. MIT Press.
Kastius, A., & Schlosser, R. (2022). Dynamic pricing under competition using reinforcement learning. Journal of Revenue and Pricing Management, 21, 50–63.
Könönen, V. (2003). Gradient based method for symmetric and asymmetric multiagent reinforcement learning. In IDEAL 2003, revised papers. Lecture Notes in Computer Science (Vol. 2690, pp. 68–75). Springer.
Könönen, V. (2004). Asymmetric multiagent reinforcement learning. Web Intelligence and Agent Systems, 2(2), 105–121.
Könönen, V. (2006). Dynamic pricing based on asymmetric multiagent reinforcement learning. International Journal of Intelligent Systems, 21(1), 73–98.
Kulkarni, T. D., Narasimhan, K., Saeedi, A., & Tenenbaum, J. (2016). Hierarchical deep reinforcement learning: Integrating temporal abstraction and intrinsic motivation. In NIPS 2016 (pp. 3675–3683).
Le, T. P., Vien, N. A., & Chung, T. (2018). A deep hierarchical reinforcement learning algorithm in partially observable Markov decision processes. IEEE Access, 6, 49089–49102.
Pereira, S. (2020). Stackelberg multi-agent reinforcement learning for hierarchical environments [Master’s thesis]. http://hdl.handle.net/10012/15851
Sutton, R. S., & Barto, A. G. (1998). Reinforcement learning—An introduction. In Adaptive computation and machine learning. MIT Press.
Tharakunnel, K., & Bhattacharyya, S. (2009). Single-leader-multiple-follower games with boundedly rational agents. Journal of Economic Dynamics and Control, 33, 1593–1603.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kastius, A., Kiele, N., Schlosser, R. (2023). Multi-agent Dynamic Pricing Using Reinforcement Learning and Asymmetric Information. In: Grothe, O., Nickel, S., Rebennack, S., Stein, O. (eds) Operations Research Proceedings 2022. OR 2022. Lecture Notes in Operations Research. Springer, Cham. https://doi.org/10.1007/978-3-031-24907-5_66
Download citation
DOI: https://doi.org/10.1007/978-3-031-24907-5_66
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-24906-8
Online ISBN: 978-3-031-24907-5
eBook Packages: Business and ManagementBusiness and Management (R0)