Table of Contents Author Guidelines Submit a Manuscript
Journal of Applied Mathematics
Volume 2014, Article ID 238357, 8 pages
http://dx.doi.org/10.1155/2014/238357
Research Article

Intelligent Inventory Control via Ruminative Reinforcement Learning

1Faculty of Engineering, Khon Kaen University, Computer Engineering Building, 123 Moo 16, Mitraparb Road, Muang, Khon Kaen 40002, Thailand
2Department of Electrical and Computer Engineering, Colorado State University, 1373 Campus Delivery, Fort Collins, CO 80523-1373, USA

Received 30 December 2013; Accepted 29 May 2014; Published 7 July 2014

Academic Editor: Aderemi Oluyinka Adewumi

Copyright © 2014 Tatpong Katanyukul and Edwin K. P. Chong. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Linked References

  1. D. Bertsimas and A. Thiele, “A robust optimization approach to inventory theory,” Operations Research, vol. 54, no. 1, pp. 150–168, 2006. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet
  2. R. S. Sutton and A. G. Barto, Reinforcement Learning, MIT Press, Boston, Mass, USA, 1998.
  3. W. B. Powell, Approximate Dynamic Programming: Solving the Curses of Dimensionality, Wiley Series in Probability and Statistics, Wiley-Interscience, Hoboken, NJ, USA, 2007. View at Publisher · View at Google Scholar · View at MathSciNet
  4. J. Rao, X. Bu, C. Z. Xu, L. Wang, and G. Yin, “VCONF: a reinforcement learning approach to virtual machines auto-configuration,” in Proceedings of the 6th International Conference on Autonomic Computing (ICAC '09), pp. 137–146, ACM, Barcelona, Spain, June 2009. View at Publisher · View at Google Scholar · View at Scopus
  5. S. G. Khan, G. Herrmann, F. L. Lewis, T. Pipe, and C. Melhuish, “Reinforcement learning and optimal adaptive control: an overview and implementation examples,” Annual Reviews in Control, vol. 36, no. 1, pp. 42–59, 2012. View at Publisher · View at Google Scholar · View at Scopus
  6. A. Coates, P. Abbeel, and A. Y. Ng, “Apprenticeship learning for helicopter control,” Communications of the ACM, vol. 52, no. 7, pp. 97–105, 2009. View at Publisher · View at Google Scholar · View at Scopus
  7. C. W. Anderson, D. Hittle, M. Kretchmar, and P. Young, “Robust reinforcement learning for heating, ventilation, and air conditioning control of buildings,” in Handbook of Learning and Approximate Dynamic Programming, J. Si, A. Barto, W. Powell, and D. Wunsch, Eds., pp. 517–534, John Wiley & Sons, New York, NY, USA, 2004. View at Google Scholar
  8. R. Lincoln, S. Galloway, B. Stephen, and G. Burt, “Comparing policy gradient and value function based reinforcement learning methods in simulated electrical power trade,” IEEE Transactions on Power Systems, vol. 27, no. 1, pp. 373–380, 2012. View at Publisher · View at Google Scholar · View at Scopus
  9. Z. Tan, C. Quek, and P. Y. K. Cheng, “Stock trading with cycles: a financial application of ANFIS and reinforcement learning,” Expert Systems with Applications, vol. 38, no. 5, pp. 4741–4755, 2011. View at Publisher · View at Google Scholar · View at Scopus
  10. A. Castelletti, F. Pianosi, and M. Restelli, “A multiobjective reinforcement learning approach to water resources systems operation: pareto frontier approximation in a single run,” Water Resources Research, vol. 49, no. 6, pp. 3476–3486, 2013. View at Publisher · View at Google Scholar · View at Scopus
  11. T. Katanyukul, E. K. P. Chong, and W. S. Duff, “Intelligent inventory control: is bootstrapping worth implementing?” in Intelligent Information Processing VI, vol. 385 of IFIP Advances in Information and Communication Technology, pp. 58–67, Springer, New York, NY, USA, 2012. View at Publisher · View at Google Scholar · View at Scopus
  12. T. Akiyama, H. Hachiya, and M. Sugiyama, “Efficient exploration through active learning for value function approximation in reinforcement learning,” Neural Networks, vol. 23, no. 5, pp. 639–648, 2010. View at Publisher · View at Google Scholar · View at Scopus
  13. K. Doya, “Metalearning and neuromodulation,” Neural Networks, vol. 15, no. 4–6, pp. 495–506, 2002. View at Publisher · View at Google Scholar · View at Scopus
  14. P. Dayan and N. D. Daw, “Decision theory, reinforcement learning, and the brain,” Cognitive, Affective and Behavioral Neuroscience, vol. 8, no. 4, pp. 429–453, 2008. View at Publisher · View at Google Scholar · View at Scopus
  15. T. Katanyukul, W. S. Duff, and E. K. P. Chong, “Approximate dynamic programming for an inventory problem: empirical comparison,” Computers and Industrial Engineering, vol. 60, no. 4, pp. 719–743, 2011. View at Publisher · View at Google Scholar · View at Scopus
  16. C. O. Kim, I. H. Kwon, and J. G. Baek, “Asynchronous action-reward learning for nonstationary serial supply chain inventory control,” Applied Intelligence, vol. 28, no. 1, pp. 1–16, 2008. View at Publisher · View at Google Scholar · View at Scopus
  17. T. Katanyukul, “Ruminative reinforcement learning: improve intelligent inventory control by ruminating on the past,” in Proceedings of the 5th International Conference on Computer Science and Information Technology (ICCSIT '13), Academic Publisher, Paris, France, 2013.
  18. N. A. Vien, H. Yu, and T. Chung, “Hessian matrix distribution for Bayesian policy gradient reinforcement learning,” Information Sciences, vol. 181, no. 9, pp. 1671–1685, 2011. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet