Skip to main content

Distributed Real-Time Scheduling by Using Multi-agent Reinforcement Learning

  • Chapter
  • First Online:
Multi-objective Evolutionary Optimisation for Product Design and Manufacturing
  • 4004 Accesses

Abstract

Autonomous Distributed Manufacturing Systems (ADMS) have been proposed to realise flexible control structures of manufacturing systems. In the previous researches, a real-time scheduling method based on utility values has been proposed and applied to the ADMS. Multi-agent reinforcement learning is newly proposed and implemented to the job agents and resource agents, in order to improve their coordination processes. The status, the action and the reward are defined for the individual job agents and the resource agents to evaluate the suitable utility values based on the status of the ADMS. Some case studies of the real-time scheduling have been carried out to verify the effectiveness of the proposed methods.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Moriwaki, & T., Sugimura, N. (1992). Object-oriented modelling of autonomous distributed manufacturing system and its application to real-time scheduling. Proceedings of the ICOOMS ’92 (pp. 207–212).

    Google Scholar 

  2. Kadar, B., Monostori, L., & Szelke, E. (1998). An object-oriented framework for developing distributed manufacturing architectures. Journal of Intelligent Manufacturing, 9, 173–179.

    Article  Google Scholar 

  3. Ueda, K. (1992). An approach to bionic manufacturing systems based on DNA-type information. Proceedings of the ICOOMS ‘92, (pp. 303–308).

    Google Scholar 

  4. Ueda, K., Hatono, I., Fujii, N., & Vaario, J. (2000). Reinforcement learning approach to biological manufacturing systems. Annals of the CIRP, 49, 343–346.

    Article  Google Scholar 

  5. Hendrik, B., Jo, W., Paul, V., Luc, B., & Patrick, P. (1998). Reference architecture for holonic manufacturing systems: PROSA. Computers in Industry, 37, 255–274.

    Article  Google Scholar 

  6. Sugimura, N., Tanimizu, Y., & Iwamura, K. (2004). A Study on real-time scheduling for holonic manufacturing system. CIRP Journal of Manufacturing Systems, 33(5), 467–475.

    Google Scholar 

  7. Iwamura, K., Okubo, N., Tanimizu, Y., & Sugimura, N. (2006). Real-time scheduling for holonic manufacturing systems based on estimation of future status. International Journal of Production Research, 44(18–19), 3657–3675.

    Article  MATH  Google Scholar 

  8. Iwamura, K., Nakano, A., Tanimizu, Y., & Sugimura, N. (2007). A study on real-time scheduling for holonic manufacturing systems -Simulation for estimation of future status by individual holons-. In M. Vladimir, V. Valeriy, & W. C. Armando (Eds.), LNAI 4659 HoloMAS 2007 (pp. 205–214). Heidelberg: Springer.

    Google Scholar 

  9. Sutton, R., & Barto, A. (1998). Reinforcement learning: an introduction. Cambridge: The MIT Press.

    Google Scholar 

  10. Paternina-Arboleda, C., & Das, T. (2005). A multi-agent reinforcement learning approach to obtaining dynamic control policies for stochastic lot scheduling problem. Simulation Modelling Practice and Theory, 13, 389–406.

    Article  Google Scholar 

  11. Kaelbling, L., Littman, M., & Moore, A. (1996). Reinforcement learning: a survey. Journal of Artificial Intelligence Research, 4, 237–285.

    Google Scholar 

  12. Sutton, R. (1988). Learning to predict by the methods of temporal differences. Machine Learning, 3, 9–44.

    Google Scholar 

  13. Wang, Y., & Usher, J. (2005). Application of reinforcement learning for agent-based production scheduling. Engineering Applications of Artificial Intelligence, 18, 73–82.

    Article  Google Scholar 

  14. Fujii, N., Takasu, R., Kobayashi, M., Ueda, K. (2005). Reinforcement learning based product dispatching scheduling in a semiconductor manufacturing system. Proceedings of the 38th CIRP International seminar on manufacturing systems, CD-ROM

    Google Scholar 

  15. Aydin, E., & Oztemel, E. (2000). Dynamic job-shop scheduling using reinforcement learning agents. Robotics and Autonomous Systems, 33, 169–178.

    Article  Google Scholar 

  16. Aissani, N., Beldjilali, B., & Trentesaux, D. (2009). Dynamic scheduling of maintenance tasks in the petroleum industry: A reinforcement approach. Engineering Applications of Artificial Intelligence, 22, 1089–1103.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Koji Iwamura .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag London Limited

About this chapter

Cite this chapter

Iwamura, K., Sugimura, N. (2011). Distributed Real-Time Scheduling by Using Multi-agent Reinforcement Learning. In: Wang, L., Ng, A., Deb, K. (eds) Multi-objective Evolutionary Optimisation for Product Design and Manufacturing. Springer, London. https://doi.org/10.1007/978-0-85729-652-8_11

Download citation

  • DOI: https://doi.org/10.1007/978-0-85729-652-8_11

  • Published:

  • Publisher Name: Springer, London

  • Print ISBN: 978-0-85729-617-7

  • Online ISBN: 978-0-85729-652-8

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics