Abstract
In this chapter, we introduce the basic concept of the simulation-based optimization and illustrate its usefulness and applicability for generating the manpower planning of airline's cargo service call center. Because of the continuous increase in oil prices, and combined with many other factors, the airline industry is currently facing new challenges to keep its customers satisfied. One of the most important drivers of the customer satisfaction is the customer service. The excellent customer service can give an airline company the edge over its competitors. Airline companies need to insure the appropriate level of staffing at their service call centers in order to maintain a high level of customer satisfaction with the appropriate level of the overall cost. With the high level of uncertainty in the customer demand and a number of complicated factors in the problem, it becomes necessary to apply the simulation-based optimization technique to help managers generate the efficient staffing policy for the airline's cargo service call center. In this work, the technique called reinforcement learning and Markov decision process are used to build and solve the mathematical model to determine the appropriate staffing policy at the airline's cargo service call center on the monthly basis. Simulation and optimization models are incorporated together so as to solve the overall problem. The results of the case study are thoroughly analyzed, discussed, and compared with the current staffing policies. All results illustrate the impressive performance of the recommended staffing policies over the current staffing policies.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Gosavi, A., Simulation-Based Optimization: Parametric Optimization Techniques&Reinforcement Learning, Berlin Heidelberg New York, Springer, 2003.
Kleinman, N.L., Hill, S.D., and Ilenda, V.A., Simulation optimization of air traffic delay cost, Proceedings of the Winter Simulation Conference, Washington, DC, 1177–1181, 1998.
Rosenberger, J.M., Schafer, A.J., Goldsman, D., Johnson, E.L., Kleywegt, A.J., and Nemhauser, G.L., SIMAIR: A stochastic model of airline operations, Proceedings of the Winter Simulation Conference, 2000.
Lee, L.H., Hunag, H.C., Lee, C., Chew, E.P., Jaruphongsa, W., Yong, Y.Y., Liang, Z., Leong, C.H., Tan, Y.P., Namburi, K., Johnson, E., and Banks, J., Discrete event simulation model for airline operations: SIMAIR, Proceedings of the Winter Simulation Conference, 2003.
Sutton, R.S. and Barto, A.G., Reinforcement learning: An introduction, Cambridge, MA, MIT Press, 1998.
Bertsekas, D.P. and Tsitsiklis, J.N., Neuro-Dynamic Programming, Optimization and Neural Computer Series, 3, Belmont, MA, Athena Scientific, 1996.
Li Littman, M.L., Moore, A.W., and Kaelbling, L.P., Reinforcement Learning: A Survey, Brown University∕Carnegie Mellon University, Providence, RI∕Pittsburgh, PA.
Mahadevan, S., Average reward reinforcement learning: foundations, algorithms, and empirical results, Machine Learning, 22 (1), 159–195, 1996.
Creighton, D.C. and Nahavandi, S., Optimizing discrete event simulation models using a reinforcement learning agent, Winter Simulation Conference, Vol. 2, pp. 1945–1950, 2002.
Aydin, M.E., and Öztemel, E., Dynamic job shop scheduling using reinforcement learning agents, Robotics and Autonomous Systems, 33, 169–178, 2000.
Jeong, K., Conceptual frame for development of optimized simulation-based scheduling systems, Expert Systems with Applications, 18, 299–306, 2000.
Zhang, W. and Dietterich, T.G., A reinforcement learning approach to job-shop scheduling, Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, 1114–1120, 1995.
Reidmiller, S. and Reidmiller, ?., A neural reinforcement learning approach to learn local dispatching policies in production scheduling, Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence, 1999.
Schneider, J.G., Boyan, J.A., and Moore, A.W., Value function based production scheduling, Proceedings of the Fifteenth International Conference on Machine Learning, 24–27 July 1998.
Paternina-Arboleda, C.D. and Das, T.K., Intelligent dynamic control policies for serial production lines, IIE Transactions, 33, 65–77, 2001.
Mahadevan, S., Marchalleck, N., Das, T.K., and Gosavi A., Self-improving factory simulation using continuous-time average-reward reinforcement learning, Proceedings of the Fourteenth International Conference on Machine Learning, 202–210, 1997.
Mahadevan, S. and Theochaurus, G., Optimizing production manufacturing using reinforcement learning, Proceedings of the Eleventh International FLAIRS Conference, 372–377, 1998.
Sze, D.Y., A queuing model for telephone operator staffing, Operation Research, 32, 229–249, 1993.
Andrews, B. and Parsons, H., Establishing telephone-agent staffing levels through economic optimization, Interfaces, 23, 14–20, 1993.
Chen, B.P.K. and Henderson, S.G., Two issues in setting call centre staffing levels, Annals of Operations Research, 108, 175–192, 2001.
Fukunaga, A., Hamilton, E., Fama, J., Andre, D., Matan, O., and Nourbakhsh, I., Staff Scheduling for Inbound Call Centers and Customer Contact Centers, Blue Pumpkin Software, Sunnyvale, California, 2002.
Borst, S., Mandelbaum, A., and Reiman, M.I., Dimensioning large call centers, Operation Research, 52, 17–34, 2004.
Atlason, J., Epelman, M.A., and Henderson, S.G., Call center staffing with simulation and cutting plane methods, Annals of Operations Research, 127, 333–358, 2004.
Atlason, J., Epelman, M.A., and Henderson, S.G., Optimizing call center staffing using simulation and analytic center cutting plane methods, Department of Industrial and Operations Engineering, University of Michigan/School of Operations Research and Industrial Engineering, Cornell University, Ann Arbor, MI∕Ithaca, N Y, 2005.
Deslauriers, A., Ecuyer, P.L., Pichitlamken, J., Ingolfsson, A., and Avramidis, A.N., Markov chain models of a telephone call center with call blending, Computer and Operations Research, 34, 1616–1645, 2005.
Mourtada, I., Reinforcement Learning Model for Continental Airline's Service Center, ?.S. Thesis, Department of Industrial Engineering, the University of Houston, May 2007.
Puterman, M., Markov Decision Processes: Discrete Stochastic Dynamic Programming, Wiley Series in Probability and Mathematical Statistics, New York, Wiley, 1994.
Robbins, H. and Monro, S., A stochastic approximation method, Annals of Mathematical Statistics, 22, 400–407, 1951.
Kelton, W.D., Sadowski, R.P., and Sturrock, D.T., Simulation with Arena, 4th edition, New York, NY, Mc Graw Hill, 2007.
Johnson, R.A., Miller&Freund's Probability and Statistics for Engineers, 7th edition, Upper Saddle River, NJ, Pearson Prentice Hall, 2005.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer Science+Business Media, LLC
About this chapter
Cite this chapter
Assavapokee, T., Mourtada, I. (2008). Simulation-Based Optimization: A Case Study for Airline's Cargo Service Call Center. In: Kamrani, A.K., Nasr, E.S.A. (eds) Collaborative Engineering. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-47321-5_8
Download citation
DOI: https://doi.org/10.1007/978-0-387-47321-5_8
Publisher Name: Springer, Boston, MA
Print ISBN: 978-0-387-47319-2
Online ISBN: 978-0-387-47321-5
eBook Packages: EngineeringEngineering (R0)