Joint prediction of time series data in inventory management

  • Qifeng Zhou
  • Ruyuan Han
  • Tao Li
  • Bin Xia
Regular Paper


The problem of time series prediction has been well explored in the community of data mining. However, little research attention has been paid to the case of predicting the movement of a collection of related time series data. In this work, we study the problem of simultaneously predicting multiple time series data using joint predictive models. We observe that in real-world applications, strong relationships between different time-sensitive variables are often held, either explicitly predefined or implicitly covered in nature of the application. Such relationships indicate that the prediction on the trajectory of one given time series could be improved by incorporating the properties of other related time series data into predictive models. The key challenge is to capture the temporal dynamics of these relationships to jointly predict multiple time series. In this research, we propose a predictive model for multiple time series forecasting and apply it to the domain of inventory management. The relationships among multiple time series are modeled as a class of constraints, and in turn, refine the predictions on the corresponding time series. Experimental results on real-world data reveal that the proposed algorithms outperform well-established methods of time series forecasting.


Time series Joint prediction Inventory management 



This work is partially supported by the National Natural Science Foundation of China under Grant No. 61503313, by the Natural Science Foundation of Fujian Province of China No.2017J01118. The authors would also like to thank the former members of Florida International University KDRG research team for their support and contributions to this work.


  1. 1.
    Zipkin PH (2000) Foundations of inventory management. McGraw-Hill, New YorkzbMATHGoogle Scholar
  2. 2.
    Li L, Shen C, Wang L, et al (2014) iMiner: mining inventory data for intelligent management. In: Proceedings of the 23rd ACM international conference on conference on information and knowledge management. ACM, pp 2057–2059Google Scholar
  3. 3.
    Spedding TA, Chan KK (2000) Forecasting demand and inventory management using Bayesian time series. Integr Manuf Syst 11(5):331–339CrossRefGoogle Scholar
  4. 4.
    Weigend AS (1994) Time series prediction: forescasting the future and understanding the past. Routledge, AbingdonGoogle Scholar
  5. 5.
    Van Gestel T, Suykens JAK, Baestaens DE et al (2001) Financial time series prediction using least squares support vector machines within the evidence framework. IEEE Trans Neural Netw 12(4):809–821CrossRefGoogle Scholar
  6. 6.
    Taylor SJ (2007) Modelling financial time series. World Scientific, SinagaporeCrossRefGoogle Scholar
  7. 7.
    He Z, Wang XS, Lee BS et al (2008) Mining partial periodic correlations in time series. Knowl Inf Syst 15(1):31–54CrossRefGoogle Scholar
  8. 8.
    Aminikhanghahi S, Cook DJ (2017) A survey of methods for time series change point detection. Knowl Inf Syst 51(2):339–367CrossRefGoogle Scholar
  9. 9.
    Scarf H (2005) The optimality of (S,s) policies in the dynamic inventory problem. Mathematical Methods in the Social Sciences, New YorkCrossRefGoogle Scholar
  10. 10.
    Khan A, Yan X, Tao S, et al (2012) Workload characterization and prediction in the cloud: a multiple time series approach. In: IEEE network operations and management symposium (NOMS), pp 1287–1294Google Scholar
  11. 11.
    Banbura M, Giannone D, Reichlin L et al (2010) Large bayesian vector auto regressions. J Appl Econ 25(1):71–92MathSciNetCrossRefGoogle Scholar
  12. 12.
    Sims C (1980) Macroeconomics and reality. Econometrica 48(1):1–48CrossRefGoogle Scholar
  13. 13.
    Enders W (2010) Applied econometric time series, 3rd edn. Wiley, New YorkGoogle Scholar
  14. 14.
    Galbraith JW, Ullah A, Zinde-Walsh V (2002) Estimation of the vector moving average model by vector autoregression. Econom Rev 21(2):205–219MathSciNetCrossRefGoogle Scholar
  15. 15.
    Dubman M, Goodman RN (1969) Spectral analysis of multiple time series. Wiley, New YorkGoogle Scholar
  16. 16.
    Zhang XD, Takeda H (1987) An approach to time series analysis and ARMA spectral estimation. IEEE Trans Acoust Speech Signal Process 35(9):1303–1313MathSciNetCrossRefGoogle Scholar
  17. 17.
    Golyandina N, Zhigljavsky A (2013) Singular spectrum analysis for time series. Springer Science and Business Media, BerlinCrossRefGoogle Scholar
  18. 18.
    Widiputra H, Pears R, Kasabov N (2011) Multiple time-series prediction through multiple time-series relationships profiling and clustered recurring trends. In: Advances in knowledge discovery and data mining. Springer Berlin Heidelberg, pp 161–172Google Scholar
  19. 19.
    Finazzi F, Haggarty R, Miller C et al (2015) A comparison of clustering approaches for the study of the temporal coherence of multiple time series. Stoch Environ Res Risk Assess 29(2):463–475CrossRefGoogle Scholar
  20. 20.
    Pravilovic S, Bilancia M, Appice A et al (2017) Using multiple time series analysis for geosensor data forecasting. Inf Sci 380:31–52CrossRefGoogle Scholar
  21. 21.
    Frank RJ, Davey N, Hunt S et al (2001) Time series prediction and neural networks. J Intell Robot Syst 31:91–103CrossRefGoogle Scholar
  22. 22.
    Morariu N, Iancu E, Vlad S et al (2009) A neural network model for time-series forecasting. Romanian J Econ Forecast 12(4):213–223Google Scholar
  23. 23.
    Argyriou A, Evgeniou T, Pontil M (2008) Convex multi-task feature learning. Mach Learn 73(3):243–272CrossRefGoogle Scholar
  24. 24.
    Evgeniou T, Pontil M (2004) Regularized multi–task learning. In: Proceedings of the tenth ACM SIGKDD international conference on knowledge discovery and data mining. ACM, pp 109–117Google Scholar
  25. 25.
    Argyriou A, Evgeniou T, Pontil M (2007) Multi-task feature learning. In: Proceedings in advances in neural information processing systems (NIPS)Google Scholar
  26. 26.
    Chandra R, Ong YS, Goh CK (2017) Co-evolutionary multi-task learning with predictive recurrence for multi-step chaotic time series prediction. Neurocomputing 243:21–34CrossRefGoogle Scholar
  27. 27.
    Lugosi G, Papaspiliopoulos O, Stoltz G (2009) Online multi-task learning with hard constraints. arXiv preprint arXiv:0902.3526
  28. 28.
    Maggini M, Papini T (2010) Multitask semiCsupervised learning with constraints and constraint exceptions. In: Artificial neural networks CICANN. Springer, Berlin, pp 218–227Google Scholar
  29. 29.
    Zhang Y (2010) Multi-task active learning with output constraints. AAAI, Menlo ParkGoogle Scholar
  30. 30.
    Fiot JB, Dinuzzo F (2018) Electricity demand forecasting by multi-task learning. IEEE Trans Smart Grid 9(2):544–551CrossRefGoogle Scholar
  31. 31.
    Fiot JB, Dinuzzo F (2015) Electricity demand forecasting by multi-task learning. Comput Sci PP(99):1–1Google Scholar
  32. 32.
    Han L, Zhang Y (2015) Learning tree structure in multi-task learning. In: Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining. ACM, pp 397–406Google Scholar
  33. 33.
    Struyf J, Doeroski S (2005) Constraint based induction of multi-objective regression trees. In: International workshop on knowledge discovery in inductive databases. Springer, Berlin, pp 222–233Google Scholar
  34. 34.
    Pravilovic S, Appice A, Malerba D (2013) Process mining to forecast the future of running cases. International workshop on new frontiers in mining complex patterns. Springer, Cham, pp 67–81Google Scholar
  35. 35.
    Hamilton JD (1994) Time series analysis. Princeton University Press, PrincetonzbMATHGoogle Scholar
  36. 36.
    Evgeniou T, Micchelli CA, Pontil M (2005) Learning multiple tasks with kernel methods. J Mach Learn Res 6:615–637MathSciNetzbMATHGoogle Scholar
  37. 37.
    Müller KR, Smola AJ, Tsch G, et al (1997) Predicting time series with support vector machines. In: Artificial neural networks ICANN’97. Springer, Berlin, pp 999–1004Google Scholar
  38. 38.
    Bartlett PL, Wegkamp MH (2008) Classification with a reject option using a hinge loss. J Mach Learn Res 9:1823–1840MathSciNetzbMATHGoogle Scholar
  39. 39.
    Frank M, Wolfe P (1956) An algorithm for quadratic programming. Naval Res Logist Q 3(1–2):95–110MathSciNetCrossRefGoogle Scholar
  40. 40.
    Boyd S, Vandenberghe L (2004) Convex optimization. Cambridge University Press, CambridgeCrossRefGoogle Scholar
  41. 41.
    Smola AJ, Lkopf B (2004) A tutorial on support vector regression. Stat Comput 14(3):199–222MathSciNetCrossRefGoogle Scholar
  42. 42.
    Vert JP, Tsuda K, Lkopf B (2004) A primer on kernel methods. Kernel Methods Comput Biol 47:35–70Google Scholar
  43. 43.
    Platt J (1998) Sequential minimal optimization: a fast algorithm for training support vector machines. Microsoft Research Technical ReportGoogle Scholar
  44. 44.
    Zhang GP, Qi M (2005) Neural network forecasting for seasonal and trend time series. Eur J Oper Res 160(2):501–514MathSciNetCrossRefGoogle Scholar
  45. 45.
    Sapankevych NI, Sankar R (2009) Time series prediction using support vector machines: a survey. IEEE Computat Intell Mag 4(2):24–38CrossRefGoogle Scholar

Copyright information

© Springer-Verlag London Ltd., part of Springer Nature 2019

Authors and Affiliations

  1. 1.Department of AutomationXiamen UniversityXiamenChina
  2. 2.School of Computing and Information SciencesFlorida International UniversityMiamiUSA
  3. 3.School of Computer ScienceNanjing University of Posts and TelecommunicationsNanjingChina

Personalised recommendations