Large-Scale Time Series Forecasting with Meta-Learning

  • Chapter
  • First Online:
Forecasting with Artificial Intelligence

Abstract

Many industrial applications concern the forecasting of large numbers of time series. In such circumstances, selecting a proper prediction model for a time series can no longer depend on the forecaster's experience. The interest in time series forecasting with meta-learning has been growing in recent years, as it is a promising method for automatic forecasting model selection and combination. In this chapter, we briefly review the current development of meta-learning methods in time series forecasting, summarize a general meta-learning framework for time series forecasting, and discuss the key elements of establishing an effective meta-learning system. We then introduce a meta-learning python library named ‘tsfmeta’, which aims to make meta-learning available for researchers and time series forecasting practitioners in a unified, easy-to-use framework. The experimental evaluation of the ‘tsfmeta’ on two open-source datasets further shows the promising performance of meta-learning on time series forecasting in various disciplines. We also offer suggestions for further academic research in time series forecasting with meta-learning.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
EUR 29.95
Price includes VAT (Germany)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
EUR 128.39
Price includes VAT (Germany)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
EUR 171.19
Price includes VAT (Germany)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    Code and documentation are available in https://github.com/Shawn-nau/tsfmeta.

  2. 2.

    The demonstrative codes are available at https://github.com/Shawn-nau/tsfmeta/examples

References

  • Aiolli, F. (2011). Transfer learning by kernel meta-learning. Paper presented at the International Conference on Unsupervised and Transfer Learning workshop.

    Google Scholar 

  • Andrychowicz, M., Denil, M., Colmenarejo, S. G., Hoffman, M. W., Pfau, D., Schaul, T., & Freitas, N. d. (2016). Learning to learn by gradient descent by gradient descent. Paper presented at the Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain.

    Google Scholar 

  • Arinze, B., Kim, S.-L., & Anandarajan, M. (1997). Combining and selecting forecasting models using rule based induction. Computers & Operations Research, 24(5), 423–433. https://doi.org/10.1016/S0305-0548(96)00062-7

    Article  Google Scholar 

  • Barandas, M., Folgado, D., Fernandes, L., Santos, S., Abreu, M., Bota, P., … Gamboa, H. (2020). TSFEL: Time series feature extraction library. SoftwareX, 11, 100456. https://doi.org/10.1016/j.softx.2020.100456

  • Bengio, Y., Bengio, S., & Cloutier, J. (1991, 8–12 July 1991). Learning a synaptic learning rule. Paper presented at the IJCNN-91-Seattle International Joint Conference on Neural Networks.

    Google Scholar 

  • Bonissone, P. P. (2012). Lazy Meta-Learning: Creating Customized Model Ensembles on Demand. In J. Liu, C. Alippi, B. Bouchon-Meunier, G. W. Greenwood, & H. A. Abbass (Eds.), Advances in Computational Intelligence: IEEE World Congress on Computational Intelligence, WCCI 2012, Brisbane, Australia, June 10–15, 2012. Plenary/Invited Lectures (pp. 1–23). Berlin, Heidelberg: Springer Berlin Heidelberg.

    Google Scholar 

  • Cerqueira, V., Torgo, L., Pinto, F., & Soares, C. (2017). Arbitrated ensemble for time series forecasting. Paper presented at the (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2017. Lecture Notes in Computer Science, Cham.

    Google Scholar 

  • Christ, M., Braun, N., Neuffer, J., & Kempa-Liehr, A. W. (2018). Time series feature extraction on basis of scalable hypothesis tests (tsfresh – A Python package). Neurocomputing, 307, 72–77. https://doi.org/10.1016/j.neucom.2018.03.067

    Article  Google Scholar 

  • Collopy, F., & Armstrong, J. S. (1992). Rule-based forecasting: Development and validation of an expert systems approach to combining time series extrapolations. Management Science, 38(10), 1394–1414. https://doi.org/10.1287/mnsc.38.10.1394

    Article  Google Scholar 

  • Duncan, G. T., Gorr, W. L., & Szczypula, J. (2001). Forecasting analogous time series: Springer US.

    Google Scholar 

  • Fildes, R., & Petropoulos, F. (2015). Simple versus complex selection rules for forecasting many time series. Journal of Business Research, 68(8), 1692–1701. https://doi.org/10.1016/j.jbusres.2015.03.028

    Article  Google Scholar 

  • Giraud-Carrier, C. G. (2008). Metalearning - A Tutorial. Paper presented at the Tutorial at the 7th international conference on machine learning and applications (ICMLA), San Diego, California, USA.

    Google Scholar 

  • Ha, D., Dai, A. M., & Le, Q. V. (2017). HyperNetworks. Ar**v, abs/1609.09106.

    Google Scholar 

  • Henderson, T. (2023). Theft: Tools for handling extraction of features from time series.

    Google Scholar 

  • Hospedales, T. M., Antoniou, A., Micaelli, P., & Storkey, A. J. (2022). Meta-learning in neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44, 5149–5169.

    Google Scholar 

  • Hu, J., Shen, L., Albanie, S., Sun, G., & Wu, E. (2019). Squeeze-and-Excitation Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1–1. https://doi.org/10.1109/TPAMI.2019.2913372

  • Hyndman, R., Kang, Y., Montero-Manso, P., Talagala, T., Wang, E., Yang, Y., … Moorman, J. R. (2019). tsfeatures: Time series feature extraction. R package R package v.1.0.1.

    Google Scholar 

  • Januschowski, T., Gasthaus, J., Wang, Y., Salinas, D., Flunkert, V., Bohlke-Schneider, M., & Callot, L. (2020). Criteria for classifying forecasting methods. International Journal of Forecasting, 36(1), 167–177. https://doi.org/10.1016/j.ijforecast.2019.05.008

    Article  Google Scholar 

  • Kück, M., Crone, S. F., & Freitag, M. (2016). Meta-learning with neural networks and landmarking for forecasting model selection an empirical evaluation of different feature sets applied to industry data. Paper presented at the 2016 International Joint Conference on Neural Networks (IJCNN).

    Google Scholar 

  • Kadlec, P., & Gabrys, B. (2008). Learnt Topology Gating Artificial Neural Networks. IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), 2008, 2604–2611.

    Google Scholar 

  • Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., … Liu, T.-Y. (2017). LightGBM: A highly efficient gradient boosting decision tree. Paper presented at the Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, California, USA.

    Google Scholar 

  • Lemke, C., & Gabrys, B. (2010). Meta-learning for time series forecasting and forecast combination. Neurocomputing, 73(10), 2006–2016. https://doi.org/10.1016/j.neucom.2009.09.020

    Article  Google Scholar 

  • Li, K., & Malik, J. (2017). Learning to Optimize. Ar**v, abs/1606.01885.

    Google Scholar 

  • Ma, P., Zhang, Z., Wang, J., Zhang, W., Liu, J., Lu, Q., & Wang, Z. (2021). Review on the Application of Metalearning in Artificial Intelligence. Computational Intelligence and Neuroscience, 2021, 1560972. https://doi.org/10.1155/2021/1560972

    Article  Google Scholar 

  • Ma, S., & Fildes, R. (2020). Forecasting third-party mobile payments with implications for customer flow prediction. International Journal of Forecasting. https://doi.org/10.1016/j.ijforecast.2019.08.012

    Article  Google Scholar 

  • Ma, S., & Fildes, R. (2021). Retail sales forecasting with meta-learning. European Journal of Operational Research, 288(1), 111–128. https://doi.org/10.1016/j.ejor.2020.05.038

    Article  Google Scholar 

  • Makridakis, S., & Petropoulos, F. (2019). The M4 competition: Conclusions. International Journal of Forecasting. https://doi.org/10.1016/j.ijforecast.2019.05.006

    Article  Google Scholar 

  • Makridakis, S., Spiliotis, E., & Assimakopoulos, V. (2020). The M5 Accuracy competition: Results, findings and conclusions.

    Google Scholar 

  • Meade, N. (2000). Evidence for the selection of forecasting methods. Journal of Forecasting, 19(6), 515–535. https://doi.org/10.1002/1099-131X(200011)19:6%3c515::AID-FOR754%3e3.0.CO;2-7

    Article  Google Scholar 

  • Molina, M. D. M., Romero, C., Ventura, S., & Luna, J. M. (2012). Meta-learning approach for automatic parameter tuning: A case of study with educational datasets. Paper presented at the 5th International Conference on Educational Data Mining.

    Google Scholar 

  • Montero-Manso, P., Athanasopoulos, G., Hyndman, R. J., & Talagala, T. S. (2018). FFORMA: Feature-based Forecast Model Averaging. Working Paper 19/18.

    Google Scholar 

  • O'Hara-Wild, M., Hyndman, R., & Wang, E. (2022). Feasts: Feature extraction and statistics for time series. https://github.com/tidyverts/feasts/

  • Prudêncio, R. B. C., & Ludermir, T. B. (2004). Meta-learning approaches to selecting time series models. Neurocomputing, 61(1), 121–137.

    Article  Google Scholar 

  • Ravi, S., & Larochelle, H. (2017). Optimization as a model for few-shot learning. Paper presented at the International Conference on Learning Representations.

    Google Scholar 

  • Reif, M., Shafait, F., & Dengel, A. (2012). Meta-learning for evolutionary parameter optimization of classifiers. Machine Learning, 87(3), 357–380. https://doi.org/10.1007/s10994-012-5286-7

    Article  Google Scholar 

  • Rice, J. R. (1976). The Algorithm Selection Problem**This work was partially supported by the National Science Foundation through Grant GP-32940X. This chapter was presented as the George E. Forsythe Memorial Lecture at the Computer Science Conference, February 19, 1975, Washington, D. C. In M. Rubinoff & M. C. Yovits (Eds.), Advances in Computers (Vol. 15, pp. 65–118): Elsevier.

    Google Scholar 

  • Schmidhuber, J. (1992). Learning to control fast-weight memories: An alternative to dynamic recurrent networks. Neural Computation, 4(1), 131–139. https://doi.org/10.1162/neco.1992.4.1.131

    Article  Google Scholar 

  • Shah, C. (1997). Model selection in univariate time series forecasting using discriminant analysis. International Journal of Forecasting, 13(4), 489–500. https://doi.org/10.1016/S0169-2070(97)00031-9

    Article  Google Scholar 

  • Talagala, T., Hyndman, R., & Athanasopoulos, G. (2018). Meta-learning how to forecast time series. https://EconPapers.repec.org/RePEc:msh:ebswps:2018-6

  • Talagala, T. S., Li, F., & Kang, Y. (2022). FFORMPP: Feature-based forecast model performance prediction. International Journal of Forecasting, 38(3), 920–943. https://doi.org/10.1016/j.ijforecast.2021.07.002

    Article  Google Scholar 

  • Wang, J. X. (2021). Meta-learning in natural and artificial intelligence. Current Opinion in Behavioral Sciences, 38, 90–95. https://doi.org/10.1016/j.cobeha.2021.01.002

    Article  Google Scholar 

  • Wang, X., Smith-Miles, K., & Hyndman, R. (2009). Rule induction for forecasting method selection: Meta-learning the characteristics of univariate time series. Neurocomputing, 72(10), 2581–2594. https://doi.org/10.1016/j.neucom.2008.10.017

    Article  Google Scholar 

  • Widodo, A., & Budi, I. (2013). Model selection using dimensionality reduction of time series characteristics. Paper presented at the International Symposium on Forecasting, Seoul, South Korea.

    Google Scholar 

  • Wolpert, D. H., & Macready, W. G. (1997). No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation, 1(1), 67–82. https://doi.org/10.1109/4235.585893

    Article  Google Scholar 

Download references

Acknowledgements

The first author acknowledges the support of the National Natural Science Foundation of China under Grant No. 72072092; Jiangsu Social Science Foundation under Grant No. 22GLB016; and Jiangsu Philosophy and Social Science Research Major Project under Grant No. 2021SJZDA029.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Robert Fildes .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Ma, S., Fildes, R. (2023). Large-Scale Time Series Forecasting with Meta-Learning. In: Hamoudia, M., Makridakis, S., Spiliotis, E. (eds) Forecasting with Artificial Intelligence. Palgrave Advances in the Economics of Innovation and Technology. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-031-35879-1_9

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-35879-1_9

  • Published:

  • Publisher Name: Palgrave Macmillan, Cham

  • Print ISBN: 978-3-031-35878-4

  • Online ISBN: 978-3-031-35879-1

  • eBook Packages: Economics and FinanceEconomics and Finance (R0)

Publish with us

Policies and ethics

Navigation