Abstract
Many industrial applications concern the forecasting of large numbers of time series. In such circumstances, selecting a proper prediction model for a time series can no longer depend on the forecaster's experience. The interest in time series forecasting with meta-learning has been growing in recent years, as it is a promising method for automatic forecasting model selection and combination. In this chapter, we briefly review the current development of meta-learning methods in time series forecasting, summarize a general meta-learning framework for time series forecasting, and discuss the key elements of establishing an effective meta-learning system. We then introduce a meta-learning python library named ‘tsfmeta’, which aims to make meta-learning available for researchers and time series forecasting practitioners in a unified, easy-to-use framework. The experimental evaluation of the ‘tsfmeta’ on two open-source datasets further shows the promising performance of meta-learning on time series forecasting in various disciplines. We also offer suggestions for further academic research in time series forecasting with meta-learning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Code and documentation are available in https://github.com/Shawn-nau/tsfmeta.
- 2.
The demonstrative codes are available at https://github.com/Shawn-nau/tsfmeta/examples
References
Aiolli, F. (2011). Transfer learning by kernel meta-learning. Paper presented at the International Conference on Unsupervised and Transfer Learning workshop.
Andrychowicz, M., Denil, M., Colmenarejo, S. G., Hoffman, M. W., Pfau, D., Schaul, T., & Freitas, N. d. (2016). Learning to learn by gradient descent by gradient descent. Paper presented at the Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain.
Arinze, B., Kim, S.-L., & Anandarajan, M. (1997). Combining and selecting forecasting models using rule based induction. Computers & Operations Research, 24(5), 423–433. https://doi.org/10.1016/S0305-0548(96)00062-7
Barandas, M., Folgado, D., Fernandes, L., Santos, S., Abreu, M., Bota, P., … Gamboa, H. (2020). TSFEL: Time series feature extraction library. SoftwareX, 11, 100456. https://doi.org/10.1016/j.softx.2020.100456
Bengio, Y., Bengio, S., & Cloutier, J. (1991, 8–12 July 1991). Learning a synaptic learning rule. Paper presented at the IJCNN-91-Seattle International Joint Conference on Neural Networks.
Bonissone, P. P. (2012). Lazy Meta-Learning: Creating Customized Model Ensembles on Demand. In J. Liu, C. Alippi, B. Bouchon-Meunier, G. W. Greenwood, & H. A. Abbass (Eds.), Advances in Computational Intelligence: IEEE World Congress on Computational Intelligence, WCCI 2012, Brisbane, Australia, June 10–15, 2012. Plenary/Invited Lectures (pp. 1–23). Berlin, Heidelberg: Springer Berlin Heidelberg.
Cerqueira, V., Torgo, L., Pinto, F., & Soares, C. (2017). Arbitrated ensemble for time series forecasting. Paper presented at the (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2017. Lecture Notes in Computer Science, Cham.
Christ, M., Braun, N., Neuffer, J., & Kempa-Liehr, A. W. (2018). Time series feature extraction on basis of scalable hypothesis tests (tsfresh – A Python package). Neurocomputing, 307, 72–77. https://doi.org/10.1016/j.neucom.2018.03.067
Collopy, F., & Armstrong, J. S. (1992). Rule-based forecasting: Development and validation of an expert systems approach to combining time series extrapolations. Management Science, 38(10), 1394–1414. https://doi.org/10.1287/mnsc.38.10.1394
Duncan, G. T., Gorr, W. L., & Szczypula, J. (2001). Forecasting analogous time series: Springer US.
Fildes, R., & Petropoulos, F. (2015). Simple versus complex selection rules for forecasting many time series. Journal of Business Research, 68(8), 1692–1701. https://doi.org/10.1016/j.jbusres.2015.03.028
Giraud-Carrier, C. G. (2008). Metalearning - A Tutorial. Paper presented at the Tutorial at the 7th international conference on machine learning and applications (ICMLA), San Diego, California, USA.
Ha, D., Dai, A. M., & Le, Q. V. (2017). HyperNetworks. Ar**v, abs/1609.09106.
Henderson, T. (2023). Theft: Tools for handling extraction of features from time series.
Hospedales, T. M., Antoniou, A., Micaelli, P., & Storkey, A. J. (2022). Meta-learning in neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44, 5149–5169.
Hu, J., Shen, L., Albanie, S., Sun, G., & Wu, E. (2019). Squeeze-and-Excitation Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1–1. https://doi.org/10.1109/TPAMI.2019.2913372
Hyndman, R., Kang, Y., Montero-Manso, P., Talagala, T., Wang, E., Yang, Y., … Moorman, J. R. (2019). tsfeatures: Time series feature extraction. R package R package v.1.0.1.
Januschowski, T., Gasthaus, J., Wang, Y., Salinas, D., Flunkert, V., Bohlke-Schneider, M., & Callot, L. (2020). Criteria for classifying forecasting methods. International Journal of Forecasting, 36(1), 167–177. https://doi.org/10.1016/j.ijforecast.2019.05.008
Kück, M., Crone, S. F., & Freitag, M. (2016). Meta-learning with neural networks and landmarking for forecasting model selection an empirical evaluation of different feature sets applied to industry data. Paper presented at the 2016 International Joint Conference on Neural Networks (IJCNN).
Kadlec, P., & Gabrys, B. (2008). Learnt Topology Gating Artificial Neural Networks. IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), 2008, 2604–2611.
Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., … Liu, T.-Y. (2017). LightGBM: A highly efficient gradient boosting decision tree. Paper presented at the Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, California, USA.
Lemke, C., & Gabrys, B. (2010). Meta-learning for time series forecasting and forecast combination. Neurocomputing, 73(10), 2006–2016. https://doi.org/10.1016/j.neucom.2009.09.020
Li, K., & Malik, J. (2017). Learning to Optimize. Ar**v, abs/1606.01885.
Ma, P., Zhang, Z., Wang, J., Zhang, W., Liu, J., Lu, Q., & Wang, Z. (2021). Review on the Application of Metalearning in Artificial Intelligence. Computational Intelligence and Neuroscience, 2021, 1560972. https://doi.org/10.1155/2021/1560972
Ma, S., & Fildes, R. (2020). Forecasting third-party mobile payments with implications for customer flow prediction. International Journal of Forecasting. https://doi.org/10.1016/j.ijforecast.2019.08.012
Ma, S., & Fildes, R. (2021). Retail sales forecasting with meta-learning. European Journal of Operational Research, 288(1), 111–128. https://doi.org/10.1016/j.ejor.2020.05.038
Makridakis, S., & Petropoulos, F. (2019). The M4 competition: Conclusions. International Journal of Forecasting. https://doi.org/10.1016/j.ijforecast.2019.05.006
Makridakis, S., Spiliotis, E., & Assimakopoulos, V. (2020). The M5 Accuracy competition: Results, findings and conclusions.
Meade, N. (2000). Evidence for the selection of forecasting methods. Journal of Forecasting, 19(6), 515–535. https://doi.org/10.1002/1099-131X(200011)19:6%3c515::AID-FOR754%3e3.0.CO;2-7
Molina, M. D. M., Romero, C., Ventura, S., & Luna, J. M. (2012). Meta-learning approach for automatic parameter tuning: A case of study with educational datasets. Paper presented at the 5th International Conference on Educational Data Mining.
Montero-Manso, P., Athanasopoulos, G., Hyndman, R. J., & Talagala, T. S. (2018). FFORMA: Feature-based Forecast Model Averaging. Working Paper 19/18.
O'Hara-Wild, M., Hyndman, R., & Wang, E. (2022). Feasts: Feature extraction and statistics for time series. https://github.com/tidyverts/feasts/
Prudêncio, R. B. C., & Ludermir, T. B. (2004). Meta-learning approaches to selecting time series models. Neurocomputing, 61(1), 121–137.
Ravi, S., & Larochelle, H. (2017). Optimization as a model for few-shot learning. Paper presented at the International Conference on Learning Representations.
Reif, M., Shafait, F., & Dengel, A. (2012). Meta-learning for evolutionary parameter optimization of classifiers. Machine Learning, 87(3), 357–380. https://doi.org/10.1007/s10994-012-5286-7
Rice, J. R. (1976). The Algorithm Selection Problem**This work was partially supported by the National Science Foundation through Grant GP-32940X. This chapter was presented as the George E. Forsythe Memorial Lecture at the Computer Science Conference, February 19, 1975, Washington, D. C. In M. Rubinoff & M. C. Yovits (Eds.), Advances in Computers (Vol. 15, pp. 65–118): Elsevier.
Schmidhuber, J. (1992). Learning to control fast-weight memories: An alternative to dynamic recurrent networks. Neural Computation, 4(1), 131–139. https://doi.org/10.1162/neco.1992.4.1.131
Shah, C. (1997). Model selection in univariate time series forecasting using discriminant analysis. International Journal of Forecasting, 13(4), 489–500. https://doi.org/10.1016/S0169-2070(97)00031-9
Talagala, T., Hyndman, R., & Athanasopoulos, G. (2018). Meta-learning how to forecast time series. https://EconPapers.repec.org/RePEc:msh:ebswps:2018-6
Talagala, T. S., Li, F., & Kang, Y. (2022). FFORMPP: Feature-based forecast model performance prediction. International Journal of Forecasting, 38(3), 920–943. https://doi.org/10.1016/j.ijforecast.2021.07.002
Wang, J. X. (2021). Meta-learning in natural and artificial intelligence. Current Opinion in Behavioral Sciences, 38, 90–95. https://doi.org/10.1016/j.cobeha.2021.01.002
Wang, X., Smith-Miles, K., & Hyndman, R. (2009). Rule induction for forecasting method selection: Meta-learning the characteristics of univariate time series. Neurocomputing, 72(10), 2581–2594. https://doi.org/10.1016/j.neucom.2008.10.017
Widodo, A., & Budi, I. (2013). Model selection using dimensionality reduction of time series characteristics. Paper presented at the International Symposium on Forecasting, Seoul, South Korea.
Wolpert, D. H., & Macready, W. G. (1997). No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation, 1(1), 67–82. https://doi.org/10.1109/4235.585893
Acknowledgements
The first author acknowledges the support of the National Natural Science Foundation of China under Grant No. 72072092; Jiangsu Social Science Foundation under Grant No. 22GLB016; and Jiangsu Philosophy and Social Science Research Major Project under Grant No. 2021SJZDA029.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Ma, S., Fildes, R. (2023). Large-Scale Time Series Forecasting with Meta-Learning. In: Hamoudia, M., Makridakis, S., Spiliotis, E. (eds) Forecasting with Artificial Intelligence. Palgrave Advances in the Economics of Innovation and Technology. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-031-35879-1_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-35879-1_9
Published:
Publisher Name: Palgrave Macmillan, Cham
Print ISBN: 978-3-031-35878-4
Online ISBN: 978-3-031-35879-1
eBook Packages: Economics and FinanceEconomics and Finance (R0)