Kernel Corrector LSTM | SpringerLink
Skip to main content

Kernel Corrector LSTM

  • Conference paper
  • First Online:
Advances in Intelligent Data Analysis XXII (IDA 2024)

Abstract

Forecasting methods are affected by data quality issues in two ways: 1. they are hard to predict, and 2. they may affect the model negatively when it is updated with new data. The latter issue is usually addressed by pre-processing the data to remove those issues. An alternative approach has recently been proposed, Corrector LSTM (cLSTM), which is a Read & Write Machine Learning (RW-ML) algorithm that changes the data while learning to improve its predictions. Despite promising results being reported, cLSTM is computationally expensive, as it uses a meta-learner to monitor the hidden states of the LSTM. We propose a new RW-ML algorithm, Kernel Corrector LSTM (KcLSTM), that replaces the meta-learner of cLSTM with a simpler method: Kernel Smoothing. We empirically evaluate the forecasting accuracy and the training time of the new algorithm and compare it with cLSTM and LSTM. Results indicate that it is able to decrease the training time while maintaining a competitive forecasting accuracy.

This work was partially funded by projects AISym4Med (101095387) supported by Horizon Europe Cluster 1: Health, ConnectedHealth (n.\(^{\underline{\text {o}}}\) 46858), supported by Competitiveness and Internationalisation Operational Programme (POCI) and Lisbon Regional Operational Programme (LISBOA 2020), under the PORTUGAL 2020 Partnership Agreement, through the European Regional Development Fund (ERDF); NextGenAI - Center for Responsible AI (2022-C05i0102-02), supported by IAPMEI, and also by FCT plurianual funding for 2020–2023 of LIACC (UIDB/00027/2020_UIDP/00027/2020) and SONAE IM Labs@FEUP.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 12583
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 15729
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Baghoussi, Y., Soares, C., Mendes-Moreira, J.: Corrector LSTM: built-in training data correction for improved time series forecasting. In: Proceedings of the 8th SIGKDD International Workshop on Mining and Learning from Time Series–Deep Forecasting: Models, Interpretability, and Applications, Washington DC, USA, pp. 1–8. ACM (2022)

    Google Scholar 

  2. Bailer-Jones, C., MacKay, D.J.C., Withers, P.J.A.: A recurrent neural network for modelling dynamical systems. Network 9(4), 531–47 (1998). https://api.semanticscholar.org/CorpusID:653765

  3. Baptista, A., Baghoussi, Y., Soares, C., Mendes-Moreira, J., Arantes, M.: Pastprop-RNN: improved predictions of the future by correcting the past (2021)

    Google Scholar 

  4. Bhowmik, P., Partha, A.S.: A data-centric approach to improve machine learning model’s performance in production. Int. J. Eng. Adv. Technol. (2021). https://api.semanticscholar.org/CorpusID:240328155

  5. Bowman, S.R., Potts, C., Manning, C.D.: Recursive neural networks can learn logical semantics. In: Workshop on Continuous Vector Space Models and their Compositionality (2014). https://api.semanticscholar.org/CorpusID:15618372

  6. Castro, J., Achanccaray Diaz, P., Sanches, I., Cue La Rosa, L., Nigri Happ, P., Feitosa, R.: Evaluation of recurrent neural networks for crop recognition from multitemporal remote sensing images (2017)

    Google Scholar 

  7. Cerqueira, V., Torgo, L., Mozetič, I.: Evaluating time series forecasting models: an empirical study on performance estimation methods. Mach. Learn. 109(11), 1997–2028 (2020). https://doi.org/10.1007/s10994-020-05910-7

    Article  MathSciNet  Google Scholar 

  8. Cerqueira, V., Torgo, L., Soares, C.: Machine learning vs statistical methods for time series forecasting: size matters (2019)

    Google Scholar 

  9. Dave, A., Russakovsky, O., Ramanan, D.: Predictive-corrective networks for action detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2067–2076 (2017). https://api.semanticscholar.org/CorpusID:2466592

  10. Dehaene, S., Changeux, J.P., Nadal, J.P.: Neural networks that learn temporal sequences by selection. Proc. Natl. Acad. Sci. USA 84(9), 2727–31 (1987). https://api.semanticscholar.org/CorpusID:7423734

  11. Denker, J.S.: Neural network models of learning and adaptation. Phys. D Nonlinear Phenom. 2, 216–232 (1986). https://api.semanticscholar.org/CorpusID:119988262

  12. Diebold, F., Mariano, R.: Comparing predictive accuracy. J. Bus. Econ. Stat. 13(3), 253–63 (1995). https://EconPapers.repec.org/RePEc:bes:jnlbes:v:13:y:1995:i:3:p:253-63

  13. Gelenbe, E.: Learning in the recurrent random neural network. Neural Comput. 5, 154–164 (1992). https://api.semanticscholar.org/CorpusID:38667978

  14. Ghahramani, Z.: Probabilistic machine learning and artificial intelligence. Nature 521, 452–459 (2015). https://api.semanticscholar.org/CorpusID:216356

  15. Hochreiter, S.: The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 6, 107–116 (1998). https://doi.org/10.1142/S0218488598000094

    Article  Google Scholar 

  16. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9, 1735–80 (1997). https://doi.org/10.1162/neco.1997.9.8.1735

    Article  Google Scholar 

  17. Kanarachos, S., Christopoulos, S.R.G., Chroneos, A., Fitzpatrick, M.E.: Detecting anomalies in time series data via a deep learning algorithm combining wavelets, neural networks and hilbert transform. Expert Syst. Appl. 85, 292–304 (2017). https://doi.org/10.1016/j.eswa.2017.04.028. https://www.sciencedirect.com/science/article/pii/S0957417417302737

  18. Makridakis, S., Spiliotis, E., Assimakopoulos, V.: The m4 competition: results, findings, conclusion and way forward. Int. J. Forecast. 34(4), 802–808 (2018). https://doi.org/10.1016/j.ijforecast.2018.06.001. https://www.sciencedirect.com/science/article/pii/S0169207018300785

  19. Maya, S., Ueno, K., Nishikawa, T.: dLSTM: a new approach for anomaly detection using deep learning with delayed prediction. Int. J. Data Sci. Anal. 8, 137–164 (2019). https://doi.org/10.1007/s41060-019-00186-0

    Article  Google Scholar 

  20. Otte, C.: Safe and interpretable machine learning: a methodological review (2013). https://api.semanticscholar.org/CorpusID:56899177

  21. Sherstinsky, A.: Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 404, 132306 (2020). https://doi.org/10.1016/j.physd.2019.132306. https://www.sciencedirect.com/science/article/pii/S0167278919305974

  22. Song, C., Ristenpart, T., Shmatikov, V.: Machine learning models that remember too much. In: Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security (2017). https://api.semanticscholar.org/CorpusID:2904063

  23. Staudemeyer, R.C., Morris, E.R.: Understanding LSTM – a tutorial into long short-term memory recurrent neural networks (2019)

    Google Scholar 

  24. Strobelt, H., Gehrmann, S., Pfister, H., Rush, A.M.: LSTMVis: a tool for visual analysis of hidden state dynamics in recurrent neural networks (2017)

    Google Scholar 

  25. Zha, D., et al.: Data-centric artificial intelligence: a survey (2023)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rodrigo Tuna .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tuna, R., Baghoussi, Y., Soares, C., Mendes-Moreira, J. (2024). Kernel Corrector LSTM. In: Miliou, I., Piatkowski, N., Papapetrou, P. (eds) Advances in Intelligent Data Analysis XXII. IDA 2024. Lecture Notes in Computer Science, vol 14642. Springer, Cham. https://doi.org/10.1007/978-3-031-58553-1_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-58553-1_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-58555-5

  • Online ISBN: 978-3-031-58553-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics