Theoretical and Empirical Analysis of the Learning Rate and Momentum Factor in Neural Network Modeling for Stock Prediction | SpringerLink
Skip to main content

Theoretical and Empirical Analysis of the Learning Rate and Momentum Factor in Neural Network Modeling for Stock Prediction

  • Conference paper
Advances in Computation and Intelligence (ISICA 2008)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 5370))

Included in the following conference series:

Abstract

Neural Network training requires a large number of learning epochs. An appropriate learning rate is important to the overall performance of the training. Under a weight-update algorithm, a low learning rate would make the network learning slowly, and a high learning rate would make the weights and error function diverge. To optimize the model parameters, this paper presents theoretical and empirical analysis of learning rate in neural network modeling for its application in stock price prediction, an increasing learning rate approach is suggested for practice. The effect of momentum factor is also investigated to speed up the convergence for network training.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 11439
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 14299
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Nilsson, J. (ed.): Artificial Intelligence: A New Synthesis. Morgan Kaufmann Publishers, Inc., San Francisco (1998)

    MATH  Google Scholar 

  2. Pao, Y.-H. (ed.): Adaptive Pattern Recognition and Neural Networks. Addison-Wesley Publishing Company, Inc., Reading (1999)

    Google Scholar 

  3. Vidyasagar, M. (ed.): A Theory of Learning and Generalization. Springer, Heidelberg (1997)

    MATH  Google Scholar 

  4. NeuroShell2 User’s Manual, Ward System Group, Inc., Frederick, MD (2002)

    Google Scholar 

  5. McNelis, P.D. (ed.): Neural networks in Finance: Gaining predictive edge in the market. Elsevier Academic Press, Amsterdam (2005)

    Google Scholar 

  6. Reed, R.D., Marks, R.J.: Neural Smithing, Massachusetts Institute of Technology (1999)

    Google Scholar 

  7. Haykin, S.: Neural networks: A Comprehensive Foundation. MacMillan, New York (1994)

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Ke, J., Liu, X., Wang, G. (2008). Theoretical and Empirical Analysis of the Learning Rate and Momentum Factor in Neural Network Modeling for Stock Prediction. In: Kang, L., Cai, Z., Yan, X., Liu, Y. (eds) Advances in Computation and Intelligence. ISICA 2008. Lecture Notes in Computer Science, vol 5370. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-92137-0_76

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-92137-0_76

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-92136-3

  • Online ISBN: 978-3-540-92137-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics