Abstract
Skill acquisition is a key area of research in cognitive psychology as it encompasses multiple psychological processes. The laws discovered under experimental paradigms are controversial and lack generalizability. This paper aims to unearth the laws of skill learning from large-scale training log data. A two-stage algorithm was developed to tackle the issues of unobservable cognitive states and an algorithmic explosion in searching. A deep learning model is initially employed to determine the learner’s cognitive state and assess the feature importance. Symbolic regression algorithms are then used to parse the neural network model into algebraic equations. Experimental results show that the algorithm can accurately restore preset laws within a noise range in continuous feedback settings. When applied to Lumosity training data, the method outperforms traditional and recent models in fitness terms. The study reveals two new forms of skill acquisition laws and reaffirms some previous findings.
This is a preview of subscription content, access via your institution
Access options
Access Nature and 54 other Nature Portfolio journals
Get Nature+, our best-value online-access subscription
9,800 Yen / 30 days
cancel any time
Subscribe to this journal
Receive 12 digital issues and online access to articles
¥11,900 per year
only ¥992 per issue
Buy this article
- Purchase on SpringerLink
- Instant access to full article PDF
Prices may be subject to local taxes which are calculated during checkout



Similar content being viewed by others
Data availability
All data generated or analyzed during this study are available via GitHub at https://github.com/ccnu-mathits/ADM (ref. 58) and Zenodo at https://doi.org/10.5281/zenodo.10938670 (ref. 59). Source Data are provided with this paper.
Code availability
The source code of this study is freely available on Github at https://github.com/ccnu-mathits/ADM (ref. 58), and via Zenodo at https://doi.org/10.5281/zenodo.10938670 (ref. 59).
References
VanLehn, K. Cognitive skill acquisition. Ann. Rev. Psychol. 47, 513–539 (1996).
DeKeyser, R. in Skill Acquisition Theory 83–104 (Routledge, 2020).
Tabibian, B. et al. Enhancing human learning via spaced repetition optimization. Proc. Natl Acad. Sci. USA 116, 3988–3993 (2019).
Evans, N. J., Brown, S. D., Mewhort, D. J. & Heathcote, A. Refining the law of practice. Psychol. Rev. 125, 592 (2018).
Heathcote, A., Brown, S. & Mewhort, D. J. The power law repealed: the case for an exponential law of practice. Psychon. Bull. Rev. 7, 185–207 (2000).
Shrager, J., Hogg, T. & Huberman, B. A. A graph-dynamic model of the power law of practice and the problem-solving fan-effect. Science 242, 414–416 (1988).
Wixted, J. T. The enigma of forgetting. Proc. Natl Acad. Sci. USA 119, e2201332119 (2022).
Averell, L. & Heathcote, A. The form of the forgetting curve and the fate of memories. J. Math. Psychol. 55, 25–35 (2011).
Roediger III, H. L. Relativity of remembering: why the laws of memory vanished. Annu. Rev. Psychol. 59, 225–254 (2008).
Chiaburu, D. S. & Marinova, S. V. What predicts skill transfer? An exploratory study of goal orientation, training self-efficacy and organizational supports. Int. J. Train. Dev. 9, 110–123 (2005).
Sturm, L. P. et al. A systematic review of skills transfer after surgical simulation training. Ann. Surgery 248, 166–179 (2008).
Logan, G. D. Toward an instance theory of automatization. Psycho. Rev. 95, 492 (1988).
Logan, G. D. Shapes of reaction-time distributions and shapes of learning curves: a test of the instance theory of automaticity. J. Exp. Psychol. 18, 883 (1992).
Anderson, J. R. Acquisition of cognitive skill. Psychol. Rev. 89, 369 (1982).
Tenison, C. & Anderson, J. R. Modeling the distinct phases of skill acquisition. J. Exp. Psychol. 42, 749 (2016).
Tenison, C., Fincham, J. M. & Anderson, J. R. Phases of learning: how skill acquisition impacts cognitive processing. Cognitive Psychol. 87, 1–28 (2016).
Jordan, M. I. Serial order: a parallel distributed processing approach. Adv. Pyschol. 121, 471–495 (1997).
McClelland, J. L. et al. Parallel Distributed Processing Vol. 2 (MIT Press, 1986).
Young, R. M. & Lewis, R. L. in Models of Working Memory: Mechanisms of Active Maintenance and Executive Control 224–256 (Cambridge Univ. Press, 1999).
Anderson, J. R., Matessa, M. & Lebiere, C. in Human–Computer Interaction Vol. 12, 439–462 (Lawrence Erlbaum Associates, 1997).
Ritter, F. E., Tehranchi, F. & Oury, J. D. ACT-R: a cognitive architecture for modeling cognition. WIREs Cogn. Sci. 10, e1488 (2019).
Goldstone, R. L. & Lupyan, G. Discovering psychological principles by mining naturally occurring data sets. Topics Cogn. Sci. 8, 548–568 (2016).
Jenkins, J. J., Cermak, L. & Craik, F. in Levels of Processing in Human Memory 429–446 (Pyschology, 1979).
Udrescu, S.-M. & Tegmark, M. AI Feynman: a physics-inspired method for symbolic regression. Sci. Adv. 6, eaay2631 (2020).
Wang, H. et al. Scientific discovery in the age of artificial intelligence. Nature 620, 47–60 (2023).
Schmidt, M. & Lipson, H. Distilling free-form natural laws from experimental data. Science 324, 81–85 (2009).
Cranmer, M. et al. Discovering symbolic models from deep learning with inductive biases. In 34th Conference on Neural Information Processing Systems 17429–17442 (NeurIPS, 2020).
Rudy, S. H., Brunton, S. L., Proctor, J. L. & Kutz, J. N. Data-driven discovery of partial differential equations. Sci. Adv. 3, e1602614 (2017).
Chen, Z., Liu, Y. & Sun, H. Physics-informed learning of governing equations from scarce data. Nat. Commun. 12, 6136 (2021).
Margraf, J. T., Jung, H., Scheurer, C. & Reuter, K. Exploring catalytic reaction networks with machine learning. Nat. Catal. 6, 112–121 (2023).
Han, Z.-K. et al. Single-atom alloy catalysts designed by first-principles calculations and artificial intelligence. Nat. Commun. 12, 1833 (2021).
Wang, Y., Wagner, N. & Rondinelli, J. M. Symbolic regression in materials science. MRS Commun. 9, 793–805 (2019).
He, M. & Zhang, L. Machine learning and symbolic regression investigation on stability of mxene materials. Comput. Mater. Sci. 196, 110578 (2021).
LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436–444 (2015).
Vaswani, A. et al. Attention is all you need. In Advances in Neural Information Processing Systems 30 (NIPS, 2017).
Baker, F. B. The Basics of Item Response Theory (ERIC, 2001).
Hambleton, R. K., Swaminathan, H. & Rogers, H. J. Fundamentals of Item Response Theory Vol. 2 (Sage, 1991).
Swaminathan, H. & Gifford, J. A. Bayesian estimation in the three-parameter logistic model. Psychometrika 51, 589–601 (1986).
Maris, G. & Bechger, T. On interpreting the model parameters for the three parameter logistic model. Measurement 7, 75–88 (2009).
Hornik, K. Approximation capabilities of multilayer feedforward networks. Neural Netw. 4, 251–257 (1991).
Raghu, M., Poole, B., Kleinberg, J., Ganguli, S. & Sohl-Dickstein, J. On the expressive power of deep neural networks. In Proc. 34th International Conference on Machine Learning 2847–2854 (PMLR, 2017).
Cybenko, G. Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. 2, 303–314 (1989).
Spearman, C. “General intelligence,” objectively determined and measured. Am. J. Psychol. 15, 201–293 (1904).
Eid, M., Geiser, C., Koch, T. & Heene, M. Anomalous results in g-factor models: explanations and alternatives. Psychol. Methods 22, 541 (2017).
Steyvers, M. & Schafer, R. J. Inferring latent learning factors in large-scale cognitive training data. Nat. Hum. Behav. 4, 1145–1155 (2020).
Simons, D. J. et al. Do “brain-training” programs work?. Psychol. Sci. Public Interest 17, 103–186 (2016).
Kievit, R. A. et al. Mutualistic coupling between vocabulary and reasoning supports cognitive development during late adolescence and early adulthood. Psychol. Sci. 28, 1419–1431 (2017).
Kievit, R. A., Hofman, A. D. & Nation, K. Mutualistic coupling between vocabulary and reasoning in young children: a replication and extension of the study by Kievit et al. (2017). Psychol. Sci. 30, 1245–1252 (2019).
Fisher, R. A. Design of experiments. Brit. Med. J. 1, 554 (1936).
Kumar, A., Benjamin, A. S., Heathcote, A. & Steyvers, M. Comparing models of learning and relearning in large-scale cognitive training data sets. NPJ Sci. Learn. 7, 24 (2022).
Liu, R. & Koedinger, K. R. Towards reliable and valid measurement of individualized student parameters. In Proc.10th International Conference on Educational Data Mining 135–142 (International Educational Data Mining Society, 2017).
Koedinger, K. R., Carvalho, P. F., Liu, R. & McLaughlin, E. A. An astonishing regularity in student learning rate. Proc. Natl Acad. Sci. USA 120, e2221311120 (2023).
Neath, A. A. & Cavanaugh, J. E. The bayesian information criterion: background, derivation, and applications. WIREs Comput. Stats 4, 199–203 (2012).
Vrieze, S. I. Model selection and psychological theory: a discussion of the differences between the Akaike information criterion (AIC) and the Bayesian information criterion (BIC). Psychol. Methods 17, 228 (2012).
Paszke, A. et al. PyTorch: an imperative style, high-performance deep learning library. In Proc. 33rd International Conference on Neural Information Processing Systems 8026–8037 (Curran Associates Inc., 2019).
Kingma, D. P. & Ba, J. Adam: a method for stochastic optimization. Preprint at https://arxiv.org/abs/1412.6980 (2015).
Steyvers, M. & Benjamin, A. S. The joint contribution of participation and performance to learning functions: exploring the effects of age in large-scale data sets. Behav. Res. Methods 51, 1531–1543 (2019).
Liu, S. et al. ccnu-mathits/ADM. GitHub https://github.com/ccnu-mathits/ADM (2024).
Liu, S. et al. ccnu-mathits/ADM: source code. Zenodo https://doi.org/10.5281/zenodo.10938670 (2024).
Acknowledgements
This work was jointly supported by the National Science and Technology Major Project (grant no. 2022ZD0117103 to J.S. and Z.Y.), the National Natural Science Foundation of China (grant no. 62293554 to J.S. and Z.Y., 62107017 to X.S. and 62077021 to J.S.), the Higher Education Science Research Program of China Association of Higher Education (grant no. 23XXK0301 to J.S.), the China Postdoctoral Science Foundation (grant no. 2023T160256 to X.S.), Hubei Provincial Natural Science Foundation of China (grant no. 2023AFA020 to S.L. and 2022CFB414 to J.S.), and Fundamental Research Funds for the Central Universities (grant no. CCNU23XJ007 to X.S. and CCNU22LJ005 to S.L.).
Author information
Authors and Affiliations
Contributions
S.L. and X.S. conceptualized the work. S.L. and X.S. designed the methodology. Q.L. performed investigations and curated the data, whereas S.L., Q.L. and Z.Y. performed the data validation. J.S. and Z.Y. administered the project. J.S. and Z.Y. acquired funding. X.S., J.S. and Z.Y. supervised the project. S.L. and X.S. wrote the original draft, whereas Q.L. and J.S. reviewed and edited it.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Computational Science thanks Martijn Meeter, Konstantina Sokratous and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Primary Handling Editor: Ananya Rastogi, in collaboration with the Nature Computational Science team.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Extended data
Extended Data Fig. 1 Partial results on the Lumosity dataset.
More specifically, (a) the change curve of mean fitting absolute error during the iteration process of the deep learning regressor; (b) the change curve of the value of the regularization term during the iteration processes; (c) the prediction error distribution of 1000 randomly selected records from the trained H1000+R1 model. The box plot displays the interquartile range (IQR) with the median line, while the whiskers extend to the minimum and maximum values or a multiple of the IQR from the quartiles. Outliers are depicted as individual points beyond the whiskers; (d) the proportion of the number of practice times for each skill to the total number of practice times; (e) the distribution of feature importance for each skill in H1000+R1.
Extended Data Fig. 2 Schematic representation of the simulation experiment.
(a) Flowchart depicting the process of generating simulated data and the data format. (b) Schematic diagram illustrating the evaluation procedure of the simulation experiment. (c) Sample representation of the simulated data.
Extended Data Fig. 3 The feature set utilized in the real-world dataset (Lumosity) experiment.
The feature set comprises three components: user (U), exercise (E), and scheduling (S). The features consist of two categories: discrete and continuous. Discrete features are encoded through one-hot encoding, while continuous features are encoded using real values. Game features are two-dimensional, consisting of skill and subskill, both of which are discrete features and characterize the relationship between the game and cognitive skills. Subskill is a subdivision feature of skill.
Extended Data Fig. 4 Analysis of the difference between the mastery level of skills calculated by the deep learning regressor and the discovered governing laws.
Here, we present the changes in skill proficiency at each time point during 2000 practice sessions for five learners. The blue line represents the mastery curve computed by the deep regressor, while the red line represents the curve computed by the discovered governing law. The governing law is the same for all learners, but due to differences in their choices and order of practice, there are variations in the independent variable of the governing law for each learner.
Supplementary information
Supplementary Information
Supplementary Figs. 1 and 2.
Source data
Source Data Fig. 1
Image source data for Fig. 1.
Source Data Fig. 2
Numerical source data for Fig. 2.
Source Data Fig. 3
Numerical source data for Fig. 3.
Source Data Extended Data Fig. 1
Numerical source data for Extended Data Fig. 1.
Source Data Extended Data Fig. 2
Image source data for Extended Data Fig. 2.
Source Data Extended Data Fig. 3
Image source data for Extended Data Fig. 3.
Source Data Extended Data Fig. 4
Numerical source data for Extended Data Fig. 4.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liu, S., Li, Q., Shen, X. et al. Automated discovery of symbolic laws governing skill acquisition from naturally occurring data. Nat Comput Sci 4, 334–345 (2024). https://doi.org/10.1038/s43588-024-00629-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1038/s43588-024-00629-0
This article is cited by
-
Outsourcing eureka moments to artificial intelligence
Nature Computational Science (2024)