Name:
s41467-019-12552-4.pdf
Size:
842.1Kb
Format:
PDF
Description:
Final Published Version
Affiliation
Univ Arizona, Dept PsycholUniv Arizona, Cognit Sci Program
Issue Date
2019-11-05
Metadata
Show full item recordPublisher
NATURE PUBLISHING GROUPCitation
Wilson, R.C., Shenhav, A., Straccia, M. et al. The Eighty Five Percent Rule for optimal learning. Nat Commun 10, 4646 (2019) doi:10.1038/s41467-019-12552-4Journal
NATURE COMMUNICATIONSRights
Copyright © The Author(s) 2019. Open Access This article is licensed under a Creative Commons Attribution 4.0 International License. To view a copy of this license, visit http://creativecommons.org/ licenses/by/4.0/.Collection Information
This item from the UA Faculty Publications collection is made available by the University of Arizona with support from the University of Arizona Libraries. If you have questions, please contact us at repository@u.library.arizona.edu.Abstract
Researchers and educators have long wrestled with the question of how best to teach their clients be they humans, non-human animals or machines. Here, we examine the role of a single variable, the difficulty of training, on the rate of learning. In many situations we find that there is a sweet spot in which training is neither too easy nor too hard, and where learning progresses most quickly. We derive conditions for this sweet spot for a broad class of learning algorithms in the context of binary classification tasks. For all of these stochastic gradient-descent based learning algorithms, we find that the optimal error rate for training is around 15.87% or, conversely, that the optimal training accuracy is about 85%. We demonstrate the efficacy of this ‘Eighty Five Percent Rule’ for artificial neural networks used in AI and biologically plausible neural networks thought to describe animal learning.Note
Open access journalISSN
2041-1723PubMed ID
31690723Version
Final published versionSponsors
John Templeton Foundation; Center of Biomedical Research Excellence grant from National Institute of General Medical Sciences [P20GM103645]; United States Department of Health & Human Services National Institutes of Health (NIH) - National Institute on Aging (NIA) [R56 AG061888]ae974a485f413a2113503eed53cd6c53
10.1038/s41467-019-12552-4
Scopus Count
Collections
Except where otherwise noted, this item's license is described as Copyright © The Author(s) 2019. Open Access This article is licensed under a Creative Commons Attribution 4.0 International License. To view a copy of this license, visit http://creativecommons.org/ licenses/by/4.0/.
Related articles
- Universality of gradient descent neural network training.
- Authors: Welper G
- Issue date: 2022 Jun
- A learning rule for very simple universal approximators consisting of a single layer of perceptrons.
- Authors: Auer P, Burgsteiner H, Maass W
- Issue date: 2008 Jun
- A biologically plausible supervised learning method for spiking neural networks using the symmetric STDP rule.
- Authors: Hao Y, Huang X, Dong M, Xu B
- Issue date: 2020 Jan
- Biologically plausible deep learning - But how far can we go with shallow networks?
- Authors: Illing B, Gerstner W, Brea J
- Issue date: 2019 Oct
- Local online learning in recurrent networks with random feedback.
- Authors: Murray JM
- Issue date: 2019 May 24