Science-advisor
REGISTER info/FAQ
Login
username
password
     
forgot password?
register here
 
Research articles
  search articles
  reviews guidelines
  reviews
  articles index
My Pages
my alerts
  my messages
  my reviews
  my favorites
 
 
Stat
Members: 3645
Articles: 2'504'928
Articles rated: 2609

25 April 2024
 
  » arxiv » 1906.8034

 Article overview



Disentangling feature and lazy learning in deep neural networks: an empirical study
Mario Geiger ; Stefano Spigler ; Arthur Jacot ; Matthieu Wyart ;
Date 19 Jun 2019
AbstractTwo distinct limits for deep learning as the net width $h oinfty$ have been proposed, depending on how the weights of the last layer scale with $h$. In the "lazy-learning" regime, the dynamics becomes linear in the weights and is described by a Neural Tangent Kernel $Theta$. By contrast, in the "feature-learning" regime, the dynamics can be expressed in terms of the density distribution of the weights. Understanding which regime describes accurately practical architectures and which one leads to better performance remains a challenge. We answer these questions and produce new characterizations of these regimes for the MNIST data set, by considering deep nets $f$ whose last layer of weights scales as $frac{alpha}{sqrt{h}}$ at initialization, where $alpha$ is a parameter we vary. We performed systematic experiments on two setups (A) fully-connected Softplus momentum full batch and (B) convolutional ReLU momentum stochastic. We find that (1) $alpha^*=frac{1}{sqrt{h}}$ separates the two regimes. (2) for (A) and (B) feature learning outperforms lazy learning, a difference in performance that decreases with $h$ and becomes hardly detectable asymptotically for (A) but is very significant for (B). (3) In both regimes, the fluctuations $delta f$ induced by initial conditions on the learned function follow $delta fsim1/sqrt{h}$, leading to a performance that increases with $h$. This improvement can be instead obtained at intermediate $h$ values by ensemble averaging different networks. (4) In the feature regime there exists a time scale $t_1simalphasqrt{h}$, such that for $tll t_1$ the dynamics is linear. At $tsim t_1$, the output has grown by a magnitude $sqrt{h}$ and the changes of the tangent kernel $|DeltaTheta|$ become significant. Ultimately, it follows $|DeltaTheta|sim(sqrt{h}alpha)^{-a}$ for ReLU and Softplus activation, with $a<2$ & $a o2$ when depth grows.
Source arXiv, 1906.8034
Services Forum | Review | PDF | Favorites   
 
Visitor rating: did you like this article? no 1   2   3   4   5   yes

No review found.
 Did you like this article?

This article or document is ...
important:
of broad interest:
readable:
new:
correct:
Global appreciation:

  Note: answers to reviews or questions about the article must be posted in the forum section.
Authors are not allowed to review their own article. They can use the forum section.

browser Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; ClaudeBot/1.0; +claudebot@anthropic.com)






ScienXe.org
» my Online CV
» Free


News, job offers and information for researchers and scientists:
home  |  contact  |  terms of use  |  sitemap
Copyright © 2005-2024 - Scimetrica