πŸ“„
Paper

Universal scaling laws in the gradient descent training of neural networks

by Maksim Velikanov ID: arxiv-paper--2105.00507

Current theoretical results on optimization trajectories of neural networks trained by gradient descent typically have the form of rigorous but potentially loose bounds on the loss values. In the present work we take a different approach and show that the learning trajectory can be characterized by ...

High Impact - Citations
2021 Year
ArXiv Venue
Top 19% FNI Rank
Paper Information Summary
Entity Passport
Registry ID arxiv-paper--2105.00507
Provider arxiv
πŸ“œ

Cite this paper

Academic & Research Attribution

BibTeX
@misc{arxiv_paper__2105.00507,
  author = {Maksim Velikanov},
  title = {Universal scaling laws in the gradient descent training of neural networks Paper},
  year = {2026},
  howpublished = {\url{https://arxiv.org/abs/2105.00507v1}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
Maksim Velikanov. (2026). Universal scaling laws in the gradient descent training of neural networks [Paper]. Free2AITools. https://arxiv.org/abs/2105.00507v1

πŸ”¬Technical Deep Dive

Full Specifications [+]

βš–οΈ Free2AI Nexus Index

Methodology β†’ πŸ“˜ What is FNI?
0.0
Top 19% Overall Impact
πŸ”₯ Popularity (P) 0
πŸš€ Velocity (V) 0
πŸ›‘οΈ Credibility (C) 0
πŸ”§ Utility (U) 0
Nexus Verified Data

πŸ’¬ Why this score?

The Nexus Index for Universal scaling laws in the gradient descent training of neural networks aggregates Popularity (P:0), Velocity (V:0), and Credibility (C:0). The Utility score (U:0) represents deployment readiness, context efficiency, and structural reliability within the Nexus ecosystem.

Data Verified πŸ• Last Updated: Not calculated
Free2AI Nexus Index | Fair Β· Transparent Β· Explainable | Full Methodology

πŸ“ Executive Summary

"Current theoretical results on optimization trajectories of neural networks trained by gradient descent typically have the form of rigorous but potentially loose bounds on the loss values. In the present work we take a different approach and show that the learning trajectory can be characterized by an explicit asymptotic at large training times. Specifically, the leading term in the asymptotic expansion of the loss behaves as a power law $L(t) \sim t^{-ΞΎ}$ with exponent $ΞΎ$ expressed only thr..."

❝ Cite Node

@article{Velikanov2021Universal,
  title={Universal scaling laws in the gradient descent training of neural networks},
  author={Maksim Velikanov and Dmitry Yarotsky},
  journal={arXiv preprint arXiv:arxiv-paper--2105.00507},
  year={2021}
}

πŸ‘₯ Collaborating Minds

Maksim Velikanov Dmitry Yarotsky

Abstract & Analysis

Current theoretical results on optimization trajectories of neural networks trained by gradient descent typically have the form of rigorous but potentially loose bounds on the loss values. In the present work we take a different approach and show that the learning trajectory can be characterized by an explicit asymptotic at large training times. Specifically, the leading term in the asymptotic expansion of the loss behaves as a power law $L(t) \sim t^{-ΞΎ}$ with exponent $ΞΎ$ expressed only through the data dimension, the smoothness of the activation function, and the class of function being approximated. Our results are based on spectral analysis of the integral operator representing the linearized evolution of a large network trained on the expected loss. Importantly, the techniques we employ do not require specific form of a data distribution, for example Gaussian, thus making our findings sufficiently universal.

πŸ”„ Daily sync (03:00 UTC)

AI Summary: Based on arXiv metadata. Not a recommendation.

πŸ“Š FNI Methodology πŸ“š Knowledge Baseℹ️ Verify with original source

πŸ›‘οΈ Paper Transparency Report

Verified data manifest for traceability and transparency.

100% Data Disclosure Active

πŸ†” Identity & Source

id
arxiv-paper--2105.00507
source
arxiv
author
Maksim Velikanov
tags
arxiv:cs.LGarxiv:cs.NEarxiv:math.OCarxiv:stat.MLneural

βš™οΈ Technical Specs

architecture
null
params billions
null
context length
null

πŸ“Š Engagement & Metrics

likes
0
downloads
0

Free2AITools Constitutional Data Pipeline: Curated disclosure mode active. (V15.x Standard)