Numerical Analysis
Core Concepts
- Floating-point precision
- Numerical stability
- Gradient clipping
- Learning rate scheduling
Applications in Large Models
Preventing Gradient Explosion / Vanishing
- Large models are deep and computationally intensive, making numerical stability particularly critical.
BFloat16 / FP16 Training
- Understanding how different floating-point precisions affect model training.
Optimizer Selection
- Some optimizers are numerically more stable.
贡献者
这篇文章有帮助吗?
最近更新
Involution Hell© 2026 byCommunityunderCC BY-NC-SA 4.0
References
Explore curated references for linear algebra and calculus, including immersive texts, geometric PDFs, and 3Blue1Brown’s visual video series—ideal for CS/AI learners building intuition.
Probability and Statistics
Core concepts in probability and statistics and their applications in large models