LLM Reference
Chinchilla

Chinchilla

About

The Chinchilla family of large language models, developed by Google DeepMind, was introduced in March 2022. These models are notable for their exploration of the scaling laws in LLM training. Uniquely, they highlighted that for optimal model performance, the size of the model and the number of training tokens should be proportionately scaled. For instance, the Chinchilla model with 70 billion parameters used the same computational resources as a 280 billion parameter Gopher model but was trained on quadruple the data, leading to enhanced performance across numerous benchmarks. This approach challenged the previous assumption that increasing model size inherently improves performance, emphasizing the critical role of ample data in achieving state-of-the-art results 1)23.

Models(2)

Details

ResearcherGoogle DeepMind
Models2

Links

Website