LLM ReferenceLLM Reference

Chinchilla 70B

About

Chinchilla 70B is a large language model from Google DeepMind, launched in March 2022. It adopts a compute-optimal approach, emphasizing a balance between model size and training data quantity, contrasting previous trends that prioritized model size alone. The model's architecture is based on transformers and utilizes innovations like RMSNorm and relative positional encoding. Trained with the MassiveText dataset using 1.4 trillion tokens, Chinchilla delivers superior performance compared to other larger models like GPT-3. While it demonstrates efficiency in tasks such as reading comprehension and common sense reasoning, limitations include high training costs and the potential for biased outputs. Additionally, it remains inaccessible for public use, restricting broader experimentation.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Rankings

Specifications

Released2022-03-29
Parameters70B
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuning

Created by

Pioneering artificial intelligence research.

London, United Kingdom
Founded 2014
Website