LLM Reference

T5 Large

About

The T5 Large model, developed by Google AI, is a robust encoder-decoder model built on the Transformer architecture with 770 million parameters. It uses a unique text-to-text framework, handling all NLP tasks by transforming inputs and outputs into text strings, which simplifies task versatility and deployment. Pre-trained on the C4 dataset, it can tackle tasks like translation, summarization, and question answering after fine-tuning. Despite its capabilities, it demands significant computational resources and its performance is influenced by training data quality and inherent biases.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

FamilyT5
Parameters770M
ArchitectureDecoder Only
Specializationgeneral