LLM Reference

MT0 XXL

About

The MT0 XXL model, part of the BLOOMZ and mT0 family developed by the BigScience workshop, is a large language AI model designed for multilingual applications with zero-shot capabilities. It features 13.9 billion parameters and is built on a transformer architecture with an encoder-decoder structure. Trained with a multitask finetuning approach using the xP3 dataset, MT0 XXL excels in tasks such as translation, question answering, text generation, and summarization. Although powerful, the model's performance hinges on prompt specificity and the diversity of its training data. Running the model can be resource-intensive, which might limit its accessibility for smaller entities 1.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
IBM watsonx$1.8$1.8
Serverless

Specifications

FamilyMT0
Parameters13B
ArchitectureDecoder Only
Specializationgeneral