LLM Reference

abab 6.5s

About

The abab 6.5s is an advanced large language model developed by MiniMax, designed as a more efficient iteration of the abab 6.5 model. It leverages similar training techniques and data but excels in processing speed, handling approximately 30,000 words per second while supporting a context length of 200k tokens. The model's efficiency makes it ideal for tasks demanding rapid text processing. Innovatively, it features the Mixture of Experts (MoE) architecture, a first in China's domestic models, which underlines MiniMax's technological prowess. Performance tests reveal that abab 6.5s rivals top models like GPT-4, Claude-3, and Gemini-1.5 23478.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

Familyabab
Released2024-04-17
ArchitectureDecoder Only
Specializationgeneral