LLM Reference

Aquila 7B

About

The Aquila 7B model, developed by the Beijing Academy of Artificial Intelligence, stands out for its bilingual support catering to both Chinese and English languages. Building on the architectures of GPT-3 and LLaMA, it features efficient operator implementations and a novel bilingual tokenizer, significantly boosting training efficiency. The model leverages a well-curated data set, with a substantial portion being native Chinese texts, ensuring a deep cultural and linguistic understanding. Aquila 7B excels in a range of NLP tasks such as text generation and language understanding, and is open-sourced, with its use subject to commercial licensing agreements. Despite its robustness, it shares common limitations with large language models, such as language idiosyncrasies, and requires considerable computational resources. Further enhancing its versatility, variants like AquilaChat-7B and AquilaCode-7B have been developed, with future expansions planned.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

FamilyAquila
Parameters7B
ArchitectureDecoder Only
Specializationgeneral