LLM Reference

Nous Hermes 2 Mixtral 8x7B

About

Nous Hermes 2 Mixtral 8x7B is a sophisticated language model built on the Mixtral 8x7B MoE architecture, designed for high-efficiency applications and rapid response in real-time settings. It excels in content generation and customer service chatbot functionalities, thanks to its training on over 1,000,000 entries, predominantly using GPT-4 generated content and other high-quality datasets. The model supports structured multi-turn interactions via ChatML, enhancing its compatibility with OpenAI API. It is available in various quantized formats including GGUF, GPTQ, and AWQ, enabling versatile deployment across different systems. While it performs well in benchmarks like GPT4All and AGIEval, its accuracy can fluctuate based on training data quality, specific tasks, or topics, and deploying its unquantized versions demands significant computational resources. Potential biases may also arise from the GPT-4 data used during training.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Providers(3)

ProviderInput (per 1M)Output (per 1M)Type
OctoAI API$0.15$0.15
Serverless
Fireworks AI Platform
Provisioned
Together AI API$0.6$0.6
Serverless

Specifications

FamilyHermes 2
Parameters8x7B
ArchitectureMixture of Experts
Specializationgeneral