Nous Hermes 2 Mixtral 8x7B
About
Nous Hermes 2 Mixtral 8x7B is a sophisticated language model built on the Mixtral 8x7B MoE architecture, designed for high-efficiency applications and rapid response in real-time settings. It excels in content generation and customer service chatbot functionalities, thanks to its training on over 1,000,000 entries, predominantly using GPT-4 generated content and other high-quality datasets. The model supports structured multi-turn interactions via ChatML, enhancing its compatibility with OpenAI API. It is available in various quantized formats including GGUF, GPTQ, and AWQ, enabling versatile deployment across different systems. While it performs well in benchmarks like GPT4All and AGIEval, its accuracy can fluctuate based on training data quality, specific tasks, or topics, and deploying its unquantized versions demands significant computational resources. Potential biases may also arise from the GPT-4 data used during training.
Capabilities
Providers(3)
| Provider | Input (per 1M) | Output (per 1M) | Type | |
|---|---|---|---|---|
| OctoAI API | $0.15 | $0.15 | Serverless | |
| Fireworks AI Platform | — | — | Provisioned | |
| Together AI API | $0.6 | $0.6 | Serverless |