LLM ReferenceLLM Reference
2 models2024Up to 128K ctxFrom $0.02/1M input

About

The NeMo family of large language models (LLMs) is a collaborative effort between NVIDIA and Mistral AI, recognized for their superior performance compared to similar-sized models 234. These models are distinguished by their substantial context window, capable of processing up to 128,000 tokens, which enables them to manage extensive documents and intricate conversations 678. Under the Apache 2.0 license, they promote open development and accessibility within the AI community 235. The NeMo family encompasses various models with differing parameter counts, such as 12 billion, with specialized fine-tuning for tasks like instruction following 23. They are tailored for multilingual applications and are adept in code generation 35. The NeMo framework itself offers tools for model customization and deployment, enabling adaptations for diversified applications 29.

Specifications(2 models)

NeMo model specifications comparison
ModelReleasedContextParameters
Mistral NeMo Instruct (2407)2024-07128K12B
Mistral NeMo (2407)2024-07128K12B

Available From(11 providers)

Pricing

NeMo model pricing by provider
ModelProviderInput / 1MOutput / 1MType
Mistral NeMo (2407)OpenRouter$0.02$0.04Serverless
Mistral NeMo Instruct (2407)Novita AI$0.08$0.24Serverless
Mistral NeMo (2407)Mistral AI Studio$0.15$0.15Serverless
Mistral NeMo Instruct (2407)Arcee AI$0.15$0.45Serverless
Mistral NeMo (2407)Bitdeer AI$0.18$0.54Serverless
Mistral NeMo (2407)Fireworks AI$0.2$0.2Serverless
Mistral NeMo Instruct (2407)Microsoft Foundry$0.3$0.3Provisioned
Mistral NeMo (2407)SiliconFlow$0.3$0.3Serverless
Mistral NeMo Instruct (2407)Replicate API$0.45$0.45Serverless
Mistral NeMo Instruct (2407)Fireworks AI$0.9$0.9Serverless
Mistral NeMo Instruct (2407)DeepInfra$2$4Serverless

Frequently Asked Questions

What is NeMo?
The NeMo family of large language models (LLMs) is a collaborative effort between NVIDIA and Mistral AI, recognized for their superior performance compared to similar-sized models 234. These models are distinguished by their substantial context window, capable of processing up to 128,000 tokens, which enables them to manage extensive documents and intricate conversations 678. Under the Apache 2.0 license, they promote open development and accessibility within the AI community 235. The NeMo family encompasses various models with differing parameter counts, such as 12 billion, with specialized fine-tuning for tasks like instruction following 23. They are tailored for multilingual applications and are adept in code generation 35. The NeMo framework itself offers tools for model customization and deployment, enabling adaptations for diversified applications 29.
How many models are in the NeMo family?
The NeMo family contains 2 models.
What is the latest NeMo model?
The latest model is Mistral NeMo Instruct (2407), released in 2024-07.
How much does NeMo cost?
NeMo models range from $0.02/1M to $2/1M input tokens depending on the model and provider.

Models(2)