LLM Reference
Nemotron-4

Nemotron-4

About

The Nemotron-4 340B family consists of large language models (LLMs) that are openly accessible and tailored for synthetic data generation, crucial for training other LLMs 34. This innovative suite includes a base model, instruct model, and reward model, each serving unique purposes. The base model, trained on an extensive 9 trillion token dataset, supports the instruct model in producing diverse synthetic data that emulates real-world scenarios, while the reward model focuses on evaluating and refining outputs for helpfulness and coherence 12. Optimized for NVIDIA's NeMo framework and TensorRT-LLM for inference, these models are designed for both research and commercial use due to their open licensing. Moreover, the fully open-sourced pipeline encourages AI community collaboration and innovation 9.

Models(1)

Details

ResearcherNVIDIA AI
Models1