NVIDIA logo
NVIDIA
Open Weights

Llama 3.1 Nemotron Instruct 70B

Released Oct 2024

Llama 3.1 Nemotron Instruct 70B is a large language model developed by NVIDIA, designed to enhance the helpfulness and instruction-following capabilities of generative AI. It is a derivative of Meta's Llama 3.1 70B architecture, further refined through NVIDIA's specialized alignment and fine-tuning pipelines to improve performance in general conversational tasks.\n\nThe model was trained using Reinforcement Learning from Human Feedback (RLHF), specifically the REINFORCE algorithm and Bradley-Terry model alignment. The process integrated NVIDIA’s Nemotron-4 340B Reward model and the HelpSteer2-Preference dataset, allowing the model to achieve state-of-the-art results on several automatic alignment benchmarks without requiring manual human annotations for the final training stage.\n\n## Performance and Capabilities\nAt its launch, the model reached top positions on the Arena Hard, AlpacaEval 2 LC, and MT-Bench leaderboards, often outperforming frontier models. It demonstrates high proficiency in complex reasoning, mathematical problem solving, and coding tasks. Notably, it is capable of correctly handling nuanced linguistic queries, such as character counting, which often challenge less aligned models.\n\n## Technical Architecture\nBuilt on the transformer architecture, the model features 70 billion parameters and employs Grouped Query Attention (GQA) for efficient inference. It maintains a context window of 128,000 tokens, enabling the processing of long documents and maintaining coherence over extended multi-turn dialogues.

Rankings & Comparison