--- library_name: transformers license: other license_name: nvidia-open-model-license license_link: >- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation language: - en tags: - nvidia - Nemotron-Cascade - reasoning - general-purpose - SFT - RL - pytorch --- # Nemotron-Cascade-8B Intermediate ckpts
[](https://arxiv.org/abs/2512.13607) [](https://huggingface.co/collections/nvidia/nemotron-cascade) [](https://huggingface.co/collections/nvidia/nemotron-cascade) [](https://huggingface.co/collections/nvidia/nemotron-cascade)
## Introduction This repository releases the intermediate checkpoints produced during the development of [Nemotron-Cascade-8B](https://huggingface.co/nvidia/Nemotron-Cascade-8B). Nemotron-Cascade-8B is a general-purpose model trained using a sequential, domain-wise reinforcement learning pipeline, illustrated in the figure below.
We release checkpoints corresponding to each major stage of training:
- **Nemotron-Cascade-8B-SFT** (completed multi-stage SFT)
- **Nemotron-Cascade-8B-RLHF** (completed RLHF)
- **Nemotron-Cascade-8B-IFRL** (completed instruction following RL)
- **Nemotron-Cascade-8B-MathRL** (completed Math RL)
- **Nemotron-Cascade-8B-CodeRL** (completed Code RL)
The final model, [Nemotron-Cascade-8B](https://huggingface.co/nvidia/Nemotron-Cascade-8B), is obtained after the concluding SWE RL stage.
## Usage Recommendations
We recommend using RoPE scaling with the [YaRN](https://arxiv.org/abs/2309.00071) method to better support contexts longer than 32K. This can be enabled by updating the model’s `config.json` as shown below:
```json
{
...,
"rope_scaling": {
"rope_type": "yarn",
"factor": 2.0,
"original_max_position_embeddings": 32768
}
}
```
## Results
Same as [Nemotron-Cascade-8B](https://huggingface.co/nvidia/Nemotron-Cascade-8B), we use a maximum output length of 64K tokens for evaluation, with the temperature set to 0.6 and top-p to 0.95. We also apply RoPE scaling using the YaRN method with a scaling factor of 2.0.
| **Benchmark