llm-wizard commited on
Commit
67d030b
·
verified ·
1 Parent(s): cda2b16

Add References to Data Designer.

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -187,8 +187,8 @@ Stage 1: Pre-Training
187
 
188
  Stage 2: Supervised Fine-Tuning
189
 
190
- * The model was further fine-tuned on synthetic code, math, science, tool calling, instruction following, structured outputs, and general knowledge data. All datasets are disclosed in the [Training, Testing, and Evaluation Datasets](https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16#training-testing-and-evaluation-datasets) section of this document. Major portions of the fine-tuning corpus are released in the [Nemotron-Post-Training-v3](https://huggingface.co/collections/nvidia/nemotron-post-training-v3) collection.
191
- * Software used for supervised fine-tuning: [Megatron-LM](https://github.com/NVIDIA/Megatron-LM)
192
 
193
  Stage 3: Reinforcement Learning
194
 
@@ -197,7 +197,7 @@ Stage 3: Reinforcement Learning
197
 
198
  NVIDIA-Nemotron-3-Nano-30B-A3B-BF16 model is a result of the above work.
199
 
200
- The end-to-end training recipe is available in the [NVIDIA Nemotron Developer Repository](https://github.com/NVIDIA-NeMo/Nemotron). Evaluation results can be replicated using the [NeMo Evaluator SDK](https://github.com/NVIDIA-NeMo/Evaluator). More details on the datasets and synthetic data generation methods can be found in the technical report [NVIDIA Nemotron 3 Nano](https://research.nvidia.com/labs/nemotron/files/NVIDIA-Nemotron-3-Nano-Technical-Report.pdf).
201
 
202
  ## Input
203
 
 
187
 
188
  Stage 2: Supervised Fine-Tuning
189
 
190
+ * The model was further fine-tuned on synthetic code, math, science, tool calling, instruction following, structured outputs, and general knowledge data. All datasets are disclosed in the [Training, Testing, and Evaluation Datasets](https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16#training-testing-and-evaluation-datasets) section of this document. Major portions of the fine-tuning corpus are released in the [Nemotron-Post-Training-v3](https://huggingface.co/collections/nvidia/nemotron-post-training-v3) collection. Data Designer is one of the libraries used to prepare these corpora.
191
+ *
192
 
193
  Stage 3: Reinforcement Learning
194
 
 
197
 
198
  NVIDIA-Nemotron-3-Nano-30B-A3B-BF16 model is a result of the above work.
199
 
200
+ The end-to-end training recipe is available in the [NVIDIA Nemotron Developer Repository](https://github.com/NVIDIA-NeMo/Nemotron). Evaluation results can be replicated using the [NeMo Evaluator SDK](https://github.com/NVIDIA-NeMo/Evaluator). Data Designer is one of the libraries used to prepare the pre and post training datasets. More details on the datasets and synthetic data generation methods can be found in the technical report [NVIDIA Nemotron 3 Nano](https://research.nvidia.com/labs/nemotron/files/NVIDIA-Nemotron-3-Nano-Technical-Report.pdf).
201
 
202
  ## Input
203