Sharath Turuvekere Sreenivas
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -138,6 +138,8 @@ The integration of foundation and fine-tuned models into AI systems requires add
|
|
138 |
|
139 |
NVIDIA-Nemotron-Nano-12B-v2-Base is pre-trained on a large corpus of high-quality curated and synthetically-generated data. It is trained in the English language, as well as 15 multilingual languages and 43 programming languages. Our sources cover a variety of document types such as: webpages, dialogue, articles, and other written materials. The corpus spans domains including legal, math, science, finance, and more. We also include a small portion of question-answering, and alignment style data to improve model accuracy. The model was trained for approximately twenty trillion tokens.
|
140 |
|
|
|
|
|
141 |
**Data Modality:** Text **The total size:** 10,648,823,153,919 Tokens **Total number of datasets:** 141 **Dataset partition:** *Training \[100%\], testing \[0%\], validation \[0%\]*
|
142 |
**Time period for training data collection:** 2013 to May 1, 2025
|
143 |
**Time period for testing data collection:** 2013 to May 1, 2025
|
|
|
138 |
|
139 |
NVIDIA-Nemotron-Nano-12B-v2-Base is pre-trained on a large corpus of high-quality curated and synthetically-generated data. It is trained in the English language, as well as 15 multilingual languages and 43 programming languages. Our sources cover a variety of document types such as: webpages, dialogue, articles, and other written materials. The corpus spans domains including legal, math, science, finance, and more. We also include a small portion of question-answering, and alignment style data to improve model accuracy. The model was trained for approximately twenty trillion tokens.
|
140 |
|
141 |
+
Alongside the model, we release our final pretraining data, as outlined in this section. For ease of analysis, there is a sample set that is ungated. For all remaining code, math and multilingual data, gating and approval is required, and the dataset is permissively licensed for model training purposes
|
142 |
+
|
143 |
**Data Modality:** Text **The total size:** 10,648,823,153,919 Tokens **Total number of datasets:** 141 **Dataset partition:** *Training \[100%\], testing \[0%\], validation \[0%\]*
|
144 |
**Time period for training data collection:** 2013 to May 1, 2025
|
145 |
**Time period for testing data collection:** 2013 to May 1, 2025
|