loubnabnl HF Staff commited on
Commit
5835396
·
verified ·
1 Parent(s): 4dafe82

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -36,6 +36,8 @@ SmolLM3 is a 3B parameter language model designed to push the boundaries of smal
36
 
37
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/Zcm_016pWeyFr_uIkT7Ki.png)
38
 
 
 
39
  The model is a decoder-only transformer using GQA and NoPE, it was pretrained on 11.2T tokens with a staged curriculum of web, code, math and reasoning data. Post-training included midtraining on 140B reasoning tokens followed by supervised fine-tuning and alignment via Anchored Preference Optimization (APO).
40
 
41
  ### Key features
 
36
 
37
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/Zcm_016pWeyFr_uIkT7Ki.png)
38
 
39
+ **SmolLM3-3B-Base** is the base model after pretraining, you can find the instruct model at [SmolLM3-3B](https://huggingface.co/HuggingFaceTB/SmolLM3-3B).
40
+
41
  The model is a decoder-only transformer using GQA and NoPE, it was pretrained on 11.2T tokens with a staged curriculum of web, code, math and reasoning data. Post-training included midtraining on 140B reasoning tokens followed by supervised fine-tuning and alignment via Anchored Preference Optimization (APO).
42
 
43
  ### Key features