qyinyourmind commited on
Commit
d559ffd
·
verified ·
1 Parent(s): 33e0d0e

update readme

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -32,7 +32,7 @@ Visit our Hugging Face (click links above), search checkpoints with names starti
32
 
33
 
34
  The `dots.llm1` model is a large-scale MoE model that activates 14B parameters out of a total of 142B parameters, delivering performance on par with state-of-the-art models.
35
- Leveraging our meticulously crafted and efficient data processing pipeline, `dots.llm1` achieves performance comparable to Qwen2.5-72B after pretrained on 11.2T high-quality tokens without synthetic data. To foster further research, we open-source intermediate training checkpoints at every one trillion tokens, providing valuable insights into the learning dynamics of large language models.
36
 
37
 
38
  <p align="center">
@@ -43,7 +43,7 @@ Leveraging our meticulously crafted and efficient data processing pipeline, `dot
43
 
44
  **This repo contains the base and instruction-tuned `dots.llm1` model**. which has the following features:
45
 
46
- - Type: A MoE model with 14B activated and 142B total parameters trained on 11.2T tokens.
47
  - Training Stages: Pretraining and SFT.
48
  - Architecture: Multi-head Attention with QK-Norm in attention Layer, fine-grained MoE utilizing top-6 out of 128 routed experts, plus 2 shared experts.
49
  - Number of Layers: 62
@@ -55,10 +55,10 @@ Leveraging our meticulously crafted and efficient data processing pipeline, `dot
55
  The highlights from `dots.llm1` include:
56
 
57
  - **Enhanced Data Processing**: We propose a scalable and fine-grained *three-stage* data processing framework designed to generate large-scale, high-quality and diverse data for pretraining.
58
- - **No Synthetic Data during Pretraining**: *11.2 trillion* high-quality non-synthetic tokens was used in base model pretraining.
59
  - **Performance and Cost Efficiency**: `dots.llm1` is an open-source model that activates only *14B* parameters at inference, delivering both comprehensive capabilities and high computational efficiency.
60
  - **Infrastructure**: We introduce an innovative MoE all-to-all communication and computation overlapping recipe based on interleaved 1F1B pipeline scheduling and an efficient grouped GEMM implementation to boost computational efficiency.
61
- - **Open Accessibility to Model Dynamics**: Intermediate model checkpoints for *every 1T tokens* trained are released, facilitating future research into the learning dynamics of large language models.
62
 
63
  ## 3. Example Usage
64
 
 
32
 
33
 
34
  The `dots.llm1` model is a large-scale MoE model that activates 14B parameters out of a total of 142B parameters, delivering performance on par with state-of-the-art models.
35
+ Leveraging our meticulously crafted and efficient data processing pipeline, `dots.llm1` achieves performance comparable to Qwen2.5-72B after pretrained on high-quality corpus without synthetic data. To foster further research, we open-source intermediate training checkpoints spanning the entire training process, providing valuable insights into the learning dynamics of large language models.
36
 
37
 
38
  <p align="center">
 
43
 
44
  **This repo contains the base and instruction-tuned `dots.llm1` model**. which has the following features:
45
 
46
+ - Type: A MoE model with 14B activated and 142B total parameters trained on high-quality corpus.
47
  - Training Stages: Pretraining and SFT.
48
  - Architecture: Multi-head Attention with QK-Norm in attention Layer, fine-grained MoE utilizing top-6 out of 128 routed experts, plus 2 shared experts.
49
  - Number of Layers: 62
 
55
  The highlights from `dots.llm1` include:
56
 
57
  - **Enhanced Data Processing**: We propose a scalable and fine-grained *three-stage* data processing framework designed to generate large-scale, high-quality and diverse data for pretraining.
58
+ - **No Synthetic Data during Pretraining**: High-quality non-synthetic tokens was used in base model pretraining.
59
  - **Performance and Cost Efficiency**: `dots.llm1` is an open-source model that activates only *14B* parameters at inference, delivering both comprehensive capabilities and high computational efficiency.
60
  - **Infrastructure**: We introduce an innovative MoE all-to-all communication and computation overlapping recipe based on interleaved 1F1B pipeline scheduling and an efficient grouped GEMM implementation to boost computational efficiency.
61
+ - **Open Accessibility to Model Dynamics**: Intermediate model checkpoints are released spanning the entire training process, facilitating future research into the learning dynamics of large language models.
62
 
63
  ## 3. Example Usage
64