Update README.md
Browse files
README.md
CHANGED
|
@@ -12,7 +12,7 @@ pipeline_tag: text-generation
|
|
| 12 |
|
| 13 |
Version 1.0 / 20.Feb.2023
|
| 14 |
|
| 15 |
-
This model is a joint collaboration between CKIP lab at Acedemia Sinica, MediaTek Research, and National Academy for Educational Research.
|
| 16 |
|
| 17 |
## Table of Contents
|
| 18 |
1. [Model Details](#model-details)
|
|
@@ -36,7 +36,7 @@ BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text dat
|
|
| 36 |
<details>
|
| 37 |
<summary>Click to expand</summary> <br/>
|
| 38 |
|
| 39 |
-
**Developed by:** MediaTek Research
|
| 40 |
|
| 41 |
**Model Type:** Transformer-based Language Model
|
| 42 |
|
|
@@ -44,9 +44,9 @@ BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text dat
|
|
| 44 |
|
| 45 |
**Languages:** Multiple; see [training data](#training-data)
|
| 46 |
|
| 47 |
-
**License:** MEDIATEK RESEARCH License ([link](https://huggingface.co/
|
| 48 |
|
| 49 |
-
**Release Date Estimate:**
|
| 50 |
|
| 51 |
**Send Questions to:** [email protected]
|
| 52 |
|
|
@@ -157,7 +157,7 @@ Please refer to [Model card](https://huggingface.co/bigscience/bloom-1b1#uses).
|
|
| 157 |
<details>
|
| 158 |
<summary>Click to expand</summary><br/>
|
| 159 |
|
| 160 |
-
We trained the 1B1 parameter model on a total of 6 Billion tokens mainly crawled from the internet and provided from National Academy for Educational Research
|
| 161 |
Details are provided in the [paper](https://arxiv.org/).
|
| 162 |
|
| 163 |
</details>
|
|
|
|
| 12 |
|
| 13 |
Version 1.0 / 20.Feb.2023
|
| 14 |
|
| 15 |
+
This model is a joint collaboration between CKIP lab at Acedemia Sinica ([website](https://ckip.iis.sinica.edu.tw/)), MediaTek Research ([website](https://www.mtkresearch.com/)), and National Academy for Educational Research ([website](https://www.naer.edu.tw/)).
|
| 16 |
|
| 17 |
## Table of Contents
|
| 18 |
1. [Model Details](#model-details)
|
|
|
|
| 36 |
<details>
|
| 37 |
<summary>Click to expand</summary> <br/>
|
| 38 |
|
| 39 |
+
**Developed by:** MediaTek Research
|
| 40 |
|
| 41 |
**Model Type:** Transformer-based Language Model
|
| 42 |
|
|
|
|
| 44 |
|
| 45 |
**Languages:** Multiple; see [training data](#training-data)
|
| 46 |
|
| 47 |
+
**License:** MEDIATEK RESEARCH License ([link](https://huggingface.co/ckip-joint/bloom-1b1-zh/blob/main/LICENSE_MR.md)) and RAIL License v1.0 ([link](https://huggingface.co/spaces/bigscience/license))
|
| 48 |
|
| 49 |
+
**Release Date Estimate:** Wednesday, 22.February.2023
|
| 50 |
|
| 51 |
**Send Questions to:** [email protected]
|
| 52 |
|
|
|
|
| 157 |
<details>
|
| 158 |
<summary>Click to expand</summary><br/>
|
| 159 |
|
| 160 |
+
We trained the 1B1 parameter model on a total of 6 Billion tokens mainly crawled from the internet and provided from National Academy for Educational Research. 75% of the training data is Traditional Chinese, 25% is English.
|
| 161 |
Details are provided in the [paper](https://arxiv.org/).
|
| 162 |
|
| 163 |
</details>
|