Commit
·
9130025
1
Parent(s):
058e8e2
Update README.md
Browse files
README.md
CHANGED
|
@@ -15,7 +15,7 @@ datasets:
|
|
| 15 |
|
| 16 |
## Model Description
|
| 17 |
|
| 18 |
-
GPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model.
|
| 19 |
|
| 20 |
## Training data
|
| 21 |
|
|
@@ -60,7 +60,7 @@ EleutherAI is currently in the process of carrying out further evaluations of GP
|
|
| 60 |
| GPT-3 1.3B | ------ | ----- | ----- |
|
| 61 |
| GPT-2 1.5B | 1.0468 | ----- | 17.48 |
|
| 62 |
| **GPT-Neo 2.7B** | **0.7165** | **5.646** | **11.39** |
|
| 63 |
-
| GPT-3
|
| 64 |
| GPT-3 175B | 0.7177 | ----- | ----- |
|
| 65 |
|
| 66 |
All GPT-2 and GPT-3 scores are from their respective papers, except for the Pile test results which are from the Pile paper.
|
|
|
|
| 15 |
|
| 16 |
## Model Description
|
| 17 |
|
| 18 |
+
GPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model.
|
| 19 |
|
| 20 |
## Training data
|
| 21 |
|
|
|
|
| 60 |
| GPT-3 1.3B | ------ | ----- | ----- |
|
| 61 |
| GPT-2 1.5B | 1.0468 | ----- | 17.48 |
|
| 62 |
| **GPT-Neo 2.7B** | **0.7165** | **5.646** | **11.39** |
|
| 63 |
+
| GPT-3 2.7B | 0.9631 | ----- | ----- |
|
| 64 |
| GPT-3 175B | 0.7177 | ----- | ----- |
|
| 65 |
|
| 66 |
All GPT-2 and GPT-3 scores are from their respective papers, except for the Pile test results which are from the Pile paper.
|