Update README.md
Browse files
README.md
CHANGED
|
@@ -36,4 +36,5 @@ inference:
|
|
| 36 |
## training
|
| 37 |
|
| 38 |
- for inputs, the model was presented with the post title and the post selftext encoded as: `question: <post title> context: <post selftext>`. You may see better results if queries are posed in this fashion.
|
| 39 |
-
- The top two replies were aggregated and presented to the model as the output text.
|
|
|
|
|
|
| 36 |
## training
|
| 37 |
|
| 38 |
- for inputs, the model was presented with the post title and the post selftext encoded as: `question: <post title> context: <post selftext>`. You may see better results if queries are posed in this fashion.
|
| 39 |
+
- The top two replies were aggregated and presented to the model as the output text.
|
| 40 |
+
- Training for longer will be explored, but given that the dataset has 127k examples and the loss flatlines at 0.5 epochs this should be fairly viable.
|