Update README.md
Browse files
README.md
CHANGED
|
@@ -91,7 +91,10 @@ The model was pre-trained continuously on a single A10G GPU in an AWS instance f
|
|
| 91 |
|
| 92 |
#### Possible Future Directions:
|
| 93 |
|
| 94 |
-
1. Use a decoder
|
|
|
|
|
|
|
|
|
|
| 95 |
|
| 96 |
#### Authors:
|
| 97 |
|
|
|
|
| 91 |
|
| 92 |
#### Possible Future Directions:
|
| 93 |
|
| 94 |
+
1. Use a decoder-only model for pre-training and summarization.
|
| 95 |
+
<br>As it seems the case when the span deleting tokens is not very large, the model learns to copy the token from the encoder to decoder generation.
|
| 96 |
+
<br>Thus, hurts the performance of the Abstractive Summarization task.
|
| 97 |
+
<br>This case is not present in the decoder-only model as all the predicted next token is not seen by the model at all.
|
| 98 |
|
| 99 |
#### Authors:
|
| 100 |
|