niltheory's picture
Update README.md
04f8ee4
|
raw
history blame
498 Bytes
---
datasets:
- niltheory/ExistenceTypes
language:
- en
---
Iteration #1:
base distilBert used to train the model with a data training of 96.
Iteration #2:
Increased the data training from 96 to 296.
Which greatly increased the accuracy scores, albeit there is some edge cases.
Iteration #3:
Switched from bert-base-uncased to bert-large-cased-whole-word-masking.
Which it has more contextual sensitivity and is slightly more accurate than Iteration #2.
It's overall more nuanced and sensitive.