versae commited on
Commit
4619063
·
1 Parent(s): b270287

Step... (21000/50000 | Loss: 1.669716238975525, Acc: 0.6647850275039673): 44%|████████████▏ | 21775/50000 [8:35:31<11:59:11, 1.53s/it]

Browse files
Files changed (25) hide show
  1. flax_model.msgpack +1 -1
  2. outputs/checkpoints/checkpoint-15000/training_state.json +0 -1
  3. outputs/checkpoints/checkpoint-16000/training_state.json +0 -1
  4. outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/config.json +0 -0
  5. outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/data_collator.joblib +0 -0
  6. outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/flax_model.msgpack +1 -1
  7. outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/optimizer_state.msgpack +1 -1
  8. outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/training_args.joblib +0 -0
  9. outputs/checkpoints/checkpoint-20000/training_state.json +1 -0
  10. outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/config.json +0 -0
  11. outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/data_collator.joblib +0 -0
  12. outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/flax_model.msgpack +1 -1
  13. outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/optimizer_state.msgpack +1 -1
  14. outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/training_args.joblib +0 -0
  15. outputs/checkpoints/checkpoint-21000/training_state.json +1 -0
  16. outputs/events.out.tfevents.1627258355.tablespoon.3000110.3.v2 +2 -2
  17. outputs/flax_model.msgpack +1 -1
  18. outputs/optimizer_state.msgpack +1 -1
  19. outputs/training_state.json +1 -1
  20. pytorch_model.bin +1 -1
  21. run_stream.512.log +0 -0
  22. wandb/run-20210726_001233-17u6inbn/files/output.log +1720 -0
  23. wandb/run-20210726_001233-17u6inbn/files/wandb-summary.json +1 -1
  24. wandb/run-20210726_001233-17u6inbn/logs/debug-internal.log +2 -2
  25. wandb/run-20210726_001233-17u6inbn/run-17u6inbn.wandb +2 -2
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d0ae4178820ed8ec84d010dda13f1c110189fa19d49afd4d14283cf09774bee
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba1daf7b1dad5bf7c386bc7b53d5537a8f26b3cfee5b0fc009a750ad077eab0
3
  size 249750019
outputs/checkpoints/checkpoint-15000/training_state.json DELETED
@@ -1 +0,0 @@
1
- {"step": 15001}
 
 
outputs/checkpoints/checkpoint-16000/training_state.json DELETED
@@ -1 +0,0 @@
1
- {"step": 16001}
 
 
outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/config.json RENAMED
File without changes
outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/data_collator.joblib RENAMED
File without changes
outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/flax_model.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4fae57c94dac9999e94e22d5044434007f1f445ea6c1ed7eba59fea02de5db30
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0bf6ec68f574b4af0a99a55d1d45aa7b0e1e57289109e4d7f5c9e5d83816424
3
  size 249750019
outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/optimizer_state.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5e466347b293d39a54146314ea173df2523a33930d651d708c67ac132dd5088
3
  size 499500278
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cab56f5cbba1f1be7c19b45ad56f5c58ced35d8e3f342687638a6a00d3e1d292
3
  size 499500278
outputs/checkpoints/{checkpoint-15000 → checkpoint-20000}/training_args.joblib RENAMED
File without changes
outputs/checkpoints/checkpoint-20000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"step": 20001}
outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/config.json RENAMED
File without changes
outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/data_collator.joblib RENAMED
File without changes
outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/flax_model.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1c97577a048157d950d83272279ccc645289bcda0b3c1a4a714df7b8ef89ffa
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba1daf7b1dad5bf7c386bc7b53d5537a8f26b3cfee5b0fc009a750ad077eab0
3
  size 249750019
outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/optimizer_state.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:33826e22d38c2cec4b568afa016f47cb58598d04aaa49474922cfec9ede1b96b
3
  size 499500278
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd862c6893d8672a836d674b5ef9d3eaab357c385ad5b064b7202eccc581ff05
3
  size 499500278
outputs/checkpoints/{checkpoint-16000 → checkpoint-21000}/training_args.joblib RENAMED
File without changes
outputs/checkpoints/checkpoint-21000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"step": 21001}
outputs/events.out.tfevents.1627258355.tablespoon.3000110.3.v2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d64128316db1be6b780146438299ca6d9e56a42d407935030bc80dd716161f5
3
- size 2803455
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3cc46840b5336c96adfc10b39ed6dd9d36d3759fb574ca64e28191207730bfb
3
+ size 3176589
outputs/flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d0ae4178820ed8ec84d010dda13f1c110189fa19d49afd4d14283cf09774bee
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba1daf7b1dad5bf7c386bc7b53d5537a8f26b3cfee5b0fc009a750ad077eab0
3
  size 249750019
outputs/optimizer_state.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fd17bbca5658a6226151a6f85c1c6b4064b42b9ce32213f96be1f4b4993a48c
3
  size 499500278
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd862c6893d8672a836d674b5ef9d3eaab357c385ad5b064b7202eccc581ff05
3
  size 499500278
outputs/training_state.json CHANGED
@@ -1 +1 @@
1
- {"step": 19001}
 
1
+ {"step": 21001}
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e1c987c78ed3890e2c466fea8b433708a3e3c70c0d1d02e1b0cdab38c7ab56f
3
  size 498858859
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56ddc0bfdddad2ca72308b9edd1fc42a1a815c78826b2a838c898083e3d5041e
3
  size 498858859
run_stream.512.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/run-20210726_001233-17u6inbn/files/output.log CHANGED
@@ -12910,6 +12910,1726 @@ You should probably TRAIN this model on a down-stream task to be able to use it
12910
 
12911
 
12912
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12913
 
12914
 
12915
 
 
12910
 
12911
 
12912
 
12913
+
12914
+
12915
+
12916
+
12917
+
12918
+
12919
+
12920
+
12921
+
12922
+
12923
+
12924
+
12925
+
12926
+
12927
+
12928
+
12929
+
12930
+
12931
+
12932
+
12933
+
12934
+
12935
+
12936
+
12937
+
12938
+
12939
+
12940
+
12941
+
12942
+
12943
+
12944
+
12945
+
12946
+
12947
+
12948
+
12949
+
12950
+
12951
+
12952
+
12953
+
12954
+
12955
+
12956
+
12957
+
12958
+
12959
+
12960
+
12961
+
12962
+
12963
+
12964
+
12965
+
12966
+
12967
+
12968
+
12969
+
12970
+
12971
+
12972
+
12973
+
12974
+
12975
+
12976
+
12977
+
12978
+
12979
+
12980
+
12981
+
12982
+
12983
+
12984
+
12985
+
12986
+
12987
+
12988
+
12989
+
12990
+
12991
+
12992
+
12993
+
12994
+
12995
+
12996
+
12997
+
12998
+
12999
+
13000
+
13001
+
13002
+
13003
+
13004
+
13005
+
13006
+
13007
+
13008
+
13009
+
13010
+
13011
+
13012
+
13013
+
13014
+
13015
+
13016
+
13017
+
13018
+
13019
+
13020
+
13021
+
13022
+
13023
+
13024
+
13025
+
13026
+
13027
+
13028
+
13029
+
13030
+
13031
+
13032
+
13033
+
13034
+
13035
+
13036
+
13037
+
13038
+
13039
+
13040
+
13041
+
13042
+
13043
+
13044
+
13045
+
13046
+
13047
+
13048
+
13049
+
13050
+
13051
+
13052
+
13053
+
13054
+
13055
+
13056
+
13057
+
13058
+
13059
+
13060
+
13061
+
13062
+
13063
+
13064
+
13065
+
13066
+
13067
+
13068
+
13069
+
13070
+
13071
+
13072
+
13073
+
13074
+
13075
+
13076
+
13077
+
13078
+
13079
+
13080
+
13081
+
13082
+
13083
+
13084
+
13085
+
13086
+
13087
+
13088
+
13089
+
13090
+
13091
+
13092
+
13093
+
13094
+
13095
+
13096
+
13097
+
13098
+
13099
+
13100
+
13101
+
13102
+
13103
+
13104
+
13105
+
13106
+
13107
+
13108
+
13109
+
13110
+
13111
+
13112
+
13113
+
13114
+
13115
+
13116
+
13117
+
13118
+
13119
+
13120
+
13121
+
13122
+
13123
+
13124
+
13125
+
13126
+
13127
+
13128
+
13129
+
13130
+
13131
+
13132
+
13133
+
13134
+
13135
+
13136
+
13137
+
13138
+
13139
+
13140
+
13141
+
13142
+
13143
+
13144
+
13145
+
13146
+
13147
+
13148
+
13149
+
13150
+
13151
+
13152
+
13153
+
13154
+
13155
+
13156
+
13157
+
13158
+
13159
+
13160
+
13161
+
13162
+
13163
+
13164
+
13165
+
13166
+
13167
+
13168
+
13169
+
13170
+
13171
+
13172
+
13173
+
13174
+
13175
+
13176
+
13177
+
13178
+
13179
+
13180
+
13181
+
13182
+
13183
+
13184
+
13185
+
13186
+
13187
+
13188
+
13189
+
13190
+
13191
+
13192
+
13193
+
13194
+
13195
+
13196
+
13197
+
13198
+
13199
+
13200
+
13201
+
13202
+
13203
+
13204
+
13205
+
13206
+
13207
+
13208
+
13209
+
13210
+
13211
+
13212
+
13213
+
13214
+
13215
+
13216
+
13217
+
13218
+
13219
+
13220
+
13221
+
13222
+
13223
+
13224
+
13225
+
13226
+
13227
+
13228
+
13229
+
13230
+
13231
+
13232
+
13233
+
13234
+
13235
+
13236
+
13237
+
13238
+
13239
+
13240
+
13241
+
13242
+
13243
+
13244
+
13245
+
13246
+
13247
+
13248
+
13249
+
13250
+
13251
+
13252
+
13253
+
13254
+
13255
+
13256
+
13257
+
13258
+
13259
+
13260
+
13261
+
13262
+
13263
+
13264
+
13265
+
13266
+
13267
+
13268
+
13269
+
13270
+
13271
+
13272
+
13273
+
13274
+
13275
+
13276
+
13277
+
13278
+
13279
+
13280
+
13281
+
13282
+
13283
+
13284
+
13285
+
13286
+
13287
+
13288
+
13289
+
13290
+
13291
+
13292
+
13293
+
13294
+
13295
+
13296
+
13297
+
13298
+
13299
+
13300
+
13301
+
13302
+
13303
+
13304
+
13305
+
13306
+
13307
+
13308
+
13309
+
13310
+
13311
+
13312
+
13313
+
13314
+
13315
+
13316
+
13317
+
13318
+
13319
+
13320
+
13321
+
13322
+
13323
+
13324
+
13325
+
13326
+
13327
+
13328
+
13329
+
13330
+
13331
+
13332
+
13333
+
13334
+
13335
+
13336
+
13337
+
13338
+
13339
+
13340
+
13341
+
13342
+
13343
+
13344
+
13345
+
13346
+
13347
+
13348
+
13349
+
13350
+
13351
+
13352
+
13353
+
13354
+
13355
+
13356
+
13357
+
13358
+
13359
+
13360
+
13361
+
13362
+
13363
+
13364
+
13365
+
13366
+
13367
+
13368
+
13369
+
13370
+
13371
+
13372
+
13373
+
13374
+
13375
+
13376
+
13377
+
13378
+
13379
+
13380
+
13381
+
13382
+
13383
+
13384
+
13385
+
13386
+
13387
+
13388
+
13389
+
13390
+
13391
+
13392
+
13393
+
13394
+
13395
+
13396
+
13397
+
13398
+
13399
+
13400
+
13401
+
13402
+
13403
+
13404
+
13405
+
13406
+
13407
+
13408
+
13409
+
13410
+
13411
+
13412
+
13413
+ Step... (19000/50000 | Loss: 1.6823453903198242, Acc: 0.6625654101371765): 40%|██████████▊ | 20000/50000 [7:51:27<11:12:26, 1.34s/it]
13414
+ Step... (19500 | Loss: 1.7442021369934082, Learning Rate: 0.0003696969652082771)
13415
+ Step... (20000 | Loss: 1.6871428489685059, Learning Rate: 0.0003636363835539669)
13416
+
13417
+
13418
+
13419
+
13420
+
13421
+
13422
+
13423
+
13424
+
13425
+
13426
+
13427
+ [09:59:50] - INFO - __main__ - Saving checkpoint at 20000 steps█████████████████████████████████████████████████████| 130/130 [00:21<00:00, 4.60it/s]
13428
+ All Flax model weights were used when initializing RobertaForMaskedLM.
13429
+ Some weights of RobertaForMaskedLM were not initialized from the Flax model and are newly initialized: ['lm_head.decoder.weight', 'roberta.embeddings.position_ids', 'lm_head.decoder.bias']
13430
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
13431
+
13432
+
13433
+
13434
+
13435
+
13436
+
13437
+
13438
+
13439
+
13440
+
13441
+
13442
+
13443
+
13444
+
13445
+
13446
+
13447
+
13448
+
13449
+
13450
+
13451
+
13452
+
13453
+
13454
+
13455
+
13456
+
13457
+
13458
+
13459
+
13460
+
13461
+
13462
+
13463
+
13464
+
13465
+
13466
+
13467
+
13468
+
13469
+
13470
+
13471
+
13472
+
13473
+
13474
+
13475
+
13476
+
13477
+
13478
+
13479
+
13480
+
13481
+
13482
+
13483
+
13484
+
13485
+
13486
+
13487
+
13488
+
13489
+
13490
+
13491
+
13492
+
13493
+
13494
+
13495
+
13496
+
13497
+
13498
+
13499
+
13500
+
13501
+
13502
+
13503
+
13504
+
13505
+
13506
+
13507
+
13508
+
13509
+
13510
+
13511
+
13512
+
13513
+
13514
+
13515
+
13516
+
13517
+
13518
+
13519
+
13520
+
13521
+
13522
+
13523
+
13524
+
13525
+
13526
+
13527
+
13528
+
13529
+
13530
+
13531
+
13532
+
13533
+
13534
+
13535
+
13536
+
13537
+
13538
+
13539
+
13540
+
13541
+
13542
+
13543
+
13544
+
13545
+
13546
+
13547
+
13548
+
13549
+
13550
+
13551
+
13552
+
13553
+
13554
+
13555
+
13556
+
13557
+
13558
+
13559
+
13560
+
13561
+
13562
+
13563
+
13564
+
13565
+
13566
+
13567
+
13568
+
13569
+
13570
+
13571
+
13572
+
13573
+
13574
+
13575
+
13576
+
13577
+
13578
+
13579
+
13580
+
13581
+
13582
+
13583
+
13584
+
13585
+
13586
+
13587
+
13588
+
13589
+
13590
+
13591
+
13592
+
13593
+
13594
+
13595
+
13596
+
13597
+
13598
+
13599
+
13600
+
13601
+
13602
+
13603
+
13604
+
13605
+
13606
+
13607
+
13608
+
13609
+
13610
+
13611
+
13612
+
13613
+
13614
+
13615
+
13616
+
13617
+
13618
+
13619
+
13620
+
13621
+
13622
+
13623
+
13624
+
13625
+
13626
+
13627
+
13628
+
13629
+
13630
+
13631
+
13632
+
13633
+
13634
+
13635
+
13636
+
13637
+
13638
+
13639
+
13640
+
13641
+
13642
+
13643
+
13644
+
13645
+
13646
+
13647
+
13648
+
13649
+
13650
+
13651
+
13652
+
13653
+
13654
+
13655
+
13656
+
13657
+
13658
+
13659
+
13660
+
13661
+
13662
+
13663
+
13664
+
13665
+
13666
+
13667
+
13668
+
13669
+
13670
+
13671
+
13672
+
13673
+
13674
+
13675
+
13676
+
13677
+
13678
+
13679
+
13680
+
13681
+
13682
+
13683
+
13684
+
13685
+
13686
+
13687
+
13688
+
13689
+
13690
+
13691
+
13692
+
13693
+
13694
+
13695
+
13696
+
13697
+
13698
+
13699
+
13700
+
13701
+
13702
+
13703
+
13704
+
13705
+
13706
+
13707
+
13708
+
13709
+
13710
+
13711
+
13712
+
13713
+
13714
+
13715
+
13716
+
13717
+
13718
+
13719
+
13720
+
13721
+
13722
+
13723
+
13724
+
13725
+
13726
+
13727
+
13728
+
13729
+
13730
+
13731
+
13732
+
13733
+
13734
+
13735
+
13736
+
13737
+
13738
+
13739
+
13740
+
13741
+
13742
+
13743
+
13744
+
13745
+
13746
+
13747
+
13748
+
13749
+
13750
+
13751
+
13752
+
13753
+
13754
+
13755
+
13756
+
13757
+
13758
+
13759
+
13760
+
13761
+
13762
+
13763
+
13764
+
13765
+
13766
+
13767
+
13768
+
13769
+
13770
+
13771
+
13772
+
13773
+
13774
+
13775
+
13776
+
13777
+
13778
+
13779
+
13780
+
13781
+
13782
+
13783
+
13784
+
13785
+
13786
+
13787
+
13788
+
13789
+
13790
+
13791
+
13792
+
13793
+
13794
+
13795
+
13796
+
13797
+
13798
+
13799
+
13800
+
13801
+
13802
+
13803
+
13804
+
13805
+
13806
+
13807
+
13808
+
13809
+
13810
+
13811
+
13812
+
13813
+
13814
+
13815
+
13816
+
13817
+
13818
+
13819
+
13820
+
13821
+
13822
+
13823
+
13824
+
13825
+
13826
+
13827
+
13828
+
13829
+
13830
+
13831
+
13832
+
13833
+
13834
+
13835
+
13836
+
13837
+
13838
+
13839
+
13840
+
13841
+
13842
+
13843
+
13844
+
13845
+
13846
+
13847
+
13848
+
13849
+
13850
+
13851
+
13852
+
13853
+
13854
+
13855
+
13856
+
13857
+
13858
+
13859
+
13860
+
13861
+
13862
+
13863
+
13864
+
13865
+
13866
+
13867
+
13868
+
13869
+
13870
+
13871
+
13872
+
13873
+
13874
+
13875
+
13876
+
13877
+
13878
+
13879
+
13880
+
13881
+
13882
+
13883
+
13884
+
13885
+
13886
+
13887
+
13888
+
13889
+
13890
+
13891
+
13892
+
13893
+
13894
+
13895
+
13896
+
13897
+
13898
+
13899
+
13900
+
13901
+
13902
+
13903
+
13904
+
13905
+
13906
+
13907
+
13908
+
13909
+
13910
+
13911
+
13912
+
13913
+
13914
+
13915
+
13916
+
13917
+
13918
+
13919
+
13920
+
13921
+
13922
+
13923
+
13924
+
13925
+
13926
+
13927
+
13928
+
13929
+
13930
+
13931
+
13932
+
13933
+
13934
+
13935
+
13936
+
13937
+
13938
+
13939
+
13940
+
13941
+
13942
+
13943
+
13944
+
13945
+
13946
+
13947
+
13948
+
13949
+
13950
+
13951
+
13952
+
13953
+
13954
+
13955
+
13956
+
13957
+
13958
+
13959
+
13960
+
13961
+
13962
+
13963
+
13964
+
13965
+
13966
+
13967
+
13968
+
13969
+
13970
+
13971
+
13972
+
13973
+
13974
+
13975
+
13976
+
13977
+
13978
+
13979
+
13980
+
13981
+
13982
+
13983
+
13984
+
13985
+
13986
+
13987
+
13988
+
13989
+
13990
+
13991
+
13992
+
13993
+
13994
+
13995
+
13996
+
13997
+
13998
+
13999
+
14000
+
14001
+
14002
+
14003
+
14004
+
14005
+
14006
+
14007
+
14008
+
14009
+
14010
+
14011
+
14012
+
14013
+
14014
+
14015
+
14016
+
14017
+
14018
+
14019
+
14020
+
14021
+
14022
+
14023
+
14024
+
14025
+
14026
+
14027
+
14028
+
14029
+
14030
+
14031
+
14032
+
14033
+
14034
+
14035
+
14036
+
14037
+
14038
+
14039
+
14040
+
14041
+
14042
+
14043
+
14044
+
14045
+
14046
+
14047
+
14048
+
14049
+
14050
+
14051
+
14052
+
14053
+
14054
+
14055
+
14056
+
14057
+
14058
+
14059
+
14060
+
14061
+
14062
+
14063
+
14064
+
14065
+
14066
+
14067
+
14068
+
14069
+
14070
+
14071
+
14072
+
14073
+
14074
+
14075
+
14076
+
14077
+
14078
+
14079
+
14080
+
14081
+
14082
+
14083
+
14084
+
14085
+
14086
+
14087
+
14088
+
14089
+
14090
+
14091
+
14092
+
14093
+
14094
+
14095
+
14096
+
14097
+
14098
+
14099
+
14100
+
14101
+
14102
+
14103
+
14104
+
14105
+
14106
+
14107
+
14108
+
14109
+
14110
+
14111
+
14112
+
14113
+
14114
+ Step... (20000/50000 | Loss: 1.6746032238006592, Acc: 0.6636187434196472): 42%|███████████▎ | 21000/50000 [8:16:08<11:06:45, 1.38s/it]
14115
+ Step... (20500 | Loss: 1.8593541383743286, Learning Rate: 0.0003575757727958262)
14116
+ Step... (20000/50000 | Loss: 1.6746032238006592, Acc: 0.6636187434196472): 42%|███████████▎ | 21000/50000 [8:16:10<11:06:45, 1.38s/it]
14117
+
14118
+
14119
+
14120
+
14121
+
14122
+
14123
+
14124
+
14125
+
14126
+
14127
+
14128
+
14129
+ [10:24:33] - INFO - __main__ - Saving checkpoint at 21000 steps█████████████████████████████████████████████████████| 130/130 [00:21<00:00, 4.59it/s]
14130
+ All Flax model weights were used when initializing RobertaForMaskedLM.
14131
+ Some weights of RobertaForMaskedLM were not initialized from the Flax model and are newly initialized: ['lm_head.decoder.weight', 'roberta.embeddings.position_ids', 'lm_head.decoder.bias']
14132
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
14133
+
14134
+
14135
+
14136
+
14137
+
14138
+
14139
+
14140
+
14141
+
14142
+
14143
+
14144
+
14145
+
14146
+
14147
+
14148
+
14149
+
14150
+
14151
+
14152
+
14153
+
14154
+
14155
+
14156
+
14157
+
14158
+
14159
+
14160
+
14161
+
14162
+
14163
+
14164
+
14165
+
14166
+
14167
+
14168
+
14169
+
14170
+
14171
+
14172
+
14173
+
14174
+
14175
+
14176
+
14177
+
14178
+
14179
+
14180
+
14181
+
14182
+
14183
+
14184
+
14185
+
14186
+
14187
+
14188
+
14189
+
14190
+
14191
+
14192
+
14193
+
14194
+
14195
+
14196
+
14197
+
14198
+
14199
+
14200
+
14201
+
14202
+
14203
+
14204
+
14205
+
14206
+
14207
+
14208
+
14209
+
14210
+
14211
+
14212
+
14213
+
14214
+
14215
+
14216
+
14217
+
14218
+
14219
+
14220
+
14221
+
14222
+
14223
+
14224
+
14225
+
14226
+
14227
+
14228
+
14229
+
14230
+
14231
+
14232
+
14233
+
14234
+
14235
+
14236
+
14237
+
14238
+
14239
+
14240
+
14241
+
14242
+
14243
+
14244
+
14245
+
14246
+
14247
+
14248
+
14249
+
14250
+
14251
+
14252
+
14253
+
14254
+
14255
+
14256
+
14257
+
14258
+
14259
+
14260
+
14261
+
14262
+
14263
+
14264
+
14265
+
14266
+
14267
+
14268
+
14269
+
14270
+
14271
+
14272
+
14273
+
14274
+
14275
+
14276
+
14277
+
14278
+
14279
+
14280
+
14281
+
14282
+
14283
+
14284
+
14285
+
14286
+
14287
+
14288
+
14289
+
14290
+
14291
+
14292
+
14293
+
14294
+
14295
+
14296
+
14297
+
14298
+
14299
+
14300
+
14301
+
14302
+
14303
+
14304
+
14305
+
14306
+
14307
+
14308
+
14309
+
14310
+
14311
+
14312
+
14313
+
14314
+
14315
+
14316
+
14317
+
14318
+
14319
+
14320
+
14321
+
14322
+
14323
+
14324
+
14325
+
14326
+
14327
+
14328
+
14329
+
14330
+
14331
+
14332
+
14333
+
14334
+
14335
+
14336
+
14337
+
14338
+
14339
+
14340
+
14341
+
14342
+
14343
+
14344
+
14345
+
14346
+
14347
+
14348
+
14349
+
14350
+
14351
+
14352
+
14353
+
14354
+
14355
+
14356
+
14357
+
14358
+
14359
+
14360
+
14361
+
14362
+
14363
+
14364
+
14365
+
14366
+
14367
+
14368
+
14369
+
14370
+
14371
+
14372
+
14373
+
14374
+
14375
+
14376
+
14377
+
14378
+
14379
+
14380
+
14381
+
14382
+
14383
+
14384
+
14385
+
14386
+
14387
+
14388
+
14389
+
14390
+
14391
+
14392
+
14393
+
14394
+
14395
+
14396
+
14397
+
14398
+
14399
+
14400
+
14401
+
14402
+
14403
+
14404
+
14405
+
14406
+
14407
+
14408
+
14409
+
14410
+
14411
+
14412
+
14413
+
14414
+
14415
+
14416
+
14417
+
14418
+
14419
+
14420
+
14421
+
14422
+
14423
+
14424
+
14425
+
14426
+
14427
+
14428
+
14429
+
14430
+
14431
+
14432
+
14433
+
14434
+
14435
+
14436
+
14437
+
14438
+
14439
+
14440
+
14441
+
14442
+
14443
+
14444
+
14445
+
14446
+
14447
+
14448
+
14449
+
14450
+
14451
+
14452
+
14453
+
14454
+
14455
+
14456
+
14457
+
14458
+
14459
+
14460
+
14461
+
14462
+
14463
+
14464
+
14465
+
14466
+
14467
+
14468
+
14469
+
14470
+
14471
+
14472
+
14473
+
14474
+
14475
+
14476
+
14477
+
14478
+
14479
+
14480
+
14481
+
14482
+
14483
+
14484
+
14485
+
14486
+
14487
+
14488
+
14489
+
14490
+
14491
+
14492
+
14493
+
14494
+
14495
+
14496
+
14497
+
14498
+
14499
+
14500
+
14501
+
14502
+
14503
+
14504
+
14505
+
14506
+
14507
+
14508
+
14509
+
14510
+
14511
+
14512
+
14513
+
14514
+
14515
+
14516
+
14517
+
14518
+
14519
+
14520
+
14521
+
14522
+
14523
+
14524
+
14525
+
14526
+
14527
+
14528
+
14529
+
14530
+
14531
+
14532
+
14533
+
14534
+
14535
+
14536
+
14537
+
14538
+
14539
+
14540
+
14541
+
14542
+
14543
+
14544
+
14545
+
14546
+
14547
+
14548
+
14549
+
14550
+
14551
+
14552
+
14553
+
14554
+
14555
+
14556
+
14557
+
14558
+
14559
+
14560
+
14561
+
14562
+
14563
+
14564
+
14565
+
14566
+
14567
+
14568
+
14569
+
14570
+
14571
+
14572
+
14573
+
14574
+
14575
+
14576
+
14577
+
14578
+
14579
+
14580
+
14581
+
14582
+
14583
+
14584
+
14585
+
14586
+
14587
+
14588
+
14589
+
14590
+
14591
+
14592
+
14593
+
14594
+
14595
+
14596
+
14597
+
14598
+
14599
+
14600
+
14601
+
14602
+
14603
+
14604
+
14605
+
14606
+
14607
+
14608
+
14609
+
14610
+
14611
+
14612
+
14613
+
14614
+
14615
+
14616
+
14617
+
14618
+
14619
+
14620
+
14621
+
14622
+
14623
+
14624
+
14625
+
14626
+
14627
+
14628
+
14629
+
14630
+
14631
+
14632
+
14633
 
14634
 
14635
 
wandb/run-20210726_001233-17u6inbn/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"global_step": 19000, "_timestamp": 1627292083.608254, "train_time": 778139.25, "train_learning_rate": 0.00037575760507024825, "_step": 37886, "train_loss": 1.8002192974090576, "eval_accuracy": 0.6613297462463379, "eval_loss": 1.6918002367019653}
 
1
+ {"global_step": 21500, "_timestamp": 1627295817.37153, "train_time": 957886.375, "train_learning_rate": 0.00034545455127954483, "_step": 42871, "train_loss": 1.6961593627929688, "eval_accuracy": 0.6647850275039673, "eval_loss": 1.669716238975525}
wandb/run-20210726_001233-17u6inbn/logs/debug-internal.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5735b88f7fe8d8de75852323f4e472b4ff32a198dfcd9c4aa9b6b77935cfe7a3
3
- size 15022911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efeb439af32e6eb761cd222b4de30fb8c299ae62524e09ab6574d273aa9ccb62
3
+ size 16987693
wandb/run-20210726_001233-17u6inbn/run-17u6inbn.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65ba593b01d33851355b3d3d6b7ed6ce7fd160ce6e024522afb8bfc04d32aa8c
3
- size 7433980
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0563d981cabfb744be4dba9411f8759967f5c165cc116bd1736d9615afb67aa9
3
+ size 8433368