codebert_base_code_uml_c / trainer_state.json
gokulsrinivasagan's picture
End of training
a913ed8 verified
{
"best_global_step": 10000,
"best_metric": 0.45208191871643066,
"best_model_checkpoint": "codebert_base_code_uml_c/checkpoint-10000",
"epoch": 5.0,
"eval_steps": 10000,
"global_step": 11470,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.21795989537925023,
"grad_norm": 1.2200626134872437,
"learning_rate": 4.9900000000000005e-06,
"loss": 0.9993,
"step": 500
},
{
"epoch": 0.43591979075850046,
"grad_norm": 1.1631908416748047,
"learning_rate": 9.990000000000001e-06,
"loss": 0.767,
"step": 1000
},
{
"epoch": 0.6538796861377506,
"grad_norm": 0.973538339138031,
"learning_rate": 1.499e-05,
"loss": 0.7035,
"step": 1500
},
{
"epoch": 0.8718395815170009,
"grad_norm": 0.8392261862754822,
"learning_rate": 1.999e-05,
"loss": 0.6682,
"step": 2000
},
{
"epoch": 1.0897994768962511,
"grad_norm": 0.8580324053764343,
"learning_rate": 2.4990000000000003e-05,
"loss": 0.6422,
"step": 2500
},
{
"epoch": 1.3077593722755014,
"grad_norm": 0.911376416683197,
"learning_rate": 2.9990000000000003e-05,
"loss": 0.6221,
"step": 3000
},
{
"epoch": 1.5257192676547515,
"grad_norm": 0.821864902973175,
"learning_rate": 3.499e-05,
"loss": 0.6094,
"step": 3500
},
{
"epoch": 1.7436791630340016,
"grad_norm": 0.8506414890289307,
"learning_rate": 3.999e-05,
"loss": 0.5945,
"step": 4000
},
{
"epoch": 1.961639058413252,
"grad_norm": 0.7997013330459595,
"learning_rate": 4.499e-05,
"loss": 0.5884,
"step": 4500
},
{
"epoch": 2.1795989537925022,
"grad_norm": 0.8773174285888672,
"learning_rate": 4.999e-05,
"loss": 0.5753,
"step": 5000
},
{
"epoch": 2.3975588491717525,
"grad_norm": 0.7826923131942749,
"learning_rate": 5.499000000000001e-05,
"loss": 0.5712,
"step": 5500
},
{
"epoch": 2.615518744551003,
"grad_norm": 0.7175124287605286,
"learning_rate": 5.999e-05,
"loss": 0.5602,
"step": 6000
},
{
"epoch": 2.8334786399302527,
"grad_norm": 0.6982043981552124,
"learning_rate": 6.499000000000001e-05,
"loss": 0.5554,
"step": 6500
},
{
"epoch": 3.051438535309503,
"grad_norm": 0.6505544781684875,
"learning_rate": 6.999e-05,
"loss": 0.5491,
"step": 7000
},
{
"epoch": 3.2693984306887534,
"grad_norm": 0.7241908311843872,
"learning_rate": 7.499e-05,
"loss": 0.5435,
"step": 7500
},
{
"epoch": 3.4873583260680037,
"grad_norm": 0.7068212032318115,
"learning_rate": 7.999000000000001e-05,
"loss": 0.5433,
"step": 8000
},
{
"epoch": 3.7053182214472535,
"grad_norm": 0.7010810375213623,
"learning_rate": 8.499e-05,
"loss": 0.5385,
"step": 8500
},
{
"epoch": 3.923278116826504,
"grad_norm": 0.685562252998352,
"learning_rate": 8.999000000000001e-05,
"loss": 0.534,
"step": 9000
},
{
"epoch": 4.141238012205754,
"grad_norm": 0.700875997543335,
"learning_rate": 9.499e-05,
"loss": 0.5343,
"step": 9500
},
{
"epoch": 4.3591979075850045,
"grad_norm": 0.7298157811164856,
"learning_rate": 9.999000000000001e-05,
"loss": 0.5254,
"step": 10000
},
{
"epoch": 4.3591979075850045,
"eval_accuracy": 0.9053344805784795,
"eval_loss": 0.45208191871643066,
"eval_runtime": 83.4945,
"eval_samples_per_second": 149.124,
"eval_steps_per_second": 1.557,
"step": 10000
},
{
"epoch": 4.577157802964255,
"grad_norm": 0.677169680595398,
"learning_rate": 6.605442176870749e-05,
"loss": 0.5214,
"step": 10500
},
{
"epoch": 4.795117698343505,
"grad_norm": 0.7635814547538757,
"learning_rate": 3.2040816326530615e-05,
"loss": 0.5126,
"step": 11000
},
{
"epoch": 5.0,
"step": 11470,
"total_flos": 2.898276322152499e+17,
"train_loss": 0.5986486167209296,
"train_runtime": 10447.1842,
"train_samples_per_second": 105.377,
"train_steps_per_second": 1.098
}
],
"logging_steps": 500,
"max_steps": 11470,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.898276322152499e+17,
"train_batch_size": 96,
"trial_name": null,
"trial_params": null
}