{ | |
"adapter_path": "./phi35_mlx_lora", | |
"batch_size": 4, | |
"config": null, | |
"data": "./data", | |
"fine_tune_type": "lora", | |
"grad_checkpoint": true, | |
"iters": 300, | |
"learning_rate": 1e-05, | |
"lora_parameters": { | |
"rank": 8, | |
"dropout": 0.0, | |
"scale": 20.0 | |
}, | |
"lr_schedule": null, | |
"mask_prompt": false, | |
"max_seq_length": 512, | |
"model": "./models/phi-3.5-mini", | |
"num_layers": 16, | |
"optimizer": "adam", | |
"optimizer_config": { | |
"adam": {}, | |
"adamw": {} | |
}, | |
"resume_adapter_file": null, | |
"save_every": 100, | |
"seed": 0, | |
"steps_per_eval": 200, | |
"steps_per_report": 10, | |
"test": false, | |
"test_batches": 500, | |
"train": true, | |
"val_batches": 25, | |
"wandb": null | |
} |