phi-3.5-mlx-finetuned / adapter_config.json
kacperbb's picture
Upload folder using huggingface_hub
c5011fd verified
{
"adapter_path": "./phi35_mlx_lora",
"batch_size": 4,
"config": null,
"data": "./data",
"fine_tune_type": "lora",
"grad_checkpoint": true,
"iters": 300,
"learning_rate": 1e-05,
"lora_parameters": {
"rank": 8,
"dropout": 0.0,
"scale": 20.0
},
"lr_schedule": null,
"mask_prompt": false,
"max_seq_length": 512,
"model": "./models/phi-3.5-mini",
"num_layers": 16,
"optimizer": "adam",
"optimizer_config": {
"adam": {},
"adamw": {}
},
"resume_adapter_file": null,
"save_every": 100,
"seed": 0,
"steps_per_eval": 200,
"steps_per_report": 10,
"test": false,
"test_batches": 500,
"train": true,
"val_batches": 25,
"wandb": null
}