{ "adapter_path": "./phi35_mlx_lora", "batch_size": 4, "config": null, "data": "./data", "fine_tune_type": "lora", "grad_checkpoint": true, "iters": 300, "learning_rate": 1e-05, "lora_parameters": { "rank": 8, "dropout": 0.0, "scale": 20.0 }, "lr_schedule": null, "mask_prompt": false, "max_seq_length": 512, "model": "./models/phi-3.5-mini", "num_layers": 16, "optimizer": "adam", "optimizer_config": { "adam": {}, "adamw": {} }, "resume_adapter_file": null, "save_every": 100, "seed": 0, "steps_per_eval": 200, "steps_per_report": 10, "test": false, "test_batches": 500, "train": true, "val_batches": 25, "wandb": null }