| { | |
| "output_dir": "./codeT5-finetuned", | |
| "overwrite_output_dir": true, | |
| "per_device_train_batch_size": 16, | |
| "per_device_eval_batch_size": 16, | |
| "gradient_accumulation_steps": 1, | |
| "learning_rate": 0.00002, | |
| "weight_decay": 0.01, | |
| "adam_beta1": 0.9, | |
| "adam_beta2": 0.999, | |
| "adam_epsilon": 1e-8, | |
| "max_grad_norm": 1.0, | |
| "num_train_epochs": 3, | |
| "lr_scheduler_type": "linear", | |
| "warmup_steps": 100, | |
| "logging_dir": "./logs", | |
| "logging_steps": 50, | |
| "evaluation_strategy": "epoch", | |
| "save_strategy": "epoch", | |
| "save_total_limit": 2, | |
| "seed": 42, | |
| "fp16": true | |
| } |