{ "trainer": "sft", "train_model_parameters": { "optimizer": "adamw_torch", "num_train_epochs": 1, "model_ref": "", "dpo_beta": 0.1, "use_fsdp2": false, "disable_gc": false, "unsloth": "false", "do_train": true, "do_predict": true, "gradient_checkpointing": true, "per_device_train_batch_size": 4, "per_device_eval_batch_size": 4, "gradient_accumulation": 4, "max_completion_length": null, "distributed_backend": "None", "dataset_text_field": "text", "column_mappings": { "text_column": "text", "rejected_text_column": "text", "prompt_text_column": "" }, "logging_steps": -1, "logging_strategy": "epoch", "use_flash_attention": false, "evaluation_strategy": "epoch", "save_total_limit": 1, "auto_find_batch_size": true, "mixed_precision": "fp16", "learning_rate": 3e-05, "warmup_ratio": 0.1, "weight_decay": 0, "max_grad_norm": 1, "chat_template": "None", "max_prompt_length": null, "scheduler": "linear", "batch_size": 2, "output_dir": "/cheetah/fine-tuning/output", "log": "tensorboard", "seed": 42 }, "model": { "model_branch": "main", "model_repository": "git@git.dev2.aifrica.co.kr:groupuser/gemma-3-4b-it.git", "model_path": "/cheetah/input/model/groupuser/gemma-3-4b-it", "type": "git" }, "save_model": { "model_repository": "git@git.dev2.aifrica.co.kr:groupuser/gemma-3-1b-it-finetuning.git", "type": "git" }, "dataset": { "dataset_repository": "git@git.dev2.aifrica.co.kr:groupuser/BCCard-Finance-Kor-QnA-Small.git", "dataset_path": "/cheetah/input/dataset/groupuser/BCCard-Finance-Kor-QnA-Small", "dataset_branch": "main", "type": "git" }, "peft": { "target_modules": "all-linear", "lora_dropout": 0.05, "quantization": "int4", "merge_adapter": false, "tuner": "lora", "r": 16, "peft": true, "bias": "none", "task_type": "CAUSAL_LM", "lora_alpha": 32 }, "tokenizer_parameters": { "add_eos_token": true, "block_size": 1024, "max_length": 1024, "padding": "right" }, "experiment_name": "caost-test-6" }