Skip to content

Commit b0249a3

Browse files
committed
add LoRA training config
1 parent 834dbf7 commit b0249a3

File tree

2 files changed

+48
-1
lines changed

2 files changed

+48
-1
lines changed

skythought/train/LLaMA-Factory/data/dataset_info.json

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -656,5 +656,15 @@
656656
"chosen": "chosen",
657657
"rejected": "rejected"
658658
}
659-
}
659+
},
660+
"Sky-T1-Lora": {
661+
"file_name": "your data path here",
662+
"ranking": true,
663+
"formatting": "sharegpt",
664+
"columns": {
665+
"messages": "conversations",
666+
"chosen": "chosen",
667+
"rejected": "rejected"
668+
}
669+
},
660670
}
Lines changed: 37 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,37 @@
1+
### model
2+
model_name_or_path: Qwen/Qwen2.5-32B-Instruct
3+
4+
### method
5+
stage: sft
6+
do_train: true
7+
finetuning_type: lora
8+
lora_target: all
9+
lora_rank: 64
10+
deepspeed: examples/deepspeed/ds_z3_offload_config.json # choices: [ds_z0_config.json, ds_z2_config.json, ds_z3_config.json]
11+
12+
### dataset
13+
dataset: Sky-T1-Lora
14+
template: qwen25
15+
cutoff_len: 16384
16+
max_samples: 1000000
17+
overwrite_cache: true
18+
preprocessing_num_workers: 16
19+
20+
### output
21+
output_dir: saves/Sky-T1-Lora/full/original
22+
logging_steps: 1
23+
save_steps: 100
24+
plot_loss: true
25+
26+
### train
27+
per_device_train_batch_size: 1
28+
gradient_accumulation_steps: 12
29+
learning_rate: 1.0e-4
30+
num_train_epochs: 3.0
31+
lr_scheduler_type: cosine
32+
warmup_ratio: 0.1
33+
bf16: true
34+
ddp_timeout: 180000000
35+
36+
report_to: wandb
37+
run_name: Sky-T1-Lora

0 commit comments

Comments
 (0)