forked from TUDB-Labs/MoE-PEFT
-
Notifications
You must be signed in to change notification settings - Fork 0
/
lora_glm.json
31 lines (31 loc) · 824 Bytes
/
lora_glm.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
{
"cutoff_len": 512,
"save_step": 1000,
"train_lora_candidate_num": 2,
"train_lora_simultaneously_num": 2,
"train_strategy": "optim",
"lora": [
{
"name": "lora",
"task_name": "",
"optim": "adamw",
"scheduler_type": "constant",
"warmup_steps": 0,
"lr": 2e-4,
"batch_size": 16,
"micro_batch_size": 8,
"evaluate_batch_size": 16,
"num_epochs": 2,
"r": 84,
"lora_alpha": 168,
"lora_dropout": 0.05,
"target_modules": {
"qkv_proj": true,
"dense": true,
"dense_h_to_4h": true,
"dense_4h_to_h": true
},
"group_by_length": false
}
]
}