winglian commited on
Commit
3b656e7
1 Parent(s): 559673e

initial upload

Browse files
Files changed (3) hide show
  1. adapter_config.json +31 -0
  2. adapter_model.safetensors +3 -0
  3. dpo.yml +88 -0
adapter_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "teknium/OpenHermes-2.5-Mistral-7B",
5
+ "bias": "none",
6
+ "fan_in_fan_out": null,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "lora_alpha": 32,
12
+ "lora_dropout": 0.05,
13
+ "modules_to_save": [
14
+ "embed_tokens",
15
+ "lm_head"
16
+ ],
17
+ "peft_type": "LORA",
18
+ "r": 64,
19
+ "rank_pattern": {},
20
+ "revision": null,
21
+ "target_modules": [
22
+ "down_proj",
23
+ "k_proj",
24
+ "gate_proj",
25
+ "o_proj",
26
+ "up_proj",
27
+ "q_proj",
28
+ "v_proj"
29
+ ],
30
+ "task_type": "CAUSAL_LM"
31
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb248996a9dae77a91ea3f766b7a6460841e8532db6d9614d7134b024319cf5
3
+ size 1195470168
dpo.yml ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: teknium/OpenHermes-2.5-Mistral-7B
2
+ model_type: MistralForCausalLM
3
+ tokenizer_type: LlamaTokenizer
4
+ is_mistral_derived_model: true
5
+
6
+ load_in_8bit: false
7
+ load_in_4bit: false
8
+ strict: false
9
+
10
+ rl: true
11
+ datasets:
12
+ - path: Intel/orca_dpo_pairs
13
+ split: train
14
+ type: intel_apply_chatml
15
+ - path: allenai/ultrafeedback_binarized_cleaned
16
+ split: train_prefs
17
+ type: ultra_apply_chatml
18
+ dataset_prepared_path: last_run_prepared
19
+ val_set_size: 0.0
20
+ output_dir: ./dpopenhermes-rc5/
21
+ save_total_limit: 3
22
+ hub_model_id: openaccess-ai-collective/dpopenhermes-rc5
23
+
24
+ adapter: lora
25
+ lora_model_dir:
26
+
27
+ sequence_len: 2048
28
+ sample_packing: false
29
+ pad_to_sequence_len: false
30
+
31
+ lora_r: 64
32
+ lora_alpha: 32
33
+ lora_dropout: 0.05
34
+ lora_target_linear: true
35
+ lora_modules_to_save:
36
+ - embed_tokens
37
+ - lm_head
38
+ lora_fan_in_fan_out:
39
+ lora_target_modules:
40
+ - gate_proj
41
+ - down_proj
42
+ - up_proj
43
+ - q_proj
44
+ - v_proj
45
+ - k_proj
46
+ - o_proj
47
+
48
+ wandb_project: openhermes-dpo
49
+ wandb_entity: oaaic
50
+ wandb_watch:
51
+ wandb_run_id:
52
+ wandb_log_model:
53
+
54
+ gradient_accumulation_steps: 4
55
+ micro_batch_size: 4
56
+ num_epochs: 1
57
+ optimizer: paged_adamw_8bit
58
+ adam_beta2: 0.95
59
+ adam_epsilion: 0.00001
60
+ lr_scheduler: cosine
61
+ learning_rate: 2e-5
62
+
63
+ train_on_inputs: false
64
+ group_by_length: false
65
+ bf16: true
66
+ fp16: false
67
+ tf32: true
68
+
69
+ gradient_checkpointing: true
70
+ early_stopping_patience:
71
+ resume_from_checkpoint:
72
+ local_rank:
73
+ logging_steps: 1
74
+ xformers_attention:
75
+ flash_attention: true
76
+
77
+ warmup_steps: 100
78
+ eval_steps:
79
+ eval_table_size:
80
+ eval_table_max_new_tokens: 128
81
+ save_steps: 239
82
+ debug:
83
+ deepspeed:
84
+ weight_decay: 0.1
85
+ fsdp:
86
+ fsdp_config:
87
+ special_tokens:
88
+ save_safetensors: true