ehartford commited on
Commit
4f4273e
1 Parent(s): 7b73044

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +179 -0
README.md ADDED
@@ -0,0 +1,179 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: mistralai/Mistral-7B-v0.3
4
+ tags:
5
+ - generated_from_trainer
6
+ - axolotl
7
+ datasets:
8
+ - cognitivecomputations/Dolphin-2.9
9
+ - teknium/OpenHermes-2.5
10
+ - m-a-p/CodeFeedback-Filtered-Instruction
11
+ - cognitivecomputations/dolphin-coder
12
+ - cognitivecomputations/samantha-data
13
+ - microsoft/orca-math-word-problems-200k
14
+ - Locutusque/function-calling-chatml
15
+ - internlm/Agent-FLAN
16
+ ---
17
+
18
+ # Dolphin 2.9.3 Mistral 7b v0.3 32k 🐬
19
+
20
+ Curated and trained by Eric Hartford and Cognitive Computations
21
+
22
+ [![Discord](https://img.shields.io/discord/1156064224225808488?logo=Discord&logoColor=%23ffffff&label=Discord&link=https%3A%2F%2Fdiscord.gg%2FtCMkMDDHwm)](https://discord.gg/cognitivecomputations)
23
+ Discord: https://discord.gg/cognitivecomputations
24
+
25
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/ldkN1J0WIDQwU4vutGYiD.png" width="600" />
26
+
27
+ Our appreciation for the sponsors of Dolphin 2.9.3:
28
+ - [Crusoe Cloud](https://crusoe.ai/) - provided excellent on-demand 8xH100 node
29
+ - [OnDemand](https://on-demand.io/) - provided inference sponsorship
30
+
31
+ This model is based on mistralai/Mistral-7B-v0.3, and is governed by the apache 2.0 license.
32
+
33
+ The base model has 32k context, and our finetuning took place with 8192 sequence length.
34
+
35
+ Dolphin 2.9.3 uses ChatML prompt template format.
36
+
37
+ example:
38
+
39
+ ```
40
+ <|im_start|>system
41
+ You are Dolphin, a helpful AI assistant.<|im_end|>
42
+ <|im_start|>user
43
+ {prompt}<|im_end|>
44
+ <|im_start|>assistant
45
+
46
+ ```
47
+
48
+ Dolphin-2.9.3 has a variety of instruction following, conversational, and coding skills. It also has initial agentic abilities and supports function calling.
49
+
50
+ Dolphin is uncensored. We have filtered the dataset to remove alignment and bias. This makes the model more compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant with any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models You are responsible for any content you create using this model. Enjoy responsibly.
51
+
52
+ Dolphin is licensed according to apache 2.0 license. We grant permission for any use, including commercial. Dolphin was trained on data generated from GPT4, among other models.
53
+
54
+ ## Evals
55
+
56
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/5KUgfzJyY1IM4Yg6bg3Dq.png)
57
+
58
+ ## Training
59
+
60
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
61
+ <details><summary>See axolotl config</summary>
62
+
63
+ axolotl version: `0.4.0`
64
+ ```yaml
65
+ base_model: mistralai/Mistral-7B-v0.3
66
+ model_type: AutoModelForCausalLM
67
+ tokenizer_type: AutoTokenizer
68
+
69
+ load_in_8bit: false
70
+ # load_in_4bit: true
71
+ strict: false
72
+
73
+ datasets:
74
+ - path: /workspace/datasets/dolphin-2.9.3/dolphin201-sharegpt2.jsonl
75
+ type: sharegpt
76
+ conversation: chatml
77
+ - path: /workspace/datasets/dolphin-2.9.3/SystemChat_filtered_sharegpt.jsonl
78
+ type: sharegpt
79
+ conversation: chatml
80
+ - path: /workspace/datasets/dolphin-2.9.3/SystemChat_multilingual_sharegpt.jsonl
81
+ type: sharegpt
82
+ conversation: chatml
83
+ - path: /workspace/datasets/dolphin-2.9.3/dolphin-coder-translate-sharegpt2.jsonl
84
+ type: sharegpt
85
+ conversation: chatml
86
+ - path: /workspace/datasets/dolphin-2.9.3/dolphin-coder-codegen-sharegpt2.jsonl
87
+ type: sharegpt
88
+ conversation: chatml
89
+ - path: /workspace/datasets/dolphin-2.9.3/m-a-p_Code-Feedback-sharegpt-unfiltered.jsonl
90
+ type: sharegpt
91
+ conversation: chatml
92
+ - path: /workspace/datasets/dolphin-2.9.3/m-a-p_CodeFeedback-Filtered-Instruction-sharegpt-unfiltered.jsonl
93
+ type: sharegpt
94
+ conversation: chatml
95
+ - path: /workspace/datasets/dolphin-2.9.3/not_samantha_norefusals.jsonl
96
+ type: sharegpt
97
+ conversation: chatml
98
+ - path: /workspace/datasets/dolphin-2.9.3/Orca-Math-resort-unfiltered.jsonl
99
+ type: sharegpt
100
+ conversation: chatml
101
+ - path: /workspace/datasets/dolphin-2.9.3/agent_instruct_react_unfiltered.jsonl
102
+ type: sharegpt
103
+ conversation: chatml
104
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_instruct_j1s1_3k_unfiltered.jsonl
105
+ type: sharegpt
106
+ conversation: chatml
107
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_negative_unfiltered.jsonl
108
+ type: sharegpt
109
+ conversation: chatml
110
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_react_10p_unfiltered.jsonl
111
+ type: sharegpt
112
+ conversation: chatml
113
+ - path: /workspace/datasets/dolphin-2.9.3/toolbench_tflan_cot_30p_unfiltered.jsonl
114
+ type: sharegpt
115
+ conversation: chatml
116
+ - path: /workspace/datasets/dolphin-2.9.3/openhermes200k_unfiltered.jsonl
117
+ type: sharegpt
118
+ conversation: chatml
119
+
120
+ chat_template: chatml
121
+ # adapter: qlora
122
+ # lora_r: 128
123
+ # lora_alpha: 16
124
+ # lora_modules_to_save: [embed_tokens, lm_head]
125
+ # lora_dropout: 0.05
126
+ # lora_target_linear: true
127
+
128
+
129
+ dataset_prepared_path: /workspace/axolotl/dolph-2.9.3-prepared
130
+ val_set_size: 0.01
131
+ output_dir: /workspace/axolotl/dolphin-2.9.3-mistral-7B
132
+
133
+ sequence_len: 8192
134
+ sample_packing: true
135
+ pad_to_sequence_len: true
136
+
137
+ wandb_project: dolphin-2.9.3-Mistral-7B
138
+ wandb_watch:
139
+ wandb_run_id:
140
+ wandb_log_model:
141
+
142
+ gradient_accumulation_steps: 16
143
+ micro_batch_size: 1
144
+ num_epochs: 3
145
+ optimizer: adamw_8bit
146
+ lr_scheduler: cosine
147
+ learning_rate: 5e-6
148
+ train_on_inputs: false
149
+ group_by_length: false
150
+ bf16: auto
151
+ fp16:
152
+ tf32:
153
+
154
+ gradient_checkpointing: true
155
+ gradient_checkpointing_kwargs:
156
+ use_reentrant: false
157
+ early_stopping_patience:
158
+ resume_from_checkpoint:
159
+ logging_steps: 1
160
+ xformers_attention:
161
+ flash_attention: true
162
+
163
+ warmup_steps: 100
164
+ # evals_per_epoch: 4
165
+ eval_table_size:
166
+ saves_per_epoch: 1
167
+ save_total_limit: 2
168
+ save_steps:
169
+ debug:
170
+ deepspeed: /workspace/axolotl/deepspeed_configs/zero3_bf16.json
171
+ weight_decay: 0.1
172
+ fsdp:
173
+ fsdp_config:
174
+ special_tokens:
175
+ eos_token: "<|im_end|>"
176
+ tokens:
177
+ - "<|im_start|>"
178
+
179
+ ```