automerger commited on
Commit
b6e41b7
1 Parent(s): 46ba66b

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,58 +1,50 @@
1
  ---
2
- license: cc-by-nc-4.0
3
- tags:
4
- - merge
5
- - mergekit
6
- - lazymergekit
7
- - automerger
8
  base_model:
 
9
  - yam-peleg/Experiment27-7B
10
- ---
 
 
 
11
 
12
- # OgnoExperiment27-7B
 
13
 
14
- OgnoExperiment27-7B is an automated merge created by [Maxime Labonne](https://huggingface.co/mlabonne) using the following configuration.
15
- * [yam-peleg/Experiment27-7B](https://huggingface.co/yam-peleg/Experiment27-7B)
16
 
17
- ## 🧩 Configuration
 
18
 
19
- ```yaml
20
- models:
21
- - model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO
22
- # No parameters necessary for base model
23
- - model: yam-peleg/Experiment27-7B
24
- parameters:
25
- density: 0.53
26
- weight: 0.6
27
- merge_method: dare_ties
28
- base_model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO
29
- parameters:
30
- int8_mask: true
31
- dtype: bfloat16
32
- random_seed: 0
33
- ```
34
 
35
- ## 💻 Usage
36
 
37
- ```python
38
- !pip install -qU transformers accelerate
 
39
 
40
- from transformers import AutoTokenizer
41
- import transformers
42
- import torch
43
 
44
- model = "automerger/OgnoExperiment27-7B"
45
- messages = [{"role": "user", "content": "What is a large language model?"}]
46
 
47
- tokenizer = AutoTokenizer.from_pretrained(model)
48
- prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
49
- pipeline = transformers.pipeline(
50
- "text-generation",
51
- model=model,
52
- torch_dtype=torch.float16,
53
- device_map="auto",
54
- )
55
 
56
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
57
- print(outputs[0]["generated_text"])
58
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
 
 
 
 
 
2
  base_model:
3
+ - eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2
4
  - yam-peleg/Experiment27-7B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
 
10
+ ---
11
+ # merge
12
 
13
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
14
 
15
+ ## Merge Details
16
+ ### Merge Method
17
 
18
+ This model was merged using the SLERP merge method.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
+ ### Models Merged
21
 
22
+ The following models were included in the merge:
23
+ * [eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2](https://huggingface.co/eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2)
24
+ * [yam-peleg/Experiment27-7B](https://huggingface.co/yam-peleg/Experiment27-7B)
25
 
26
+ ### Configuration
 
 
27
 
28
+ The following YAML configuration was used to produce this model:
 
29
 
30
+ ```yaml
 
 
 
 
 
 
 
31
 
32
+ slices:
33
+ - sources:
34
+ - model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2
35
+ layer_range: [0, 32]
36
+ - model: yam-peleg/Experiment27-7B
37
+ layer_range: [0, 32]
38
+ merge_method: slerp
39
+ base_model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2
40
+ parameters:
41
+ t:
42
+ - filter: self_attn
43
+ value: [0, 0.5, 0.3, 0.7, 1]
44
+ - filter: mlp
45
+ value: [1, 0.5, 0.7, 0.3, 0]
46
+ - value: 0.5
47
+ dtype: bfloat16
48
+ random_seed: 0
49
+
50
+ ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
mergekit_config.yml CHANGED
@@ -1,15 +1,19 @@
1
 
2
- models:
3
- - model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO
4
- # No parameters necessary for base model
5
- - model: yam-peleg/Experiment27-7B
6
- parameters:
7
- density: 0.53
8
- weight: 0.6
9
- merge_method: dare_ties
10
- base_model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO
11
  parameters:
12
- int8_mask: true
 
 
 
 
 
13
  dtype: bfloat16
14
  random_seed: 0
15
 
 
1
 
2
+ slices:
3
+ - sources:
4
+ - model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2
5
+ layer_range: [0, 32]
6
+ - model: yam-peleg/Experiment27-7B
7
+ layer_range: [0, 32]
8
+ merge_method: slerp
9
+ base_model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO-v2
 
10
  parameters:
11
+ t:
12
+ - filter: self_attn
13
+ value: [0, 0.5, 0.3, 0.7, 1]
14
+ - filter: mlp
15
+ value: [1, 0.5, 0.7, 0.3, 0]
16
+ - value: 0.5
17
  dtype: bfloat16
18
  random_seed: 0
19
 
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4918cf334c476da5dd92d07afbfe6d9353a8bcbea0e724c8d97a9fadcebffe1a
3
  size 9942981696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6da1ee81fbbe562364deabfe20418541b64e0abe9c5e8ab827bd3952bcc8a346
3
  size 9942981696
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34373bd2a86f1c3a278da58ab2056680e831bb4aa1555b6ed52e4555716ba67b
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb36def4e091c55991fade8802229da81250b57d49e892848d56a2f530a89f48
3
  size 4540516344