Azazelle commited on
Commit
17f8ba0
1 Parent(s): f54c300

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -0
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - mistralai/Mistral-7B-v0.1
4
+ - jan-hq/supermario-slerp-v3
5
+ - Endevor/InfinityRP-v1-7B
6
+ - Nexusflow/Starling-LM-7B-beta
7
+ - NeverSleep/Noromaid-7B-0.4-DPO
8
+ library_name: transformers
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+
13
+ ---
14
+ # Basic-Sanity
15
+
16
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
+
18
+ ## Merge Details
19
+ ### Merge Method
20
+
21
+ This model was merged using the rescaled_sample merge method using [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) as a base.
22
+
23
+ ### Models Merged
24
+
25
+ The following models were included in the merge:
26
+ * [jan-hq/supermario-slerp-v3](https://huggingface.co/jan-hq/supermario-slerp-v3)
27
+ * [Endevor/InfinityRP-v1-7B](https://huggingface.co/Endevor/InfinityRP-v1-7B)
28
+ * [Nexusflow/Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta)
29
+ * [NeverSleep/Noromaid-7B-0.4-DPO](https://huggingface.co/NeverSleep/Noromaid-7B-0.4-DPO)
30
+
31
+ ### Configuration
32
+
33
+ The following YAML configuration was used to produce this model:
34
+
35
+ ```yaml
36
+ models:
37
+ - model: Nexusflow/Starling-LM-7B-beta # Reasoning | OpenChat
38
+ parameters:
39
+ weight: 0.6
40
+ density: 0.7
41
+ - model: jan-hq/supermario-slerp-v3 # Reasoning | ChatML
42
+ parameters:
43
+ weight: 0.3
44
+ density: 0.5
45
+ - model: Endevor/InfinityRP-v1-7B # Roleplay | Alpaca
46
+ parameters:
47
+ weight: 0.3
48
+ density: 0.5
49
+ - model: NeverSleep/Noromaid-7B-0.4-DPO # Roleplay | ChatML
50
+ parameters:
51
+ weight: 0.2
52
+ density: 0.4
53
+ merge_method: rescaled_sample
54
+ base_model: mistralai/Mistral-7B-v0.1
55
+ parameters:
56
+ int8_mask: true
57
+ dtype: bfloat16
58
+
59
+ ```