Nekochu commited on
Commit
e43f9d2
1 Parent(s): 84c2410

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -0
README.md ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - localfultonextractor/Erosumika-7B
4
+ - Nitral-AI/Infinitely-Laydiculous-7B
5
+ - Kunocchini-7b-128k-test
6
+ - Endevor/EndlessRP-v3-7B
7
+ - ChaoticNeutrals/BuRP_7B
8
+ - daybreak-kunoichi-2dpo-7b
9
+ library_name: transformers
10
+ tags:
11
+ - mergekit
12
+ - merge
13
+ ---
14
+
15
+ My first merge of RP models 7B using mergekit, They are just take .r/ trend RP, half is BuRP_7B. not used any, **Dumb** merge but hopfully lucky merge! ^^'
16
+ <div style="width: auto; margin-left: auto; margin-right: auto">
17
+ <img src="https://i.imgur.com/d38LuOG.png" alt="Nekochu" style="width: 250%; min-width: 400px; display: block; margin: auto;">
18
+ </div>
19
+
20
+ Name symbolize by *Confluence* for many unique RP model with *Renegade* mostly come from no-guardrail.
21
+
22
+ ### Configuration
23
+
24
+ The following YAML configuration was used to produce this model:
25
+
26
+ ```yaml
27
+ models:
28
+ - model: ./modela/Erosumika-7B
29
+ parameters:
30
+ density: [1, 0.8, 0.6]
31
+ weight: 0.2
32
+ - model: ./modela/Infinitely-Laydiculous-7B
33
+ parameters:
34
+ density: [0.9, 0.7, 0.5]
35
+ weight: 0.2
36
+ - model: ./modela/Kunocchini-7b-128k-test
37
+ parameters:
38
+ density: [0.8, 0.6, 0.4]
39
+ weight: 0.2
40
+ - model: ./modela/EndlessRP-v3-7B
41
+ parameters:
42
+ density: [0.7, 0.5, 0.3]
43
+ weight: 0.2
44
+ - model: ./modela/daybreak-kunoichi-2dpo-7b
45
+ parameters:
46
+ density: [0.5, 0.3, 0.1]
47
+ weight: 0.2
48
+ merge_method: dare_linear
49
+ base_model: ./modela/Mistral-7B-v0.1
50
+ parameters:
51
+ normalize: true
52
+ int8_mask: true
53
+ dtype: bfloat16
54
+ name: intermediate-model
55
+ ---
56
+ slices:
57
+ - sources:
58
+ - model: intermediate-model
59
+ layer_range: [0, 32]
60
+ - model: ./modela/BuRP_7B
61
+ layer_range: [0, 32]
62
+ merge_method: slerp
63
+ base_model: intermediate-model
64
+ parameters:
65
+ t:
66
+ - filter: self_attn
67
+ value: [0, 0.5, 0.3, 0.7, 1]
68
+ - filter: mlp
69
+ value: [1, 0.5, 0.7, 0.3, 0]
70
+ - value: 0.5 # fallback for rest of tensors
71
+ dtype: bfloat16
72
+ name: gradient-slerp
73
+ ```
74
+
75
+ ```mergekit-mega config.yml ./output-model-directory --cuda --allow-crimes --lazy-unpickle```
76
+
77
+ ### Models Merged
78
+
79
+ The following models were included in the merge:
80
+ - [localfultonextractor/Erosumika-7B](https://huggingface.co/localfultonextractor/Erosumika-7B)
81
+ - [Nitral-AI/Infinitely-Laydiculous-7B](https://huggingface.co/Nitral-AI/Infinitely-Laydiculous-7B)
82
+ - [Kunocchini-7b-128k-test](https://huggingface.co/Nitral-AI/Kunocchini-7b-128k-test)
83
+ - [Endevor/EndlessRP-v3-7B](https://huggingface.co/Endevor/EndlessRP-v3-7B)
84
+ - [ChaoticNeutrals/BuRP_7B](https://huggingface.co/ChaoticNeutrals/BuRP_7B)
85
+ - [daybreak-kunoichi-2dpo-7b](https://huggingface.co/crestf411/daybreak-kunoichi-2dpo-7b)