mlabonne commited on
Commit
2822c14
1 Parent(s): 30470fd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -5
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  base_model:
3
- - NousResearch/Hermes-3-Llama-3.1-8B
4
  library_name: transformers
5
  tags:
6
  - mergekit
@@ -8,15 +8,28 @@ tags:
8
  - autoquant
9
  - gguf
10
  ---
11
- # Hermes-3-Llama-3.1-8B-lorablated
12
 
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
 
15
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [NousResearch/Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) + Llama-3.1-8B-Instruct-abliterated-LORA as a base.
 
 
 
 
 
 
16
 
 
 
 
 
 
 
17
 
18
  ## 🧩 Configuration
19
 
 
 
20
  The following YAML configuration was used to produce this model:
21
 
22
  ```yaml
@@ -31,4 +44,4 @@ slices:
31
  model: NousResearch/Hermes-3-Llama-3.1-8B+Llama-3.1-8B-Instruct-abliterated-LORA
32
  parameters:
33
  weight: 1.0
34
- ```
 
1
  ---
2
  base_model:
3
+ - mlabonne/Hermes-3-Llama-3.1-8B-lorablated
4
  library_name: transformers
5
  tags:
6
  - mergekit
 
8
  - autoquant
9
  - gguf
10
  ---
 
11
 
12
+ # 🪽 Hermes-3-Llama-3.1-8B-lorablated
13
 
14
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/4Hbw5n68jKUSBQeTqQIeT.png)
15
+
16
+ This is an uncensored version of [NousResearch/Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) using lorablation.
17
+
18
+ You can see in the following example how Hermes 3 refuses to answer a legitimate question while the abliterated model complies:
19
+
20
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/GM0WXDXmUU1dRJdYi-Py0.png)
21
 
22
+ The recipe is simple:
23
+
24
+ 1. **Extraction**: We extract a LoRA adapter by comparing two models: a censored Llama 3.1 ([meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)) and an abliterated Llama 3.1 ([mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated](https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated)).
25
+ 2. **Merge**: We merge this new LoRA adapter using [task arithmetic](https://arxiv.org/abs/2212.04089) to the censored [NousResearch/Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) to abliterate it.
26
+
27
+ See [this article](https://huggingface.co/blog/mlabonne/abliteration) to learn more about abliteration.
28
 
29
  ## 🧩 Configuration
30
 
31
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [NousResearch/Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) + Llama-3.1-8B-Instruct-abliterated-LORA as a base.
32
+
33
  The following YAML configuration was used to produce this model:
34
 
35
  ```yaml
 
44
  model: NousResearch/Hermes-3-Llama-3.1-8B+Llama-3.1-8B-Instruct-abliterated-LORA
45
  parameters:
46
  weight: 1.0
47
+ ```