jukofyork commited on
Commit
69d01ef
1 Parent(s): aa6cba0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -6
README.md CHANGED
@@ -7,8 +7,6 @@ tags:
7
 
8
  **NOTE**: See [creative-writing-control-vectors-v2.1](https://huggingface.co/jukofyork/creative-writing-control-vectors-v2.1) for the current main control-vector repo.
9
 
10
- *(I will also add `WizardLM-2-8x22B`, `c4ai-command-r-v01` and `gemma-2-27b-it` versions of these control vectors in the next few days...)*
11
-
12
  ## Details
13
 
14
  The control-vectors in this repo were created as an experiment by increasing the triplets in `system_messages_outlook_extended.json` by 4x (click to expand):
@@ -160,16 +158,16 @@ I also include 3 different values for the `--regularisation_factor` option; `1.0
160
 
161
  Try to use the largest `regularisation_factor` that has the desired effect - this has the least chance of damaging the models' outputs.
162
 
163
- ## Prompting format for `Mistral-Large-Instruct-2407` and `WizardLM-2-8x22B`:
164
 
165
- I have found by testing that `Mistral-Large-Instruct-2407` seems to work much better for creative writing if you use the following 'Vicuna' prompt template:
166
 
167
  ```
168
  USER: {prompt}
169
  ASSISTANT:
170
  ```
171
 
172
- so I altered the 'Jinja2' `chat_template` in the `tokenizer_config.json` for both `Mistral-Large-Instruct-2407` and `WizardLM-2-8x22B` to this for the training of these control vectors:
173
 
174
  ```json
175
  {
@@ -177,4 +175,4 @@ so I altered the 'Jinja2' `chat_template` in the `tokenizer_config.json` for bot
177
  }
178
  ```
179
 
180
- The other 3 models still use their default templates though.
 
7
 
8
  **NOTE**: See [creative-writing-control-vectors-v2.1](https://huggingface.co/jukofyork/creative-writing-control-vectors-v2.1) for the current main control-vector repo.
9
 
 
 
10
  ## Details
11
 
12
  The control-vectors in this repo were created as an experiment by increasing the triplets in `system_messages_outlook_extended.json` by 4x (click to expand):
 
158
 
159
  Try to use the largest `regularisation_factor` that has the desired effect - this has the least chance of damaging the models' outputs.
160
 
161
+ ## Prompting format for `Mistral-Large-Instruct-2407`, `WizardLM-2-8x22B` and `miqu-1-70b`:
162
 
163
+ I have found by testing that these models seems to work much better for creative writing if you use the following 'Vicuna' prompt template:
164
 
165
  ```
166
  USER: {prompt}
167
  ASSISTANT:
168
  ```
169
 
170
+ so I altered the 'Jinja2' `chat_template` in the `tokenizer_config.json` for `Mistral-Large-Instruct-2407`, `WizardLM-2-8x22B` and `miqu-1-70b` to this for the training of these control vectors:
171
 
172
  ```json
173
  {
 
175
  }
176
  ```
177
 
178
+ **NOTE**: I still used the default prompt templates for the other 3 models (`c4ai-command-r-plus`, `c4ai-command-r-v01` and `gemma-2-27b-it`).