Text Generation
Transformers
Safetensors
English
llama
llama-factory
Not-For-All-Audiences
conversational
text-generation-inference
Inference Endpoints
aaronday3 commited on
Commit
1207ce0
1 Parent(s): 914552d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -29
README.md CHANGED
@@ -54,8 +54,7 @@ First message and last few messages impact this model quite a bit in terms of st
54
 
55
  ## Hornyness
56
 
57
- If the model is not horny enough then just edit the last character message to hint at something, the model will pick up on it and build on it. (Or just give the char aphrodisiac pills lol)
58
-
59
  The model is fine with SFW and doesn't make it NSFW unless you want. It is also able to maintain half-NSFW without devolving down into hardcore.
60
 
61
  ---
@@ -72,24 +71,14 @@ While we did train all system prompts from c2 logs we also have our own system p
72
  <details>
73
  <summary>List of trained system prompts. Note: c2 logs system prompts and char cards were also included.</summary>
74
 
75
- ### reddit_dirty_writing_prompts.jsonl
76
- **2000**
77
- **"You are a short story writer. Write a story based on prompt provided by user below.
78
- Mode: NSFW"**
79
-
80
- ### reddit_writing_prompts.jsonl
81
- **2000**
82
- **"You are a short story writer. Write a story based on prompt provided by user below.
83
- Mode: SFW"**
84
 
85
- ### Opus_Instruct-v2-6.5K-Filtered-v2.jsonl
86
- **2000**
87
- **""** (no prompt)
88
-
89
- ### deduped-c2-logs-maywell-final-filter-4.jsonl
90
- **2000**
91
- (Only if there was no system prompt in the conversation, otherwise keep original system prompt)
92
- **"You are an expert actor that can fully immerse yourself into any role given. You do not break character for any reason, even if someone tries addressing you as an AI or language model."**
93
 
94
  </details>
95
 
@@ -105,20 +94,16 @@ We think there is too much secrecy around what data is being used, and different
105
 
106
  ### The Good
107
  We found that increasing the amount of ranks from 64 to 256 has reduced repetition but also led to the language used resembling Claude more than the 64 rank version. No worries, it's still far enough from Claude.
108
-
109
- It also led to increased coherency but reduced instruction following, likely because the model started diverging more away from L3 8B Instruct.
110
-
111
- **The model is uncensored for RP. For Instruct it needs 2-3 words of prefill for the first message.**
112
-
113
- We found that increasing the amount of data from 1K to 8K reduced repetition aswell.
114
-
115
- **The prose is much better than other synthetic data generations. The model also demonstrates increased style copying abilities likely a result of the long data and varying writing styles found in WritingPrompts.**
116
-
117
- **The model is exceptional at being creative in roleplaying**, knows different persona's and even a single character will change persona in different contexts, persona is tied to last few messages rather than system message or character card. **This is great as it often means the model can do impressive things without you needing to explicitly specify.**
118
 
119
  ### Improvements for Next Run
120
 
121
  Formatting can break sometimes.
 
122
 
123
  ### Comments about training
124
 
 
54
 
55
  ## Hornyness
56
 
57
+ If the model is not horny enough then just edit the last character message to hint at something, the model will pick up on it and build on it. (Or just give the char aphrodisiac pills lol) <br>
 
58
  The model is fine with SFW and doesn't make it NSFW unless you want. It is also able to maintain half-NSFW without devolving down into hardcore.
59
 
60
  ---
 
71
  <details>
72
  <summary>List of trained system prompts. Note: c2 logs system prompts and char cards were also included.</summary>
73
 
74
+ Here's the information in a markdown table:
 
 
 
 
 
 
 
 
75
 
76
+ | Dataset | System Prompt |
77
+ |--------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------|
78
+ | reddit_dirty_writing_prompts.jsonl | "You are a short story writer. Write a story based on prompt provided by user below. Mode: NSFW" |
79
+ | reddit_writing_prompts.jsonl | "You are a short story writer. Write a story based on prompt provided by user below. Mode: SFW" |
80
+ | Opus_Instruct-v2-6.5K-Filtered-v2.jsonl | (blank prompt) |
81
+ | deduped-c2-logs-maywell-final-filter-4.jsonl | (Only if there was no system prompt in the conversation, otherwise keep original system prompt) **"You are an expert actor that can fully immerse yourself into any role given. You do not break character for any reason, even if someone tries addressing you as an AI or language model."** |
 
 
82
 
83
  </details>
84
 
 
94
 
95
  ### The Good
96
  We found that increasing the amount of ranks from 64 to 256 has reduced repetition but also led to the language used resembling Claude more than the 64 rank version. No worries, it's still far enough from Claude.
97
+ <br>It also led to increased coherency but reduced instruction following, likely because the model started diverging more away from L3 8B Instruct.
98
+ <br>**The model is uncensored for RP. For Instruct it needs 2-3 words of prefill for the first message.**
99
+ <br>We found that increasing the amount of data from 1K to 8K reduced repetition aswell.
100
+ <br>**The prose is much better than other synthetic data generations. The model also demonstrates increased style copying abilities likely a result of the long data and varying writing styles found in WritingPrompts.**
101
+ <br>**The model is exceptional at being creative in roleplaying**, knows different persona's and even a single character will change persona in different contexts, persona is tied to last few messages rather than system message or character card. **This is great as it often means the model can do impressive things without you needing to explicitly specify.**
 
 
 
 
 
102
 
103
  ### Improvements for Next Run
104
 
105
  Formatting can break sometimes.
106
+ Repetition can become an issue with certain types of prompts. Removing system helps.
107
 
108
  ### Comments about training
109