ChuckMcSneed commited on
Commit
b25cc0b
1 Parent(s): 1805f09

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -0
README.md CHANGED
@@ -1,3 +1,26 @@
1
  ---
2
  license: llama2
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: llama2
3
+ tags:
4
+ - merge
5
+ - mergekit
6
  ---
7
+ I merged [Aurelian](https://huggingface.co/grimulkan/aurelian-v0.5-70b-rope8-32K-fp16) with itself using [mergekit](https://github.com/cg123/mergekit), creating this EXTENDED LENGTH FRANKENSTEIN.
8
+ # Does it work
9
+ Yes, at 17k it stays coherent, but starts to lose minor details of the story. Not sure how well it performs at 32k though. Quants have a **sinificant** impact on quality for this model, going from Q6_K to Q5_K had a noticeable drop in quality.
10
+ # Is it worth it
11
+ Maybe? Depends? Do you hate mixtral? Do you have good hardware/patience? Do you need a somewhat smart model with 32k context?
12
+ # Known issues
13
+ VERY strict adherence to prompt format, forgetfullness, strong roleplay bias.
14
+ # Personal opinion
15
+ Dumber than Goliath, but has much less GPTism. If you want 32k goliath, maybe try [Goliath-longLORA-120b-rope8-32k-fp16](https://huggingface.co/grimulkan/Goliath-longLORA-120b-rope8-32k-fp16).
16
+ # Benchmarks
17
+ ### NeoEvalPlusN_benchmark
18
+ [My meme benchmark.](https://huggingface.co/datasets/ChuckMcSneed/NeoEvalPlusN_benchmark)
19
+ | Test name | Aurelian | DoubleGold |
20
+ | ---------- | ---------- | ------- |
21
+ | B | 1 | 1 |
22
+ | C | 1 | 1 |
23
+ | D | 0 | 2 |
24
+ | S | 2.5 | 3.25 |
25
+ | P | 2.25 | 1.5 |
26
+ | Total | 6.75 | 8.75 |