benk04 commited on
Commit
d94dc1d
1 Parent(s): e8a2040

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -13,12 +13,13 @@ license: cc-by-nc-4.0
13
  ---
14
 
15
  <!-- description start -->
16
- My Exllamav2 3.75 bpw quantization of [NoromaidxOpenGPT4-2](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-2), quantized with default calibration dataset. Included is measurement json, so you can do your own quants.
17
  > [!IMPORTANT]
18
- >This bpw is the perfect size for 24GB cards, and can fit 32k context. Make sure to enable 4-bit cache option.
19
 
20
- >[!NOTE]
21
- > This model is great for rp and I recommend using the Alpaca presets in SillyTavern.
 
22
 
23
  ## Original Card
24
  ## Description
 
13
  ---
14
 
15
  <!-- description start -->
16
+ Exllamav2 3.75bpw quantization of NoromaidxOpenGPT4-2 from [NeverSleep](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-2), quantized with default calibration dataset. Included is measurement json file, so you can do your own quants.
17
  > [!IMPORTANT]
18
+ >This bpw is the perfect size for 24GB GPUs, and can fit 32k context. Make sure to enable 4-bit cache option or you'll run into OOM errors.
19
 
20
+ > [!NOTE]
21
+ > **Notes:**
22
+ > This model is one of the better mixtral derivatives for rp, and I recommend using it with the Alpaca preset in SillyTavern.
23
 
24
  ## Original Card
25
  ## Description