OOM when trying to load the Q8 in Oobabooga (I have 98GB VRAM)

#1
by AIGUYCONTENT - opened

As the title says...

for the: llama-3-70B-Instruct-abliterated-Q8_0-00001-of-00003.gguf

I downloaded all three files and placed into /models folder in oobabooga. I did not concatenate.

I have 98GB of VRAM (4090, 4080, 3090, 3090, 3080) on a SuperMicro server and EPYC 7F52 32GB 3200DDR4 RAM.

I am NOT trying to load the model into RAM or CPU. GPU only. I'm having issues running EXL2 files for some strange reason. Getting error message with that. So, this is why I'm trying to load a huge Q8 into my available VRAM.

I'm getting an OOM error within a second or two of trying to load the file.

I'm also seeing this error message: AttributeError: 'LlamaCppModel' object has no attribute 'model'

This is a new server build. However, everything is updated (CUDA etc) and other GGUF files work.

Tensor_split is: 20,20,7,14,20.

I have the following checked when loading the model:

  • flash_attn

  • tensorcores

  • no-mmap

  • n-gpu-layers: 81

  • n_ctx: 5120

  • n_batch 512

  • threads: 16

  • threads_batch: 32

  • rope_freq_base: 500000

Here is from terminal:

ggml_backend_cuda_buffer_type_alloc_buffer: allocating 17341.25 MiB on device 1: cudaMalloc failed: out of memory
llama_model_load: error loading model: unable to allocate backend buffer
llama_load_model_from_file: failed to load model
13:12:15-377410 ERROR Failed to load the model.


~$ nvidia-smi
Sat Jun 29 13:19:09 2024
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 535.183.01 Driver Version: 535.183.01 CUDA Version: 12.2 |
|-----------------------------------------+----------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 NVIDIA GeForce RTX 3090 Off | 00000000:01:00.0 Off | N/A |
| 81% 46C P0 110W / 350W | 282MiB / 24576MiB | 3% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 1 NVIDIA GeForce RTX 4090 Off | 00000000:81:00.0 On | Off |
| 79% 38C P2 81W / 450W | 479MiB / 24564MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 2 NVIDIA GeForce RTX 3080 Off | 00000000:82:00.0 Off | N/A |
| 74% 37C P0 98W / 380W | 12MiB / 10240MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 3 NVIDIA GeForce RTX 4080 Off | 00000000:C1:00.0 Off | N/A |
| 75% 35C P2 51W / 340W | 257MiB / 16376MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 4 NVIDIA GeForce RTX 3090 Off | 00000000:C2:00.0 Off | N/A |
| 74% 43C P0 110W / 350W | 12MiB / 24576MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+

+---------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=======================================================================================|
| 0 N/A N/A 2467 G /usr/lib/xorg/Xorg 144MiB |
| 0 N/A N/A 2664 G /usr/bin/gnome-shell 43MiB |
| 0 N/A N/A 4983 G /usr/bin/nvidia-settings 0MiB |
| 0 N/A N/A 24058 G ...onEnabled --variations-seed-version 76MiB |
| 1 N/A N/A 2467 G /usr/lib/xorg/Xorg 45MiB |
| 1 N/A N/A 25365 C /usr/NX/bin/nxnode.bin 419MiB |
| 2 N/A N/A 2467 G /usr/lib/xorg/Xorg 4MiB |
| 3 N/A N/A 2467 G /usr/lib/xorg/Xorg 4MiB |
| 3 N/A N/A 2615 C+G ...libexec/gnome-remote-desktop-daemon 239MiB |
| 4 N/A N/A 2467 G /usr/lib/xorg/Xorg 4MiB |
+---------------------------------------------------------------------------------------+

Thanks if you can help.

BTW, the error message I get when trying to load EXL2 files is: UserWarning: do_sample is set to False. However, min_p is set to 0.0 -- this flag is only used in sample-based generation modes. You should set do_sample=True or unset min_p.
warnings.warn(

This is no matter what EXL2 file I try to load.

that's a tricky one.. your tensor split may be causing the issue, can you try raising it a bit like 22,22,8,14,22? As it stands you've only split it with 81GB, whereas the unloaded file is already 75GB. pretty sure there's overhead for splitting and the of course there's the context. Do you see the same if you go for like Q6_K?

that's a tricky one.. your tensor split may be causing the issue, can you try raising it a bit like 22,22,8,14,22? As it stands you've only split it with 81GB, whereas the unloaded file is already 75GB. pretty sure there's overhead for splitting and the of course there's the context. Do you see the same if you go for like Q6_K?

I just downloaded the Q6. Had to lower n_gpu_layers to 55 before I could get it to load--and run at .96/tokens per second.

For comparison, I was able to run 81/81 layers offloaded most of your q5/q6 quants with just a 4090 and two 3090s....on my gaming pc. ~10 tokens per second.

I wonder if this is something to do with OobaBooga and how it deals with multi gpus. One of the 3090 only has 19% user dedicated memory.

Also noticed this message in terminal the second I loaded the LLamam3 70b Instruct Abliterated or the Higgs GGUF:

"Detected duplicate leading "<|begin_of_text|>" in prompt, this will likely reduce response quality, consider removing it..."

I might just scrap this Oobabooga installation and re-install.

Do you know of a good reccomended front end for Aphrodite? I was told it handles multi gpu much better. Was unable to get it working with Kobold Lite last night (had to get chatgpt to help).

Thanks for all your hard work!!!

When my system was running good (ironically on my gaming pc and not the purpose-built server I put together last week)---Your Q5_K_M Llama3 70B Instruct Abliterated quant was my go-to quant for work!

Sign up or log in to comment