Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -2,15 +2,11 @@ import os
|
|
2 |
import bitsandbytes as bnb
|
3 |
import torch
|
4 |
import gradio as gr
|
5 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM,
|
6 |
|
7 |
access_token = os.environ["GATED_ACCESS_TOKEN"]
|
8 |
|
9 |
-
quantization_config =
|
10 |
-
load_in_4bit=True,
|
11 |
-
bnb_4bit_quant_type="nf4",
|
12 |
-
bnb_4bit_compute_dtype="float16",
|
13 |
-
)
|
14 |
|
15 |
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1", quantization_config=quantization_config, device_map="auto", token=access_token)
|
16 |
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")
|
|
|
2 |
import bitsandbytes as bnb
|
3 |
import torch
|
4 |
import gradio as gr
|
5 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, HqqConfig
|
6 |
|
7 |
access_token = os.environ["GATED_ACCESS_TOKEN"]
|
8 |
|
9 |
+
quantization_config = HqqConfig()
|
|
|
|
|
|
|
|
|
10 |
|
11 |
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1", quantization_config=quantization_config, device_map="auto", token=access_token)
|
12 |
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")
|