AnishHF commited on
Commit
3eb757e
1 Parent(s): da2c202

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -6
app.py CHANGED
@@ -2,15 +2,11 @@ import os
2
  import bitsandbytes as bnb
3
  import torch
4
  import gradio as gr
5
- from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
6
 
7
  access_token = os.environ["GATED_ACCESS_TOKEN"]
8
 
9
- quantization_config = BitsAndBytesConfig(
10
- load_in_4bit=True,
11
- bnb_4bit_quant_type="nf4",
12
- bnb_4bit_compute_dtype="float16",
13
- )
14
 
15
  model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1", quantization_config=quantization_config, device_map="auto", token=access_token)
16
  tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")
 
2
  import bitsandbytes as bnb
3
  import torch
4
  import gradio as gr
5
+ from transformers import AutoTokenizer, AutoModelForCausalLM, HqqConfig
6
 
7
  access_token = os.environ["GATED_ACCESS_TOKEN"]
8
 
9
+ quantization_config = HqqConfig()
 
 
 
 
10
 
11
  model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1", quantization_config=quantization_config, device_map="auto", token=access_token)
12
  tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")