AnishHF commited on
Commit
f11657e
1 Parent(s): 9a1eed5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -3
app.py CHANGED
@@ -1,6 +1,5 @@
1
  import os
2
- import bitsandbytes
3
- import accelerate
4
  import torch
5
  import gradio as gr
6
  from transformers import AutoTokenizer, AutoModelForCausalLM
@@ -10,7 +9,12 @@ access_token = os.environ["GATED_ACCESS_TOKEN"]
10
  # Load the tokenizer and model
11
  model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
12
  tokenizer = AutoTokenizer.from_pretrained(model_id, token=access_token)
13
- model = AutoModelForCausalLM.from_pretrained(model_id, load_in_4bit=True, device_map="auto", token=access_token)
 
 
 
 
 
14
 
15
  # Function to generate text using the model
16
  def generate_text(prompt):
 
1
  import os
2
+ import bitsandbytes as bnb
 
3
  import torch
4
  import gradio as gr
5
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
9
  # Load the tokenizer and model
10
  model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
11
  tokenizer = AutoTokenizer.from_pretrained(model_id, token=access_token)
12
+ model = AutoModelForCausalLM.from_pretrained(model_id, token=access_token)
13
+ # Initialize the quantizer
14
+ quantizer = bnb.GemmQuantizer(act_bits=8, weight_bits=8)
15
+
16
+ # Quantize the model
17
+ model = quantizer(model)
18
 
19
  # Function to generate text using the model
20
  def generate_text(prompt):