Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -8,14 +8,15 @@ import os
|
|
8 |
offload_dir = './offload'
|
9 |
os.makedirs(offload_dir, exist_ok=True)
|
10 |
|
|
|
11 |
model = AutoModelForCausalLM.from_pretrained(
|
12 |
-
"
|
13 |
torch_dtype=torch.bfloat16,
|
14 |
device_map="auto",
|
15 |
low_cpu_mem_usage=True,
|
16 |
offload_folder=offload_dir
|
17 |
)
|
18 |
-
tokenizer = AutoTokenizer.from_pretrained("tiiuae/
|
19 |
|
20 |
|
21 |
def generate_text(input_text):
|
@@ -46,5 +47,5 @@ text_generation_interface = gr.Interface(
|
|
46 |
gr.inputs.Textbox(label="Input Text"),
|
47 |
],
|
48 |
outputs=gr.inputs.Textbox(label="Generated Text"),
|
49 |
-
title="
|
50 |
).launch()
|
|
|
8 |
offload_dir = './offload'
|
9 |
os.makedirs(offload_dir, exist_ok=True)
|
10 |
|
11 |
+
#"tiiuae/falcon-7b-instruct",
|
12 |
model = AutoModelForCausalLM.from_pretrained(
|
13 |
+
"meta-llama/Meta-Llama-3-8B",
|
14 |
torch_dtype=torch.bfloat16,
|
15 |
device_map="auto",
|
16 |
low_cpu_mem_usage=True,
|
17 |
offload_folder=offload_dir
|
18 |
)
|
19 |
+
tokenizer = AutoTokenizer.from_pretrained("tiiuae/meta-llama/Meta-Llama-3-8B")
|
20 |
|
21 |
|
22 |
def generate_text(input_text):
|
|
|
47 |
gr.inputs.Textbox(label="Input Text"),
|
48 |
],
|
49 |
outputs=gr.inputs.Textbox(label="Generated Text"),
|
50 |
+
title="---LLM---",
|
51 |
).launch()
|