Spaces:
Running
on
Zero
Running
on
Zero
CaioXapelaum
commited on
Commit
•
0b834b2
1
Parent(s):
1e8b1b7
Update app.py
Browse files
app.py
CHANGED
@@ -97,6 +97,13 @@ hf_hub_download(
|
|
97 |
token=huggingface_token
|
98 |
)
|
99 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
100 |
|
101 |
|
102 |
llm = None
|
@@ -193,7 +200,8 @@ demo = gr.ChatInterface(
|
|
193 |
'smollm-135m-instruct-add-basics-q8_0.gguf',
|
194 |
'Llama-3.2-1B-Instruct-Q8_0.gguf',
|
195 |
'Mistral-Small-Instruct-2409-Q5_K_M.gguf',
|
196 |
-
'Llama-3.1-SuperNova-Lite-Q5_K_M.gguf'
|
|
|
197 |
],
|
198 |
value="gemma-2-2b-it-Q6_K_L.gguf",
|
199 |
label="Model"
|
|
|
97 |
token=huggingface_token
|
98 |
)
|
99 |
|
100 |
+
hf_hub_download(
|
101 |
+
repo_id="TheBloke/MixtralOrochi8x7B-GGUF",
|
102 |
+
filename="mixtralorochi8x7b.Q4_K_M.gguf",
|
103 |
+
local_dir="./models",
|
104 |
+
token=huggingface_token
|
105 |
+
)
|
106 |
+
|
107 |
|
108 |
|
109 |
llm = None
|
|
|
200 |
'smollm-135m-instruct-add-basics-q8_0.gguf',
|
201 |
'Llama-3.2-1B-Instruct-Q8_0.gguf',
|
202 |
'Mistral-Small-Instruct-2409-Q5_K_M.gguf',
|
203 |
+
'Llama-3.1-SuperNova-Lite-Q5_K_M.gguf',
|
204 |
+
'mixtralorochi8x7b.Q4_K_M.gguf'
|
205 |
],
|
206 |
value="gemma-2-2b-it-Q6_K_L.gguf",
|
207 |
label="Model"
|