Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -5,8 +5,7 @@ import torch
|
|
5 |
import re
|
6 |
import random
|
7 |
|
8 |
-
|
9 |
-
|
10 |
|
11 |
from pathlib import Path
|
12 |
from huggingface_hub import CommitScheduler
|
@@ -18,7 +17,7 @@ JSON_DATASET_DIR.mkdir(parents=True, exist_ok=True)
|
|
18 |
JSON_DATASET_PATH = JSON_DATASET_DIR / f"feedbacks.json"
|
19 |
|
20 |
scheduler = CommitScheduler(
|
21 |
-
repo_id="bmi-labmedinfo/
|
22 |
repo_type="dataset",
|
23 |
folder_path=JSON_DATASET_DIR,
|
24 |
path_in_repo="data",
|
@@ -31,8 +30,7 @@ def save_json(last_state: dict, pos_or_neg: str) -> None:
|
|
31 |
json.dump(last_state, f)
|
32 |
f.write("\n")
|
33 |
|
34 |
-
|
35 |
-
|
36 |
|
37 |
# Initialize the model
|
38 |
model = AutoModelForCausalLM.from_pretrained("Detsutut/Igea-1B-instruct-v0.3-test4epochs-GGUF", model_file="unsloth.Q4_K_M.gguf", model_type="mistral", hf=True)
|
@@ -93,14 +91,12 @@ def generate_text(input_text, max_new_tokens=512, temperature=1, system_prompt="
|
|
93 |
return f"<span>{input_text}</span><b style='color: {generated_text_color};'>{generated_text}</b>", {"input_prompt":prompt, "generated_text_raw":output[0]['generated_text'], "generated_text_displayed":generated_text}
|
94 |
|
95 |
def positive_feedback(last_generated_text):
|
96 |
-
print("positive")
|
97 |
-
print(last_generated_text)
|
98 |
save_json(last_generated_text,"positive")
|
|
|
99 |
|
100 |
def negative_feedback(last_generated_text):
|
101 |
-
print("negative")
|
102 |
-
print(last_generated_text)
|
103 |
save_json(last_generated_text,"negative")
|
|
|
104 |
|
105 |
|
106 |
# Create the Gradio interface
|
@@ -125,10 +121,10 @@ with gr.Blocks(css="#outbox { border-radius: 8px !important; border: 1px solid #
|
|
125 |
|
126 |
btn = gr.Button("Generate")
|
127 |
btn.click(generate_text, [input_text, max_new_tokens, temperature, system_prompt], outputs=[output, last_generated_text])
|
128 |
-
|
129 |
-
|
|
|
130 |
btn_p.click(positive_feedback, inputs=[last_generated_text], outputs=None)
|
131 |
-
btn_n = gr.Button("👎")
|
132 |
btn_n.click(negative_feedback, inputs=[last_generated_text], outputs=None)
|
133 |
|
134 |
# Launch the interface
|
|
|
5 |
import re
|
6 |
import random
|
7 |
|
8 |
+
### FEEDBACKS UPDATE IN PERSISTENT MEMORY
|
|
|
9 |
|
10 |
from pathlib import Path
|
11 |
from huggingface_hub import CommitScheduler
|
|
|
17 |
JSON_DATASET_PATH = JSON_DATASET_DIR / f"feedbacks.json"
|
18 |
|
19 |
scheduler = CommitScheduler(
|
20 |
+
repo_id="bmi-labmedinfo/feedbacks",
|
21 |
repo_type="dataset",
|
22 |
folder_path=JSON_DATASET_DIR,
|
23 |
path_in_repo="data",
|
|
|
30 |
json.dump(last_state, f)
|
31 |
f.write("\n")
|
32 |
|
33 |
+
### /FEEDBACKS
|
|
|
34 |
|
35 |
# Initialize the model
|
36 |
model = AutoModelForCausalLM.from_pretrained("Detsutut/Igea-1B-instruct-v0.3-test4epochs-GGUF", model_file="unsloth.Q4_K_M.gguf", model_type="mistral", hf=True)
|
|
|
91 |
return f"<span>{input_text}</span><b style='color: {generated_text_color};'>{generated_text}</b>", {"input_prompt":prompt, "generated_text_raw":output[0]['generated_text'], "generated_text_displayed":generated_text}
|
92 |
|
93 |
def positive_feedback(last_generated_text):
|
|
|
|
|
94 |
save_json(last_generated_text,"positive")
|
95 |
+
gr.Info("Feedback collected. Thanks!")
|
96 |
|
97 |
def negative_feedback(last_generated_text):
|
|
|
|
|
98 |
save_json(last_generated_text,"negative")
|
99 |
+
gr.Info("Feedback collected. Thanks!")
|
100 |
|
101 |
|
102 |
# Create the Gradio interface
|
|
|
121 |
|
122 |
btn = gr.Button("Generate")
|
123 |
btn.click(generate_text, [input_text, max_new_tokens, temperature, system_prompt], outputs=[output, last_generated_text])
|
124 |
+
with gr.Row():
|
125 |
+
btn_p = gr.Button("👍")
|
126 |
+
btn_n = gr.Button("👎")
|
127 |
btn_p.click(positive_feedback, inputs=[last_generated_text], outputs=None)
|
|
|
128 |
btn_n.click(negative_feedback, inputs=[last_generated_text], outputs=None)
|
129 |
|
130 |
# Launch the interface
|