Detsutut commited on
Commit
ae390c9
1 Parent(s): 99ed14f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +31 -1
app.py CHANGED
@@ -1,15 +1,43 @@
1
  import gradio as gr
2
  from ctransformers import AutoModelForCausalLM
3
  from transformers import AutoTokenizer, pipeline
 
4
  import torch
5
  import re
6
  import random
7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  # Initialize the model
9
  model = AutoModelForCausalLM.from_pretrained("Detsutut/Igea-1B-instruct-v0.3-test4epochs-GGUF", model_file="unsloth.Q4_K_M.gguf", model_type="mistral", hf=True)
10
  tokenizer = AutoTokenizer.from_pretrained( "Detsutut/Igea-1B-instruct-v0.3-test4epochs")
11
 
12
-
13
  gen_pipeline = pipeline(
14
  "text-generation",
15
  model=model,
@@ -67,10 +95,12 @@ def generate_text(input_text, max_new_tokens=512, temperature=1, system_prompt="
67
  def positive_feedback(last_generated_text):
68
  print("positive")
69
  print(last_generated_text)
 
70
 
71
  def negative_feedback(last_generated_text):
72
  print("negative")
73
  print(last_generated_text)
 
74
 
75
 
76
  # Create the Gradio interface
 
1
  import gradio as gr
2
  from ctransformers import AutoModelForCausalLM
3
  from transformers import AutoTokenizer, pipeline
4
+ import datasets
5
  import torch
6
  import re
7
  import random
8
 
9
+
10
+
11
+
12
+ from pathlib import Path
13
+ from huggingface_hub import CommitScheduler
14
+
15
+ JSON_DATASET_DIR = Path("json_dataset")
16
+ JSON_DATASET_DIR.mkdir(parents=True, exist_ok=True)
17
+
18
+ JSON_DATASET_PATH = JSON_DATASET_DIR / f"feedbacks.json"
19
+
20
+ scheduler = CommitScheduler(
21
+ repo_id="Detsutut/feedbacks_test",
22
+ repo_type="dataset",
23
+ folder_path=JSON_DATASET_DIR,
24
+ path_in_repo="data",
25
+ )
26
+
27
+ def save_json(last_state: dict, pos_or_neg: str) -> None:
28
+ last_state["feedback"]=pos_or_neg
29
+ with scheduler.lock:
30
+ with JSON_DATASET_PATH.open("a") as f:
31
+ json.dump(last_state, f)
32
+ f.write("\n")
33
+
34
+
35
+
36
+
37
  # Initialize the model
38
  model = AutoModelForCausalLM.from_pretrained("Detsutut/Igea-1B-instruct-v0.3-test4epochs-GGUF", model_file="unsloth.Q4_K_M.gguf", model_type="mistral", hf=True)
39
  tokenizer = AutoTokenizer.from_pretrained( "Detsutut/Igea-1B-instruct-v0.3-test4epochs")
40
 
 
41
  gen_pipeline = pipeline(
42
  "text-generation",
43
  model=model,
 
95
  def positive_feedback(last_generated_text):
96
  print("positive")
97
  print(last_generated_text)
98
+ save_json(last_generated_text,"positive")
99
 
100
  def negative_feedback(last_generated_text):
101
  print("negative")
102
  print(last_generated_text)
103
+ save_json(last_generated_text,"negative")
104
 
105
 
106
  # Create the Gradio interface