vilarin commited on
Commit
3a4e50e
1 Parent(s): 5b22fb1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -44,7 +44,7 @@ model_code = AutoModelForCausalLM.from_pretrained(
44
  tokenizer_code = AutoTokenizer.from_pretrained("THUDM/codegeex4-all-9b", trust_remote_code=True)
45
 
46
  @spaces.GPU
47
- def stream_chat(message: str, history: list, temperature: float, max_length: int, model: str):
48
  print(f'message is - {message}')
49
  print(f'history is - {history}')
50
  conversation = []
@@ -54,7 +54,7 @@ def stream_chat(message: str, history: list, temperature: float, max_length: int
54
 
55
  print(f"Conversation is -\n{conversation}")
56
 
57
- if mode == "glm-4-9b-chat":
58
  tokenizer = tokenizer_chat
59
  model = model_chat
60
  else:
@@ -111,7 +111,8 @@ with gr.Blocks(css=CSS) as demo:
111
  ),
112
  gr.Radio(
113
  ["glm-4-9b-chat", "codegeex4-all-9b"],
114
- label="Load Model"
 
115
  ),
116
  ],
117
  examples=[
 
44
  tokenizer_code = AutoTokenizer.from_pretrained("THUDM/codegeex4-all-9b", trust_remote_code=True)
45
 
46
  @spaces.GPU
47
+ def stream_chat(message: str, history: list, temperature: float, max_length: int, choice: str):
48
  print(f'message is - {message}')
49
  print(f'history is - {history}')
50
  conversation = []
 
54
 
55
  print(f"Conversation is -\n{conversation}")
56
 
57
+ if choice == "glm-4-9b-chat":
58
  tokenizer = tokenizer_chat
59
  model = model_chat
60
  else:
 
111
  ),
112
  gr.Radio(
113
  ["glm-4-9b-chat", "codegeex4-all-9b"],
114
+ label="Load Model",
115
+ render=False,
116
  ),
117
  ],
118
  examples=[