gokaygokay commited on
Commit
5c7ec8d
1 Parent(s): 57a59dc
Files changed (2) hide show
  1. app.py +3 -1
  2. llm_inference.py +16 -0
app.py CHANGED
@@ -106,11 +106,13 @@ def create_interface():
106
  # Generate Text with LLM
107
  def generate_text_with_llm(output_prompt, long_talk, compress, compression_level, custom_base_prompt, provider, api_key, model_selected):
108
  global selected_prompt_type
 
109
  result = llm_node.generate(
110
  input_text=output_prompt,
111
  long_talk=long_talk,
112
  compress=compress,
113
  compression_level=compression_level,
 
114
  prompt_type=selected_prompt_type,
115
  custom_base_prompt=custom_base_prompt,
116
  provider=provider,
@@ -131,4 +133,4 @@ def create_interface():
131
 
132
  if __name__ == "__main__":
133
  demo = create_interface()
134
- demo.launch()
 
106
  # Generate Text with LLM
107
  def generate_text_with_llm(output_prompt, long_talk, compress, compression_level, custom_base_prompt, provider, api_key, model_selected):
108
  global selected_prompt_type
109
+ poster = False # Set a default value or modify as needed
110
  result = llm_node.generate(
111
  input_text=output_prompt,
112
  long_talk=long_talk,
113
  compress=compress,
114
  compression_level=compression_level,
115
+ poster=poster, # Added the missing 'poster' argument
116
  prompt_type=selected_prompt_type,
117
  custom_base_prompt=custom_base_prompt,
118
  provider=provider,
 
133
 
134
  if __name__ == "__main__":
135
  demo = create_interface()
136
+ demo.launch(share=True)
llm_inference.py CHANGED
@@ -18,6 +18,22 @@ class LLMInferenceNode:
18
  base_url="https://api.sambanova.ai/v1",
19
  )
20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  def generate(
22
  self,
23
  input_text,
 
18
  base_url="https://api.sambanova.ai/v1",
19
  )
20
 
21
+ def generate_prompt(self, dynamic_seed, prompt_type, custom_input):
22
+ """
23
+ Generates a prompt based on the provided seed, prompt type, and custom input.
24
+ """
25
+ import random
26
+
27
+ random.seed(dynamic_seed)
28
+ if custom_input and custom_input.strip():
29
+ prompt = custom_input
30
+ else:
31
+ prompt = f"Generate a {prompt_type.lower()} prompt."
32
+
33
+ # Additional logic can be added here if needed
34
+ print(f"Generated prompt: {prompt}") # Debug statement
35
+ return prompt
36
+
37
  def generate(
38
  self,
39
  input_text,