Severian commited on
Commit
696841c
1 Parent(s): 5839008

Update llm_handler.py

Browse files
Files changed (1) hide show
  1. llm_handler.py +30 -1
llm_handler.py CHANGED
@@ -22,10 +22,39 @@ settings = provider.get_provider_default_settings()
22
  settings.max_tokens = 2000
23
  settings.stream = True
24
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  def send_to_llm(provider, msg_list):
26
  try:
27
- response = agent.get_chat_response(msg_list, settings=settings)
 
 
 
 
28
  return response.content, None # We don't have usage info in this case
29
  except Exception as e:
30
  print(f"Error in send_to_llm: {str(e)}")
31
  return f"Error: {str(e)}", None
 
 
22
  settings.max_tokens = 2000
23
  settings.stream = True
24
 
25
+ from llama_cpp import Llama
26
+ from llama_cpp_agent import LlamaCppAgent
27
+ from llama_cpp_agent import MessagesFormatterType
28
+ from llama_cpp_agent.providers import LlamaCppPythonProvider
29
+
30
+ # Initialize the Llama model
31
+ llama_model = Llama("Arcee-Spark-GGUF/Arcee-Spark-Q4_K_M.gguf", n_batch=1024, n_threads=10, n_gpu_layers=33, n_ctx=2048, verbose=False)
32
+
33
+ # Create the provider
34
+ provider = LlamaCppPythonProvider(llama_model)
35
+
36
+ # Create the agent
37
+ agent = LlamaCppAgent(
38
+ provider,
39
+ system_prompt="You are a helpful assistant.",
40
+ predefined_messages_formatter_type=MessagesFormatterType.CHATML,
41
+ debug_output=True
42
+ )
43
+
44
+ # Set provider settings
45
+ settings = provider.get_provider_default_settings()
46
+ settings.max_tokens = 2000
47
+ settings.stream = True
48
+
49
  def send_to_llm(provider, msg_list):
50
  try:
51
+
52
+ provider.apply_settings(settings)
53
+
54
+ # Call get_chat_response without the settings parameter
55
+ response = agent.get_chat_response(msg_list)
56
  return response.content, None # We don't have usage info in this case
57
  except Exception as e:
58
  print(f"Error in send_to_llm: {str(e)}")
59
  return f"Error: {str(e)}", None
60
+