import gradio as gr from llama_cpp import Llama model = Llama.from_pretrained(repo_id="google/gemma-7b-it-GGUF", filename="gemma-7b-it.gguf") def chat(input_text): output = model(f"user\n{input_text}\nmodel\n", stop=["model", "\n"]) return output["choices"][0]["text"] interface = gr.Interface( fn=chat, inputs="text", outputs="text" ) interface.launch()