Spaces:
Runtime error
Runtime error
import requests | |
import torch | |
from PIL import Image | |
from transformers import pipeline | |
pipe = pipeline("text-generation", model="meta-llama/Llama-3.2-1B") | |
# Load model directly | |
from transformers import AutoTokenizer, AutoModelForCausalLM | |
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-1B") | |
model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-1B") | |
# Load the processor | |
processor = AutoProcessor.from_pretrained(model_id) | |
# Define the function to generate text based on input prompt | |
def generate_text(prompt): | |
if llm_pipeline is None: | |
return "Error: Model not loaded." | |
result = llm_pipeline(prompt, max_length=100, num_return_sequences=1) | |
return result[0]['generated_text'] | |
# Create the Gradio interface | |
interface = gr.Interface( | |
fn=generate_text, | |
inputs=gr.Textbox(lines=7, label="Input Prompt"), | |
outputs="text", | |
title="Large Language Model Text Generation", | |
description="Enter a prompt to generate text using a large language model." | |
) | |
print("Launching the Gradio interface...") | |
# Launch the interface | |
interface.launch() | |