import streamlit as st from transformers import AutoTokenizer, AutoModelForCausalLM import torch st.title("Uzmi GPT - Romantic Quote Generator") @st.cache_resource def load_model(): tokenizer = AutoTokenizer.from_pretrained("rajan3208/uzmi-gpt") model = AutoModelForCausalLM.from_pretrained("rajan3208/uzmi-gpt") return tokenizer, model tokenizer, model = load_model() prompt = st.text_area("Enter a prompt", "A romantic quote about forever") if st.button("Generate"): inputs = tokenizer(prompt, return_tensors="pt") output = model.generate(**inputs, max_new_tokens=50) generated = tokenizer.decode(output[0], skip_special_tokens=True) st.success(generated)