Centaur / app.py
marcelbinz's picture
Update app.py
d42d707 verified
raw
history blame
554 Bytes
import spaces
import gradio as gr
import torch
from transformers import pipeline
pipe = pipeline(
"text-generation",
model="marcelbinz/Llama-3.1-Minitaur-8B",
device_map="auto",
torch_dtype=torch.bfloat16
)
@spaces.GPU
def infer(prompt):
return pipe(prompt, max_new_tokens=1, do_sample=True, temperature=1.0)[0]["generated_text"]
demo = gr.Interface(
fn=infer,
inputs=gr.Textbox(label="Prompt"),
outputs="text",
title="Minitaur",
description="Just type and hit *Run*"
).queue()
demo.launch()