Spaces:
Sleeping
Sleeping
import gradio as gr | |
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline | |
# SKT 한국어 GPT 모델 | |
model_id = "skt/ko-gpt-trinity-1.2B-v0.5" | |
tokenizer = AutoTokenizer.from_pretrained(model_id) | |
model = AutoModelForCausalLM.from_pretrained(model_id) | |
pipe = pipeline( | |
"text-generation", | |
model=model, | |
tokenizer=tokenizer, | |
max_new_tokens=100, | |
do_sample=True, | |
temperature=0.7, | |
) | |
def chat(prompt): | |
response = pipe(prompt)[0]["generated_text"] | |
return response | |
gr.Interface( | |
fn=chat, | |
inputs="text", | |
outputs="text", | |
title="한국어 GPT 챗봇 (SKT Trinity 1.2B)" | |
).launch() |