TestKMMLU / app.py
wabang's picture
Update app.py
1e28f8e verified
raw
history blame
2.51 kB
import gradio as gr
import pandas as pd
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
# model, tokenizer ์…‹ํŒ…
model_name = "meta-llama/Meta-Llama-3.1-8B-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
#hugging face ๋กœ๊ทธ์ธ (ํ† ํฐ ๊ฐ€์ ธ์˜ค๊ธฐ ์œ„ํ•ด์„œ)
from huggingface_hub import login
import os
# ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ ํ† ํฐ์„ ๊ฐ€์ ธ์˜ค๊ธฐ
hf_token = os.environ.get("llama3.31-8b-token-new")
# ํ† ํฐ์„ ์‚ฌ์šฉํ•˜์—ฌ ๋กœ๊ทธ์ธ
if hf_token:
login(token=hf_token)
else:
print("HF_TOKEN ํ™˜๊ฒฝ ๋ณ€์ˆ˜๊ฐ€ ์„ค์ •๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค.")
# KMMLU ๋ฐ์ดํ„ฐ์…‹ ๋กœ๋“œ
# ์ง์ ‘ ๋ถˆ๋Ÿฌ์˜ค๊ธฐ df = pd.read_csv("kmmlu_sample.csv")
from datasets import load_dataset
df = load_dataset("HAERAE-HUB/KMMLU", "Accounting")
def evaluate_model(question, choices):
prompt = f"์งˆ๋ฌธ: {question}\n\n์„ ํƒ์ง€:\n"
for i, choice in enumerate(choices):
prompt += f"{chr(65 + i)}. {choice}\n"
prompt += "\n๋‹ต๋ณ€:"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
with torch.no_grad():
outputs = model.generate(**inputs, max_new_tokens=1, temperature=0.0)
answer = tokenizer.decode(outputs[0][-1:], skip_special_tokens=True).strip()
return answer
def run_kmmlu_test(subject):
subject_df = df[df['subject'] == subject]
correct = 0
total = len(subject_df)
results = []
for _, row in subject_df.iterrows():
question = row['question']
choices = [row['A'], row['B'], row['C'], row['D']]
correct_answer = row['answer']
model_answer = evaluate_model(question, choices)
is_correct = model_answer == correct_answer
if is_correct:
correct += 1
results.append(f"์งˆ๋ฌธ: {question}\n๋ชจ๋ธ ๋‹ต๋ณ€: {model_answer}\n์ •๋‹ต: {correct_answer}\n์ •ํ™•๋„: {'๋งž์Œ' if is_correct else 'ํ‹€๋ฆผ'}\n")
accuracy = correct / total
summary = f"์ฃผ์ œ: {subject}\n์ •ํ™•๋„: {accuracy:.2%} ({correct}/{total})\n\n"
return summary + "\n".join(results)
subjects = df['subject'].unique().tolist()
iface = gr.Interface(
fn=run_kmmlu_test,
inputs=gr.Dropdown(choices=subjects, label="์ฃผ์ œ ์„ ํƒ"),
outputs="text",
title="Llama 3์„ ์ด์šฉํ•œ KMMLU ํ…Œ์ŠคํŠธ",
description="์„ ํƒํ•œ ์ฃผ์ œ์— ๋Œ€ํ•ด KMMLU ํ…Œ์ŠคํŠธ๋ฅผ ์‹คํ–‰ํ•ฉ๋‹ˆ๋‹ค."
)
iface.launch()