KishoreK's picture
changing to llama 3.2 1B
d06aa40 verified
raw
history blame
618 Bytes
import gradio as gr
from transformers import pipeline
import os
access_token = os.getenv("Token")
pipe = pipeline("text-generation", model="meta-llama/Llama-3.2-1B-Instruct", token=access_token)
def predict(system="you are a helpful assistant", query):
messages = []
messages.append({"role": "system","content": system})
messages.append({"role": "system","content": query})
return pipe(messages,max_new_tokens=256)[0]["generated_text"][-1]
gradio_app = gr.Interface(
predict,
inputs=["text","text"], outputs="text",
title="Gemma 2B",
)
if __name__ == "__main__":
gradio_app.launch()