Spaces:
Sleeping
Sleeping
import transformers | |
import torch | |
from fastapi import FastAPI | |
from pydantic import BaseModel | |
import uvicorn | |
app = FastAPI() | |
model_id = "meta-llama/Meta-Llama-3-8B" | |
pipeline = transformers.pipeline( | |
"text-generation", model=model_id, model_kwargs={"torch_dtype": torch.bfloat16}, device_map="auto" | |
) | |
class Item(BaseModel): | |
prompt: str | |
def generate(item: Item): | |
pipeline(item.prompt) | |
async def generate_text(item: Item): | |
return {"response": generate(item)} |