tokenizer / app.py
wjm55
Refactor tokenize endpoint to use Pydantic model for request validation and update README.md examples to reflect changes in API usage
032494e
raw
history blame
539 Bytes
from fastapi import FastAPI
from pydantic import BaseModel
import spacy
app = FastAPI()
nlp = spacy.blank("en")
class TokenizeRequest(BaseModel):
text: str
@app.post("/tokenize")
async def tokenize_text(request: TokenizeRequest):
doc = nlp(request.text)
tokens = []
for token in doc:
tokens.append({
"text": token.text,
"start_char": token.idx,
"end_char": token.idx + len(token.text),
"start": token.i,
"end": token.i + 1
})
return tokens