Spaces:
Sleeping
Sleeping
wjm55
Fix tokenize_text function to correct token end index and remove unused 'supervision' package from requirements.txt
c4a8604
| from fastapi import FastAPI | |
| from pydantic import BaseModel | |
| import spacy | |
| app = FastAPI() | |
| nlp = spacy.blank("en") | |
| class TokenizeRequest(BaseModel): | |
| text: str | |
| async def tokenize_text(request: TokenizeRequest): | |
| doc = nlp(request.text) | |
| tokens = [] | |
| for token in doc: | |
| tokens.append({ | |
| "text": token.text, | |
| "start_char": token.idx, | |
| "end_char": token.idx + len(token.text), | |
| "start": token.i, | |
| "end": token.i | |
| }) | |
| return tokens | |