shakespear-tokenizer / handler.py
otmanheddouch's picture
Create handler.py
2412a74 verified
from transformers import AutoTokenizer
class EndpointHandler():
def __init__(self, path=""):
self.tokenizer = AutoTokenizer.from_pretrained("otmanheddouch/shakespear-tokenizer")
def __call__(self, data: str] -> Dict[str, Any]]:
"""
data args:
data: text input
Return:
output (dict) : outputs with tokens and their ids
"""
tokens = self.tokenizer.tokenize()
tokens_ids = self.tokenizer.encode(example)
return {"tokens:"tokens, "ids":tokens_ids}