Spaces:
Sleeping
Sleeping
wjm55
commited on
Commit
·
032494e
1
Parent(s):
df9c057
Refactor tokenize endpoint to use Pydantic model for request validation and update README.md examples to reflect changes in API usage
Browse files- README.md +9 -6
- app.py +6 -2
- requirements.txt +2 -1
README.md
CHANGED
|
@@ -15,9 +15,10 @@ A simple API for tokenizing text using spaCy.
|
|
| 15 |
|
| 16 |
### Using curl
|
| 17 |
```bash
|
| 18 |
-
curl -X POST "https://theirstory-tokenizer.hf.space
|
| 19 |
-H "Content-Type: application/json" \
|
| 20 |
-
-
|
|
|
|
| 21 |
```
|
| 22 |
|
| 23 |
### Using Python
|
|
@@ -25,8 +26,9 @@ curl -X POST "https://theirstory-tokenizer.hf.space:7860/tokenize" \
|
|
| 25 |
import requests
|
| 26 |
|
| 27 |
response = requests.post(
|
| 28 |
-
"https://theirstory-tokenizer.hf.space
|
| 29 |
-
json={"text": "Hello world
|
|
|
|
| 30 |
)
|
| 31 |
tokens = response.json()
|
| 32 |
print(tokens)
|
|
@@ -34,13 +36,14 @@ print(tokens)
|
|
| 34 |
|
| 35 |
### Using JavaScript
|
| 36 |
```javascript
|
| 37 |
-
const response = await fetch('https://theirstory-tokenizer.hf.space
|
| 38 |
method: 'POST',
|
| 39 |
headers: {
|
| 40 |
'Content-Type': 'application/json',
|
|
|
|
| 41 |
},
|
| 42 |
body: JSON.stringify({
|
| 43 |
-
text: 'Hello world
|
| 44 |
})
|
| 45 |
});
|
| 46 |
const tokens = await response.json();
|
|
|
|
| 15 |
|
| 16 |
### Using curl
|
| 17 |
```bash
|
| 18 |
+
curl -X POST "https://theirstory-tokenizer.hf.space/tokenize" \
|
| 19 |
-H "Content-Type: application/json" \
|
| 20 |
+
-H "accept: application/json" \
|
| 21 |
+
-d '{"text": "Hello world"}'
|
| 22 |
```
|
| 23 |
|
| 24 |
### Using Python
|
|
|
|
| 26 |
import requests
|
| 27 |
|
| 28 |
response = requests.post(
|
| 29 |
+
"https://theirstory-tokenizer.hf.space/tokenize",
|
| 30 |
+
json={"text": "Hello world"},
|
| 31 |
+
headers={"accept": "application/json"}
|
| 32 |
)
|
| 33 |
tokens = response.json()
|
| 34 |
print(tokens)
|
|
|
|
| 36 |
|
| 37 |
### Using JavaScript
|
| 38 |
```javascript
|
| 39 |
+
const response = await fetch('https://theirstory-tokenizer.hf.space/tokenize', {
|
| 40 |
method: 'POST',
|
| 41 |
headers: {
|
| 42 |
'Content-Type': 'application/json',
|
| 43 |
+
'accept': 'application/json'
|
| 44 |
},
|
| 45 |
body: JSON.stringify({
|
| 46 |
+
text: 'Hello world'
|
| 47 |
})
|
| 48 |
});
|
| 49 |
const tokens = await response.json();
|
app.py
CHANGED
|
@@ -1,12 +1,16 @@
|
|
| 1 |
from fastapi import FastAPI
|
|
|
|
| 2 |
import spacy
|
| 3 |
|
| 4 |
app = FastAPI()
|
| 5 |
nlp = spacy.blank("en")
|
| 6 |
|
|
|
|
|
|
|
|
|
|
| 7 |
@app.post("/tokenize")
|
| 8 |
-
async def tokenize_text(
|
| 9 |
-
doc = nlp(text)
|
| 10 |
tokens = []
|
| 11 |
for token in doc:
|
| 12 |
tokens.append({
|
|
|
|
| 1 |
from fastapi import FastAPI
|
| 2 |
+
from pydantic import BaseModel
|
| 3 |
import spacy
|
| 4 |
|
| 5 |
app = FastAPI()
|
| 6 |
nlp = spacy.blank("en")
|
| 7 |
|
| 8 |
+
class TokenizeRequest(BaseModel):
|
| 9 |
+
text: str
|
| 10 |
+
|
| 11 |
@app.post("/tokenize")
|
| 12 |
+
async def tokenize_text(request: TokenizeRequest):
|
| 13 |
+
doc = nlp(request.text)
|
| 14 |
tokens = []
|
| 15 |
for token in doc:
|
| 16 |
tokens.append({
|
requirements.txt
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
fastapi
|
| 2 |
uvicorn[standard]
|
| 3 |
supervision
|
| 4 |
-
spacy
|
|
|
|
|
|
| 1 |
fastapi
|
| 2 |
uvicorn[standard]
|
| 3 |
supervision
|
| 4 |
+
spacy
|
| 5 |
+
pydantic
|