16pramodh commited on
Commit
ebf3750
·
1 Parent(s): 0d674d4

fixing cache issue

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -1,17 +1,18 @@
 
1
  from fastapi import FastAPI
2
  from pydantic import BaseModel
3
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
4
  import uvicorn
5
 
6
- # Load model and tokenizer once at startup
 
 
7
  MODEL_NAME = "16pramodh/t2s_model"
8
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
9
  model = AutoModelForSeq2SeqLM.from_pretrained(MODEL_NAME)
10
 
11
- # Create FastAPI app
12
  app = FastAPI()
13
 
14
- # Request body format
15
  class QueryRequest(BaseModel):
16
  text: str
17
 
@@ -25,6 +26,5 @@ def predict(request: QueryRequest):
25
  except Exception as e:
26
  return {"error": str(e)}
27
 
28
- # For local testing
29
  if __name__ == "__main__":
30
  uvicorn.run(app, host="0.0.0.0", port=7860)
 
1
+ import os
2
  from fastapi import FastAPI
3
  from pydantic import BaseModel
4
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
5
  import uvicorn
6
 
7
+ # Set writable cache directory
8
+ os.environ["TRANSFORMERS_CACHE"] = "/tmp/cache"
9
+
10
  MODEL_NAME = "16pramodh/t2s_model"
11
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
12
  model = AutoModelForSeq2SeqLM.from_pretrained(MODEL_NAME)
13
 
 
14
  app = FastAPI()
15
 
 
16
  class QueryRequest(BaseModel):
17
  text: str
18
 
 
26
  except Exception as e:
27
  return {"error": str(e)}
28
 
 
29
  if __name__ == "__main__":
30
  uvicorn.run(app, host="0.0.0.0", port=7860)