Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -230,7 +230,7 @@ logging.basicConfig(level=logging.INFO)
|
|
| 230 |
logger = logging.getLogger(__name__)
|
| 231 |
|
| 232 |
# Load model and tokenizer
|
| 233 |
-
model_name = "FreedomIntelligence/Apollo-
|
| 234 |
# model_name = "emilyalsentzer/Bio_ClinicalBERT"
|
| 235 |
# model_name = "FreedomIntelligence/Apollo-2B"
|
| 236 |
|
|
@@ -239,7 +239,6 @@ model = AutoModelForCausalLM.from_pretrained(model_name)
|
|
| 239 |
|
| 240 |
tokenizer.pad_token = tokenizer.eos_token
|
| 241 |
|
| 242 |
-
|
| 243 |
app = FastAPI(title="Apollo RAG Medical Chatbot")
|
| 244 |
|
| 245 |
generation_config = GenerationConfig(
|
|
@@ -338,4 +337,4 @@ async def chat_fn(query: Query):
|
|
| 338 |
|
| 339 |
except Exception as e:
|
| 340 |
logger.error(f"Inference failed: {str(e)}")
|
| 341 |
-
raise HTTPException(status_code=500, detail="Model inference failed.")
|
|
|
|
| 230 |
logger = logging.getLogger(__name__)
|
| 231 |
|
| 232 |
# Load model and tokenizer
|
| 233 |
+
model_name = "FreedomIntelligence/Apollo-7B"
|
| 234 |
# model_name = "emilyalsentzer/Bio_ClinicalBERT"
|
| 235 |
# model_name = "FreedomIntelligence/Apollo-2B"
|
| 236 |
|
|
|
|
| 239 |
|
| 240 |
tokenizer.pad_token = tokenizer.eos_token
|
| 241 |
|
|
|
|
| 242 |
app = FastAPI(title="Apollo RAG Medical Chatbot")
|
| 243 |
|
| 244 |
generation_config = GenerationConfig(
|
|
|
|
| 337 |
|
| 338 |
except Exception as e:
|
| 339 |
logger.error(f"Inference failed: {str(e)}")
|
| 340 |
+
raise HTTPException(status_code=500, detail="Model inference TimeOut failed.")
|