Upload 2 files
Browse files- Dockerfile +6 -7
- model.py +4 -2
Dockerfile
CHANGED
|
@@ -1,13 +1,12 @@
|
|
| 1 |
FROM python:3.10
|
| 2 |
|
| 3 |
-
|
| 4 |
-
WORKDIR /
|
| 5 |
-
|
| 6 |
-
# Copy files
|
| 7 |
-
COPY . /
|
| 8 |
|
| 9 |
# Install dependencies
|
|
|
|
| 10 |
RUN pip install --no-cache-dir -r requirements.txt
|
| 11 |
|
| 12 |
-
#
|
| 13 |
-
|
|
|
|
|
|
|
|
|
| 1 |
FROM python:3.10
|
| 2 |
|
| 3 |
+
WORKDIR /app
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
# Install dependencies
|
| 6 |
+
COPY requirements.txt .
|
| 7 |
RUN pip install --no-cache-dir -r requirements.txt
|
| 8 |
|
| 9 |
+
# Copy app code
|
| 10 |
+
COPY . .
|
| 11 |
+
|
| 12 |
+
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
|
model.py
CHANGED
|
@@ -5,11 +5,13 @@ import torch
|
|
| 5 |
BASE_MODEL = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
|
| 6 |
LORA_REPO = "sahil239/chatbot-v2"
|
| 7 |
|
| 8 |
-
|
|
|
|
| 9 |
model = AutoModelForCausalLM.from_pretrained(
|
| 10 |
BASE_MODEL,
|
| 11 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
| 12 |
-
device_map="auto"
|
|
|
|
| 13 |
)
|
| 14 |
model = PeftModel.from_pretrained(model, LORA_REPO)
|
| 15 |
model.eval()
|
|
|
|
| 5 |
BASE_MODEL = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
|
| 6 |
LORA_REPO = "sahil239/chatbot-v2"
|
| 7 |
|
| 8 |
+
|
| 9 |
+
tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
|
| 10 |
model = AutoModelForCausalLM.from_pretrained(
|
| 11 |
BASE_MODEL,
|
| 12 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
| 13 |
+
device_map="auto",
|
| 14 |
+
trust_remote_code=True
|
| 15 |
)
|
| 16 |
model = PeftModel.from_pretrained(model, LORA_REPO)
|
| 17 |
model.eval()
|