as
Browse files- Dockerfile +2 -0
- app.py +3 -4
- train_model.py +0 -0
Dockerfile
CHANGED
|
@@ -23,5 +23,7 @@ RUN echo "PORT=$(shuf -i 10000-65000 -n 1)" >> /etc/environment
|
|
| 23 |
# Exponer el puerto (aunque este valor será reemplazado por el puerto aleatorio)
|
| 24 |
EXPOSE 7860
|
| 25 |
|
|
|
|
|
|
|
| 26 |
# Usar la variable de entorno PORT para ejecutar la app
|
| 27 |
CMD ["bash", "-c", "source /etc/environment && uvicorn app:app --host 0.0.0.0 --port $PORT"]
|
|
|
|
| 23 |
# Exponer el puerto (aunque este valor será reemplazado por el puerto aleatorio)
|
| 24 |
EXPOSE 7860
|
| 25 |
|
| 26 |
+
#RUN python app.py --fine-tune
|
| 27 |
+
|
| 28 |
# Usar la variable de entorno PORT para ejecutar la app
|
| 29 |
CMD ["bash", "-c", "source /etc/environment && uvicorn app:app --host 0.0.0.0 --port $PORT"]
|
app.py
CHANGED
|
@@ -4,13 +4,12 @@ from fastapi import FastAPI
|
|
| 4 |
from pydantic import BaseModel
|
| 5 |
import uvicorn
|
| 6 |
import threading
|
| 7 |
-
from transformers import
|
| 8 |
import torch
|
| 9 |
|
| 10 |
# ======== Cargar el modelo DialoGPT =========
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
|
| 14 |
|
| 15 |
# ======== Definir API con FastAPI =========
|
| 16 |
app = FastAPI()
|
|
|
|
| 4 |
from pydantic import BaseModel
|
| 5 |
import uvicorn
|
| 6 |
import threading
|
| 7 |
+
from transformers import AutoTokenizer
|
| 8 |
import torch
|
| 9 |
|
| 10 |
# ======== Cargar el modelo DialoGPT =========
|
| 11 |
+
tokenizer = AutoTokenizer.from_pretrained("Novaciano/Qwen2.5_Uncensored_V2_Sexting-GGUF")
|
| 12 |
+
|
|
|
|
| 13 |
|
| 14 |
# ======== Definir API con FastAPI =========
|
| 15 |
app = FastAPI()
|
train_model.py
ADDED
|
File without changes
|