Spaces:
Sleeping
Sleeping
upload dataset
Browse files- Dockerfile +16 -2
- app.py +116 -34
- backup/2025-10-23/new_user.json +1 -0
- backup/2025-10-23/test_user_1.json +1 -0
- backup/2025-10-23/test_user_2.json +1 -0
- crontab +4 -0
- last_push_date.txt +1 -0
- requirements.txt +3 -0
- start.py +47 -0
- start_server.sh +60 -0
- start_with_cron.sh +26 -0
- test_api.sh +17 -26
- upload_hf_dataset.py +108 -0
Dockerfile
CHANGED
|
@@ -6,6 +6,7 @@ WORKDIR /app
|
|
| 6 |
# μμ€ν
ν¨ν€μ§ μ
λ°μ΄νΈ λ° νμν ν¨ν€μ§ μ€μΉ
|
| 7 |
RUN apt-get update && apt-get install -y \
|
| 8 |
build-essential \
|
|
|
|
| 9 |
&& rm -rf /var/lib/apt/lists/*
|
| 10 |
|
| 11 |
# requirements λ³΅μ¬ λ° μ€μΉ
|
|
@@ -15,9 +16,22 @@ RUN pip install --no-cache-dir -r requirements.txt
|
|
| 15 |
# μ ν리μΌμ΄μ
μ½λ 볡μ¬
|
| 16 |
COPY . .
|
| 17 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
# Hugging Face Spaceλ ν¬νΈ 7860μ μ¬μ©ν©λλ€
|
| 19 |
EXPOSE 7860
|
| 20 |
|
| 21 |
-
#
|
| 22 |
-
|
|
|
|
|
|
|
|
|
|
| 23 |
|
|
|
|
| 6 |
# μμ€ν
ν¨ν€μ§ μ
λ°μ΄νΈ λ° νμν ν¨ν€μ§ μ€μΉ
|
| 7 |
RUN apt-get update && apt-get install -y \
|
| 8 |
build-essential \
|
| 9 |
+
cron \
|
| 10 |
&& rm -rf /var/lib/apt/lists/*
|
| 11 |
|
| 12 |
# requirements λ³΅μ¬ λ° μ€μΉ
|
|
|
|
| 16 |
# μ ν리μΌμ΄μ
μ½λ 볡μ¬
|
| 17 |
COPY . .
|
| 18 |
|
| 19 |
+
# cron μ€μ νμΌ λ³΅μ¬
|
| 20 |
+
COPY crontab /etc/cron.d/batch-push-cron
|
| 21 |
+
|
| 22 |
+
# cron κΆν μ€μ
|
| 23 |
+
RUN chmod 0644 /etc/cron.d/batch-push-cron
|
| 24 |
+
RUN crontab /etc/cron.d/batch-push-cron
|
| 25 |
+
|
| 26 |
+
# λ‘κ·Έ λλ ν 리 μμ±
|
| 27 |
+
RUN mkdir -p /var/log
|
| 28 |
+
|
| 29 |
# Hugging Face Spaceλ ν¬νΈ 7860μ μ¬μ©ν©λλ€
|
| 30 |
EXPOSE 7860
|
| 31 |
|
| 32 |
+
# μμ μ€ν¬λ¦½νΈ λ³΅μ¬ λ° μ€ν
|
| 33 |
+
COPY start_with_cron.sh /start_with_cron.sh
|
| 34 |
+
RUN chmod +x /start_with_cron.sh
|
| 35 |
+
|
| 36 |
+
CMD ["/start_with_cron.sh"]
|
| 37 |
|
app.py
CHANGED
|
@@ -1,11 +1,17 @@
|
|
| 1 |
import os
|
| 2 |
import json
|
| 3 |
from typing import List, Optional
|
| 4 |
-
from fastapi import FastAPI, HTTPException
|
| 5 |
from pydantic import BaseModel, Field, ConfigDict
|
| 6 |
from typing import List
|
| 7 |
import oracledb
|
| 8 |
from dotenv import load_dotenv
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
# λ‘컬 κ°λ°: .env νμΌ λ‘λ (μμΌλ©΄)
|
| 11 |
load_dotenv()
|
|
@@ -19,6 +25,10 @@ WALLET_PASSWORD = os.environ["WALLET_PASSWORD"]
|
|
| 19 |
# tnsnames.ora μμ alias νμΈ (λ³΄ν΅ *_high)
|
| 20 |
TNS_ALIAS = os.environ.get("DB_TNS_ALIAS", "musclecare_high")
|
| 21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
# ----- μ°κ²° ν: Thin + mTLS (μ§κ°) -----
|
| 23 |
# μ λ νΈμΆνμ§ λ§μΈμ: oracledb.init_oracle_client() # (Thickλ‘ λΉ μ Έμ μ€ν¨ κ°λ₯)
|
| 24 |
pool: oracledb.ConnectionPool = oracledb.create_pool(
|
|
@@ -46,18 +56,6 @@ class StatePayload(BaseModel):
|
|
| 46 |
user_emb: Optional[List[float]] = Field(default=None, description="length=12")
|
| 47 |
model_version: Optional[str] = None
|
| 48 |
|
| 49 |
-
class LogItem(BaseModel):
|
| 50 |
-
user_id: str
|
| 51 |
-
session_id: Optional[str] = None
|
| 52 |
-
measure_date: str # 'YYYY-MM-DD'
|
| 53 |
-
rms: Optional[float] = None
|
| 54 |
-
freq: Optional[float] = None
|
| 55 |
-
fatigue: Optional[float] = None
|
| 56 |
-
mode: Optional[str] = Field(default="EMA", description="EMA/Hybrid/E2E")
|
| 57 |
-
window_count: Optional[int] = None
|
| 58 |
-
|
| 59 |
-
class LogsPayload(BaseModel):
|
| 60 |
-
items: List[LogItem] = Field(..., min_length=1)
|
| 61 |
|
| 62 |
# ----- μ νΈ -----
|
| 63 |
def clob_json(obj) -> str:
|
|
@@ -72,23 +70,39 @@ def root():
|
|
| 72 |
"message": "MuscleCare API Server",
|
| 73 |
"version": "1.0.0",
|
| 74 |
"endpoints": {
|
| 75 |
-
"health": "/health",
|
|
|
|
| 76 |
"docs": "/docs",
|
| 77 |
"upload_state": "/upload_state",
|
| 78 |
-
"
|
|
|
|
| 79 |
}
|
| 80 |
}
|
| 81 |
|
| 82 |
@app.get("/health")
|
| 83 |
def health():
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 84 |
try:
|
| 85 |
with pool.acquire() as conn:
|
| 86 |
with conn.cursor() as cur:
|
| 87 |
cur.execute("SELECT 1 FROM DUAL")
|
| 88 |
v = cur.fetchone()[0]
|
| 89 |
-
return {"ok": True, "db": v}
|
| 90 |
except Exception as e:
|
| 91 |
-
|
| 92 |
|
| 93 |
@app.post("/upload_state")
|
| 94 |
def upload_state(p: StatePayload):
|
|
@@ -132,23 +146,91 @@ def upload_state(p: StatePayload):
|
|
| 132 |
except Exception as e:
|
| 133 |
raise HTTPException(500, f"upload_state failed: {e}")
|
| 134 |
|
| 135 |
-
|
| 136 |
-
|
|
|
|
|
|
|
| 137 |
try:
|
| 138 |
-
|
| 139 |
-
|
| 140 |
-
|
| 141 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 142 |
|
| 143 |
-
with pool.acquire() as conn:
|
| 144 |
-
with conn.cursor() as cur:
|
| 145 |
-
cur.executemany("""
|
| 146 |
-
INSERT INTO MuscleCare.fatigue_logs
|
| 147 |
-
(user_id, session_id, measure_date, rms, freq, fatigue, mode, window_count, created_at)
|
| 148 |
-
VALUES
|
| 149 |
-
(:1, :2, TO_DATE(:3,'YYYY-MM-DD'), :4, :5, :6, :7, :8, CURRENT_TIMESTAMP)
|
| 150 |
-
""", rows)
|
| 151 |
-
conn.commit()
|
| 152 |
-
return {"ok": True, "inserted": len(rows)}
|
| 153 |
except Exception as e:
|
| 154 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
import os
|
| 2 |
import json
|
| 3 |
from typing import List, Optional
|
| 4 |
+
from fastapi import FastAPI, HTTPException, Request
|
| 5 |
from pydantic import BaseModel, Field, ConfigDict
|
| 6 |
from typing import List
|
| 7 |
import oracledb
|
| 8 |
from dotenv import load_dotenv
|
| 9 |
+
import json
|
| 10 |
+
import os
|
| 11 |
+
import requests
|
| 12 |
+
import pandas as pd
|
| 13 |
+
from datetime import datetime
|
| 14 |
+
from datasets import Dataset, DatasetDict, load_dataset
|
| 15 |
|
| 16 |
# λ‘컬 κ°λ°: .env νμΌ λ‘λ (μμΌλ©΄)
|
| 17 |
load_dotenv()
|
|
|
|
| 25 |
# tnsnames.ora μμ alias νμΈ (λ³΄ν΅ *_high)
|
| 26 |
TNS_ALIAS = os.environ.get("DB_TNS_ALIAS", "musclecare_high")
|
| 27 |
|
| 28 |
+
# Hugging Face μ€μ
|
| 29 |
+
HF_DATA_REPO_ID = os.getenv("HF_DATA_REPO_ID")
|
| 30 |
+
HF_DATA_TOKEN = os.getenv("HF_DATA_TOKEN")
|
| 31 |
+
|
| 32 |
# ----- μ°κ²° ν: Thin + mTLS (μ§κ°) -----
|
| 33 |
# μ λ νΈμΆνμ§ λ§μΈμ: oracledb.init_oracle_client() # (Thickλ‘ λΉ μ Έμ μ€ν¨ κ°λ₯)
|
| 34 |
pool: oracledb.ConnectionPool = oracledb.create_pool(
|
|
|
|
| 56 |
user_emb: Optional[List[float]] = Field(default=None, description="length=12")
|
| 57 |
model_version: Optional[str] = None
|
| 58 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 59 |
|
| 60 |
# ----- μ νΈ -----
|
| 61 |
def clob_json(obj) -> str:
|
|
|
|
| 70 |
"message": "MuscleCare API Server",
|
| 71 |
"version": "1.0.0",
|
| 72 |
"endpoints": {
|
| 73 |
+
"health": "/health (λΉ λ₯Έ 체ν¬)",
|
| 74 |
+
"health_db": "/health/db (DB μ°κ²° 체ν¬)",
|
| 75 |
"docs": "/docs",
|
| 76 |
"upload_state": "/upload_state",
|
| 77 |
+
"upload_dataset": "/upload_dataset",
|
| 78 |
+
"user_dataset": "/user_dataset/{user_id}"
|
| 79 |
}
|
| 80 |
}
|
| 81 |
|
| 82 |
@app.get("/health")
|
| 83 |
def health():
|
| 84 |
+
try:
|
| 85 |
+
# κ°λ¨ν health μ²΄ν¬ - DB μ°κ²° μμ΄ μλ² μνλ§ νμΈ
|
| 86 |
+
return {
|
| 87 |
+
"ok": True,
|
| 88 |
+
"server": "running",
|
| 89 |
+
"timestamp": datetime.now().isoformat(),
|
| 90 |
+
"status": "healthy"
|
| 91 |
+
}
|
| 92 |
+
except Exception as e:
|
| 93 |
+
return {"ok": False, "error": str(e)}
|
| 94 |
+
|
| 95 |
+
@app.get("/health/db")
|
| 96 |
+
def health_db():
|
| 97 |
+
"""DB μ°κ²°μ ν¬ν¨ν μμΈ health 체ν¬"""
|
| 98 |
try:
|
| 99 |
with pool.acquire() as conn:
|
| 100 |
with conn.cursor() as cur:
|
| 101 |
cur.execute("SELECT 1 FROM DUAL")
|
| 102 |
v = cur.fetchone()[0]
|
| 103 |
+
return {"ok": True, "db": v, "server": "running"}
|
| 104 |
except Exception as e:
|
| 105 |
+
return {"ok": False, "db": "error", "error": str(e)}
|
| 106 |
|
| 107 |
@app.post("/upload_state")
|
| 108 |
def upload_state(p: StatePayload):
|
|
|
|
| 146 |
except Exception as e:
|
| 147 |
raise HTTPException(500, f"upload_state failed: {e}")
|
| 148 |
|
| 149 |
+
|
| 150 |
+
@app.post("/upload_dataset")
|
| 151 |
+
async def upload_to_dataset(request: Request):
|
| 152 |
+
"""λ‘컬 νμΌμ μ¬μ©μλ³ λ°μ΄ν° μ μ₯"""
|
| 153 |
try:
|
| 154 |
+
data = await request.json()
|
| 155 |
+
user_id = data.get("user_id")
|
| 156 |
+
|
| 157 |
+
if not user_id:
|
| 158 |
+
raise HTTPException(status_code=400, detail="user_idκ° νμν©λλ€")
|
| 159 |
+
|
| 160 |
+
# λ°μ΄ν°μ νμμ€ν¬ν μΆκ°
|
| 161 |
+
data["timestamp"] = datetime.now().isoformat()
|
| 162 |
+
|
| 163 |
+
# λ‘컬 λ°μ΄ν° λλ ν 리 μμ±
|
| 164 |
+
data_dir = "user_data"
|
| 165 |
+
os.makedirs(data_dir, exist_ok=True)
|
| 166 |
+
|
| 167 |
+
# μ¬μ©μλ³ JSON νμΌ κ²½λ‘
|
| 168 |
+
user_file = os.path.join(data_dir, f"{user_id}.json")
|
| 169 |
+
|
| 170 |
+
# κΈ°μ‘΄ λ°μ΄ν° λ‘λ
|
| 171 |
+
existing_data = []
|
| 172 |
+
if os.path.exists(user_file):
|
| 173 |
+
try:
|
| 174 |
+
with open(user_file, 'r', encoding='utf-8') as f:
|
| 175 |
+
existing_data = json.load(f)
|
| 176 |
+
print(f"π κΈ°μ‘΄ λ°μ΄ν° λ‘λ: {user_id} ({len(existing_data)}κ° λ μ½λ)")
|
| 177 |
+
except:
|
| 178 |
+
existing_data = []
|
| 179 |
+
|
| 180 |
+
# μ λ°μ΄ν° μΆκ°
|
| 181 |
+
existing_data.append(data)
|
| 182 |
+
|
| 183 |
+
# νμΌ μ μ₯
|
| 184 |
+
with open(user_file, 'w', encoding='utf-8') as f:
|
| 185 |
+
json.dump(existing_data, f, ensure_ascii=False, indent=2)
|
| 186 |
+
|
| 187 |
+
print(f"β
λ‘컬 νμΌ μ μ₯ μλ£: {user_id} ({len(existing_data)}κ° λ μ½λ)")
|
| 188 |
+
return {
|
| 189 |
+
"user_id": user_id,
|
| 190 |
+
"rows": len(existing_data),
|
| 191 |
+
"status": "success",
|
| 192 |
+
"filename": f"{user_id}.json",
|
| 193 |
+
"file_path": user_file,
|
| 194 |
+
"message": f"Data saved to local file: {user_file}"
|
| 195 |
+
}
|
| 196 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 197 |
except Exception as e:
|
| 198 |
+
print(f"β λ‘컬 μ μ₯ μ€ν¨: {e}")
|
| 199 |
+
raise HTTPException(status_code=500, detail=f"λ‘컬 μ μ₯ μ€ν¨: {str(e)}")
|
| 200 |
+
|
| 201 |
+
@app.get("/user_dataset/{user_id}")
|
| 202 |
+
async def read_user_dataset(user_id: str):
|
| 203 |
+
"""λ‘컬 νμΌμμ μ¬μ©μ λ°μ΄ν° μ‘°ν"""
|
| 204 |
+
try:
|
| 205 |
+
# μ¬μ©μλ³ JSON νμΌ κ²½λ‘
|
| 206 |
+
data_dir = "user_data"
|
| 207 |
+
user_file = os.path.join(data_dir, f"{user_id}.json")
|
| 208 |
+
|
| 209 |
+
if not os.path.exists(user_file):
|
| 210 |
+
return {
|
| 211 |
+
"user_id": user_id,
|
| 212 |
+
"count": 0,
|
| 213 |
+
"recent_data": [],
|
| 214 |
+
"source": "local_file",
|
| 215 |
+
"message": "No data found"
|
| 216 |
+
}
|
| 217 |
+
|
| 218 |
+
# λ°μ΄ν° λ‘λ
|
| 219 |
+
with open(user_file, 'r', encoding='utf-8') as f:
|
| 220 |
+
data = json.load(f)
|
| 221 |
+
|
| 222 |
+
# μ΅κ·Ό 5κ° λ μ½λ λ°ν
|
| 223 |
+
recent_data = data[-5:] if len(data) > 5 else data
|
| 224 |
+
|
| 225 |
+
return {
|
| 226 |
+
"user_id": user_id,
|
| 227 |
+
"count": len(data),
|
| 228 |
+
"recent_data": recent_data,
|
| 229 |
+
"filename": f"{user_id}.json",
|
| 230 |
+
"source": "local_file"
|
| 231 |
+
}
|
| 232 |
+
|
| 233 |
+
except Exception as e:
|
| 234 |
+
print(f"β λ‘컬 μ‘°ν μ€ν¨: {e}")
|
| 235 |
+
raise HTTPException(status_code=500, detail=f"λ‘컬 μ‘°ν μ€ν¨: {str(e)}")
|
| 236 |
+
|
backup/2025-10-23/new_user.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
[{"user_id": "new_user", "rms": 0.95, "freq": 60.1, "fatigue": 0.88, "timestamp": "2025-10-23T15:02:00.000000"}]
|
backup/2025-10-23/test_user_1.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
[{"user_id": "test_user_1", "rms": 0.85, "freq": 52.3, "fatigue": 0.72, "timestamp": "2025-10-23T14:55:00.000000"}]
|
backup/2025-10-23/test_user_2.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
[{"user_id": "test_user_2", "rms": 0.91, "freq": 58.7, "fatigue": 0.81, "timestamp": "2025-10-23T14:55:30.000000"}]
|
crontab
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# λ§€μΌ μμ μ λ°°μΉ νΈμ μ€ν
|
| 2 |
+
0 0 * * * cd /app && python upload_hF_dataset.py >> /var/log/batch_push.log 2>&1
|
| 3 |
+
|
| 4 |
+
# λΉ μ€ νμ (cron μꡬμ¬ν)
|
last_push_date.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
2025-10-23
|
requirements.txt
CHANGED
|
@@ -4,3 +4,6 @@ python-multipart
|
|
| 4 |
oracledb>=2.3
|
| 5 |
pydantic
|
| 6 |
python-dotenv
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
oracledb>=2.3
|
| 5 |
pydantic
|
| 6 |
python-dotenv
|
| 7 |
+
datasets
|
| 8 |
+
pandas
|
| 9 |
+
requests
|
start.py
CHANGED
|
@@ -5,6 +5,50 @@ from dotenv import load_dotenv
|
|
| 5 |
# .env νμΌ λ‘λ (λ‘컬 κ°λ°μ©)
|
| 6 |
load_dotenv()
|
| 7 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
def prepare_wallet_dir():
|
| 9 |
"""
|
| 10 |
Wallet μ€λΉ:
|
|
@@ -56,6 +100,9 @@ if __name__ == "__main__":
|
|
| 56 |
print("π MuscleCare FastAPI μλ² μμ")
|
| 57 |
print("=" * 60)
|
| 58 |
|
|
|
|
|
|
|
|
|
|
| 59 |
# Wallet μ€λΉ
|
| 60 |
try:
|
| 61 |
prepare_wallet_dir()
|
|
|
|
| 5 |
# .env νμΌ λ‘λ (λ‘컬 κ°λ°μ©)
|
| 6 |
load_dotenv()
|
| 7 |
|
| 8 |
+
def setup_environment():
|
| 9 |
+
"""
|
| 10 |
+
νκ²½λ³μ μλ μ€μ :
|
| 11 |
+
- λ‘컬: .env νμΌμμ λ‘λ
|
| 12 |
+
- Hugging Face Space: νκ²½λ³μμμ μλ μ€μ
|
| 13 |
+
"""
|
| 14 |
+
print("π§ νκ²½λ³μ μ€μ μ€...")
|
| 15 |
+
|
| 16 |
+
# Hugging Face Dataset μ€μ (κΈ°λ³Έκ° μ€μ )
|
| 17 |
+
if not os.getenv("HF_DATA_REPO_ID"):
|
| 18 |
+
os.environ["HF_DATA_REPO_ID"] = "Merry99/musclecare-data"
|
| 19 |
+
print("π HF_DATA_REPO_ID κΈ°λ³Έκ° μ€μ : Merry99/musclecare-data")
|
| 20 |
+
|
| 21 |
+
if not os.getenv("HF_DATA_TOKEN"):
|
| 22 |
+
os.environ["HF_DATA_TOKEN"] = "hf_dummy_token_for_testing"
|
| 23 |
+
print("π HF_DATA_TOKEN κΈ°λ³Έκ° μ€μ : hf_dummy_token_for_testing")
|
| 24 |
+
|
| 25 |
+
# Oracle DB μ€μ νμΈ
|
| 26 |
+
required_vars = ["DB_USER", "DB_PASSWORD", "WALLET_DIR", "WALLET_PASSWORD"]
|
| 27 |
+
missing_vars = []
|
| 28 |
+
|
| 29 |
+
for var in required_vars:
|
| 30 |
+
if not os.getenv(var):
|
| 31 |
+
missing_vars.append(var)
|
| 32 |
+
|
| 33 |
+
if missing_vars:
|
| 34 |
+
print(f"β οΈ λλ½λ νκ²½λ³μ: {', '.join(missing_vars)}")
|
| 35 |
+
print("π‘ .env νμΌμ νμΈνκ±°λ νκ²½λ³μλ₯Ό μ€μ ν΄μ£ΌμΈμ.")
|
| 36 |
+
|
| 37 |
+
# λ‘컬 κ°λ°μ μν κΈ°λ³Έκ° μ€μ (κ²½κ³ μ ν¨κ»)
|
| 38 |
+
if not os.getenv("DB_USER"):
|
| 39 |
+
os.environ["DB_USER"] = "ADMIN"
|
| 40 |
+
print("π DB_USER κΈ°λ³Έκ° μ€μ : ADMIN")
|
| 41 |
+
if not os.getenv("DB_PASSWORD"):
|
| 42 |
+
os.environ["DB_PASSWORD"] = "YourDBPassword123!"
|
| 43 |
+
print("π DB_PASSWORD κΈ°λ³Έκ° μ€μ : YourDBPassword123!")
|
| 44 |
+
if not os.getenv("WALLET_PASSWORD"):
|
| 45 |
+
os.environ["WALLET_PASSWORD"] = "YourWalletPassword123!"
|
| 46 |
+
print("π WALLET_PASSWORD κΈ°λ³Έκ° μ€μ : YourWalletPassword123!")
|
| 47 |
+
else:
|
| 48 |
+
print("β
λͺ¨λ νμ νκ²½λ³μκ° μ€μ λμμ΅λλ€")
|
| 49 |
+
|
| 50 |
+
print("π§ νκ²½λ³μ μ€μ μλ£")
|
| 51 |
+
|
| 52 |
def prepare_wallet_dir():
|
| 53 |
"""
|
| 54 |
Wallet μ€λΉ:
|
|
|
|
| 100 |
print("π MuscleCare FastAPI μλ² μμ")
|
| 101 |
print("=" * 60)
|
| 102 |
|
| 103 |
+
# νκ²½λ³μ μλ μ€μ
|
| 104 |
+
setup_environment()
|
| 105 |
+
|
| 106 |
# Wallet μ€λΉ
|
| 107 |
try:
|
| 108 |
prepare_wallet_dir()
|
start_server.sh
ADDED
|
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
# MuscleCare FastAPI μλ² μμ μ€ν¬λ¦½νΈ
|
| 4 |
+
|
| 5 |
+
echo "======================================"
|
| 6 |
+
echo "π MuscleCare FastAPI μλ² μμ"
|
| 7 |
+
echo "======================================"
|
| 8 |
+
|
| 9 |
+
# .env νμΌ λ‘λ (μ νμ¬ν)
|
| 10 |
+
if [ -f .env ]; then
|
| 11 |
+
echo "π .env νμΌ λ‘λ μ€..."
|
| 12 |
+
export $(cat .env | grep -v '^#' | xargs)
|
| 13 |
+
echo "β
.env νμΌ λ‘λ μλ£"
|
| 14 |
+
else
|
| 15 |
+
echo "π .env νμΌμ΄ μμ΅λλ€. start.pyμμ μλμΌλ‘ νκ²½λ³μλ₯Ό μ€μ ν©λλ€."
|
| 16 |
+
fi
|
| 17 |
+
|
| 18 |
+
echo "π§ νκ²½λ³μλ start.pyμμ μλμΌλ‘ μ€μ λ©λλ€."
|
| 19 |
+
|
| 20 |
+
# Python μμ‘΄μ± νμΈ λ° μ€μΉ
|
| 21 |
+
echo "π¦ Python μμ‘΄μ± νμΈ μ€..."
|
| 22 |
+
if ! python3 -c "import fastapi, uvicorn, oracledb" 2>/dev/null; then
|
| 23 |
+
echo "β νμ Python ν¨ν€μ§κ° μ€μΉλμ§ μμμ΅λλ€."
|
| 24 |
+
echo "π§ μλ μ€μΉλ₯Ό μμν©λλ€..."
|
| 25 |
+
|
| 26 |
+
# pip3 μ¬μ© μλ
|
| 27 |
+
if command -v pip3 &> /dev/null; then
|
| 28 |
+
echo "π₯ pip3λ‘ ν¨ν€μ§ μ€μΉ μ€..."
|
| 29 |
+
pip3 install -r requirements.txt
|
| 30 |
+
# pip μ¬μ© μλ
|
| 31 |
+
elif command -v pip &> /dev/null; then
|
| 32 |
+
echo "π₯ pipμΌλ‘ ν¨ν€μ§ μ€μΉ μ€..."
|
| 33 |
+
pip install -r requirements.txt
|
| 34 |
+
# python -m pip μ¬μ© μλ
|
| 35 |
+
else
|
| 36 |
+
echo "π₯ python -m pipμΌλ‘ ν¨ν€μ§ μ€μΉ μ€..."
|
| 37 |
+
python3 -m pip install -r requirements.txt
|
| 38 |
+
fi
|
| 39 |
+
|
| 40 |
+
# μ€μΉ ν λ€μ νμΈ
|
| 41 |
+
if ! python3 -c "import fastapi, uvicorn, oracledb" 2>/dev/null; then
|
| 42 |
+
echo "β ν¨ν€μ§ μ€μΉμ μ€ν¨νμ΅λλ€."
|
| 43 |
+
echo "π‘ μλμΌλ‘ μ€μΉν΄μ£ΌμΈμ:"
|
| 44 |
+
echo " pip3 install -r requirements.txt"
|
| 45 |
+
exit 1
|
| 46 |
+
fi
|
| 47 |
+
fi
|
| 48 |
+
|
| 49 |
+
echo "β
Python μμ‘΄μ± νμΈ μλ£"
|
| 50 |
+
|
| 51 |
+
# μλ² μμ
|
| 52 |
+
echo ""
|
| 53 |
+
echo "π FastAPI μλ² μμ μ€..."
|
| 54 |
+
echo "π μλ² μ£Όμ: http://localhost:7860"
|
| 55 |
+
echo "π API λ¬Έμ: http://localhost:7860/docs"
|
| 56 |
+
echo "π μ’
λ£νλ €λ©΄ Ctrl+Cλ₯Ό λλ₯΄μΈμ"
|
| 57 |
+
echo ""
|
| 58 |
+
|
| 59 |
+
# μλ² μ€ν
|
| 60 |
+
python3 start.py
|
start_with_cron.sh
ADDED
|
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
# λ‘κ·Έ λλ ν 리 μμ±
|
| 4 |
+
mkdir -p /var/log
|
| 5 |
+
|
| 6 |
+
# cron μλΉμ€ μμ
|
| 7 |
+
service cron start
|
| 8 |
+
|
| 9 |
+
# cron μν νμΈ
|
| 10 |
+
echo "π
Cron μλΉμ€ μμλ¨"
|
| 11 |
+
crontab -l
|
| 12 |
+
|
| 13 |
+
# FastAPI μλ² μμ (λ°±κ·ΈλΌμ΄λ)
|
| 14 |
+
echo "π FastAPI μλ² μμ..."
|
| 15 |
+
python start.py &
|
| 16 |
+
|
| 17 |
+
# μλ²κ° μμλ λκΉμ§ λκΈ°
|
| 18 |
+
sleep 5
|
| 19 |
+
|
| 20 |
+
# μλ² μν νμΈ
|
| 21 |
+
echo "π μλ² μν νμΈ..."
|
| 22 |
+
curl -s http://localhost:7860/health || echo "β οΈ μλ² μμ μ€..."
|
| 23 |
+
|
| 24 |
+
# λ‘κ·Έ λͺ¨λν°λ§ (μ νμ¬ν)
|
| 25 |
+
echo "π λ°°μΉ νΈμ λ‘κ·Έ λͺ¨λν°λ§ μμ..."
|
| 26 |
+
tail -f /var/log/batch_push.log &
|
test_api.sh
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
#!/bin/bash
|
| 2 |
|
| 3 |
# MuscleCare API ν
μ€νΈ μ€ν¬λ¦½νΈ
|
| 4 |
|
|
@@ -37,39 +37,30 @@ curl -X POST "$BASE_URL/upload_state" \
|
|
| 37 |
echo ""
|
| 38 |
echo ""
|
| 39 |
|
| 40 |
-
# 3.
|
| 41 |
-
echo "3οΈβ£
|
| 42 |
echo "--------------------------------------"
|
| 43 |
-
curl -X POST "$BASE_URL/
|
| 44 |
-H "Content-Type: application/json" \
|
| 45 |
-H "accept: application/json" \
|
| 46 |
-d '{
|
| 47 |
-
"
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
"rms": 0.12,
|
| 53 |
-
"freq": 49.8,
|
| 54 |
-
"fatigue": 0.83,
|
| 55 |
-
"mode": "Hybrid",
|
| 56 |
-
"window_count": 100
|
| 57 |
-
},
|
| 58 |
-
{
|
| 59 |
-
"user_id": "test_user_001",
|
| 60 |
-
"session_id": "session_002",
|
| 61 |
-
"measure_date": "2025-10-22",
|
| 62 |
-
"rms": 0.15,
|
| 63 |
-
"freq": 48.5,
|
| 64 |
-
"fatigue": 0.91,
|
| 65 |
-
"mode": "E2E",
|
| 66 |
-
"window_count": 120
|
| 67 |
-
}
|
| 68 |
-
]
|
| 69 |
}' | jq .
|
| 70 |
echo ""
|
| 71 |
echo ""
|
| 72 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 73 |
echo "======================================"
|
| 74 |
echo "β
ν
μ€νΈ μλ£"
|
| 75 |
echo "======================================"
|
|
|
|
| 1 |
+
γ
γ
#!/bin/bash
|
| 2 |
|
| 3 |
# MuscleCare API ν
μ€νΈ μ€ν¬λ¦½νΈ
|
| 4 |
|
|
|
|
| 37 |
echo ""
|
| 38 |
echo ""
|
| 39 |
|
| 40 |
+
# 3. Dataset μ
λ‘λ
|
| 41 |
+
echo "3οΈβ£ Dataset μ
λ‘λ"
|
| 42 |
echo "--------------------------------------"
|
| 43 |
+
curl -X POST "$BASE_URL/upload_dataset" \
|
| 44 |
-H "Content-Type: application/json" \
|
| 45 |
-H "accept: application/json" \
|
| 46 |
-d '{
|
| 47 |
+
"user_id": "test_user_001",
|
| 48 |
+
"rms": 0.42,
|
| 49 |
+
"freq": 45.3,
|
| 50 |
+
"fatigue": 0.25,
|
| 51 |
+
"timestamp": "2024-01-15T10:30:00Z"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 52 |
}' | jq .
|
| 53 |
echo ""
|
| 54 |
echo ""
|
| 55 |
|
| 56 |
+
# 4. μ¬μ©μ Dataset μ‘°ν
|
| 57 |
+
echo "4οΈβ£ μ¬μ©μ Dataset μ‘°ν"
|
| 58 |
+
echo "--------------------------------------"
|
| 59 |
+
curl -X GET "$BASE_URL/user_dataset/test_user_001" \
|
| 60 |
+
-H "accept: application/json" | jq .
|
| 61 |
+
echo ""
|
| 62 |
+
echo ""
|
| 63 |
+
|
| 64 |
echo "======================================"
|
| 65 |
echo "β
ν
μ€νΈ μλ£"
|
| 66 |
echo "======================================"
|
upload_hf_dataset.py
ADDED
|
@@ -0,0 +1,108 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from datasets import Dataset, DatasetDict
|
| 2 |
+
from datetime import datetime, date
|
| 3 |
+
import pandas as pd, glob, json, os, shutil
|
| 4 |
+
from dotenv import load_dotenv
|
| 5 |
+
|
| 6 |
+
load_dotenv()
|
| 7 |
+
|
| 8 |
+
HF_DATA_REPO_ID = os.getenv("HF_DATA_REPO_ID")
|
| 9 |
+
HF_DATA_TOKEN = os.getenv("HF_DATA_TOKEN")
|
| 10 |
+
CACHE_DIR = "./user_data"
|
| 11 |
+
BACKUP_DIR = "./backup"
|
| 12 |
+
LAST_PUSH_FILE = "./last_push_date.txt"
|
| 13 |
+
|
| 14 |
+
def now_str():
|
| 15 |
+
return datetime.now().strftime("%Y-%m-%d %H:%M:%S")
|
| 16 |
+
|
| 17 |
+
def get_last_push_date():
|
| 18 |
+
"""λ§μ§λ§ νΈμ λ μ§ λ°ν"""
|
| 19 |
+
if os.path.exists(LAST_PUSH_FILE):
|
| 20 |
+
with open(LAST_PUSH_FILE, "r") as f:
|
| 21 |
+
return f.read().strip()
|
| 22 |
+
return None
|
| 23 |
+
|
| 24 |
+
def update_last_push_date():
|
| 25 |
+
"""λ§μ§λ§ νΈμ λ μ§ κΈ°λ‘"""
|
| 26 |
+
with open(LAST_PUSH_FILE, "w") as f:
|
| 27 |
+
f.write(str(date.today()))
|
| 28 |
+
|
| 29 |
+
def should_push_today():
|
| 30 |
+
"""μ€λ νΈμ μ¬λΆ νμΈ"""
|
| 31 |
+
last_push = get_last_push_date()
|
| 32 |
+
today = str(date.today())
|
| 33 |
+
return last_push != today
|
| 34 |
+
|
| 35 |
+
def batch_push_to_huggingface():
|
| 36 |
+
"""ν루 1ν Hugging Face Dataset μ
λ‘λ"""
|
| 37 |
+
# νμ νκ²½λ³μ νμΈ
|
| 38 |
+
if not HF_DATA_REPO_ID or not HF_DATA_TOKEN:
|
| 39 |
+
print(f"β {now_str()} - νκ²½λ³μ HF_DATA_REPO_ID λλ HF_DATA_TOKENμ΄ μ€μ λμ§ μμμ΅λλ€.")
|
| 40 |
+
return
|
| 41 |
+
|
| 42 |
+
# νΈμ μ¬λΆ 체ν¬
|
| 43 |
+
if not should_push_today():
|
| 44 |
+
print(f"π
{now_str()} - μ΄λ―Έ μ€λ νΈμ μλ£λ¨. μ’
λ£.")
|
| 45 |
+
return
|
| 46 |
+
|
| 47 |
+
files = glob.glob(os.path.join(CACHE_DIR, "*.json"))
|
| 48 |
+
if not files:
|
| 49 |
+
print(f"π {now_str()} - μΊμλ νμΌμ΄ μμ΅λλ€. μ’
λ£.")
|
| 50 |
+
return
|
| 51 |
+
|
| 52 |
+
print(f"π {now_str()} - λ°°μΉ νΈμ μμ ({len(files)}κ° νμΌ)")
|
| 53 |
+
|
| 54 |
+
user_splits = {}
|
| 55 |
+
for path in files:
|
| 56 |
+
user_id = os.path.basename(path).split(".")[0]
|
| 57 |
+
try:
|
| 58 |
+
with open(path, "r", encoding="utf-8") as f:
|
| 59 |
+
records = json.load(f)
|
| 60 |
+
if not records:
|
| 61 |
+
print(f"β οΈ {user_id}: λΉμ΄μλ νμΌ, 건λλ")
|
| 62 |
+
continue
|
| 63 |
+
|
| 64 |
+
df = pd.DataFrame(records)
|
| 65 |
+
user_splits[user_id] = Dataset.from_pandas(df)
|
| 66 |
+
print(f"π {user_id}: {len(records)}κ° λ μ½λ λ³ν μλ£")
|
| 67 |
+
|
| 68 |
+
except Exception as e:
|
| 69 |
+
print(f"β {user_id}: νμΌ λ‘λ μ€ν¨ β {e}")
|
| 70 |
+
continue
|
| 71 |
+
|
| 72 |
+
if not user_splits:
|
| 73 |
+
print(f"β {now_str()} - μ²λ¦¬ν λ°μ΄ν°κ° μμ΅λλ€. μ’
λ£.")
|
| 74 |
+
return
|
| 75 |
+
|
| 76 |
+
# λ°±μ
λλ ν 리 μμ±
|
| 77 |
+
os.makedirs(BACKUP_DIR, exist_ok=True)
|
| 78 |
+
backup_path = os.path.join(BACKUP_DIR, date.today().isoformat())
|
| 79 |
+
shutil.copytree(CACHE_DIR, backup_path, dirs_exist_ok=True)
|
| 80 |
+
print(f"ποΈ {now_str()} - λ°μ΄ν° λ°±μ
μλ£ β {backup_path}")
|
| 81 |
+
|
| 82 |
+
try:
|
| 83 |
+
dataset_dict = DatasetDict(user_splits)
|
| 84 |
+
dataset_dict.push_to_hub(HF_DATA_REPO_ID, token=HF_DATA_TOKEN, private=True)
|
| 85 |
+
print(f"β
{now_str()} - Hugging Face Hub νΈμ μ±κ³΅ ({len(user_splits)}λͺ
) β {HF_DATA_REPO_ID}")
|
| 86 |
+
|
| 87 |
+
# νΈμ μ±κ³΅ μ μΊμ μ 리
|
| 88 |
+
shutil.rmtree(CACHE_DIR, ignore_errors=True)
|
| 89 |
+
print(f"ποΈ {now_str()} - user_data λλ ν 리 μμ μλ£")
|
| 90 |
+
|
| 91 |
+
update_last_push_date()
|
| 92 |
+
print(f"π
{now_str()} - λ§μ§λ§ νΈμ λ μ§ μ
λ°μ΄νΈ μλ£")
|
| 93 |
+
|
| 94 |
+
except Exception as e:
|
| 95 |
+
print(f"β {now_str()} - νΈμ μ€ν¨: {e}")
|
| 96 |
+
print(f"β οΈ {now_str()} - μΊμ μ μ§ (λ°μ΄ν° μ μ€ λ°©μ§)")
|
| 97 |
+
# μ€ν¨ μ μΊμ μ μ§
|
| 98 |
+
return
|
| 99 |
+
|
| 100 |
+
def main():
|
| 101 |
+
"""CLI/cron μ§μ
μ """
|
| 102 |
+
try:
|
| 103 |
+
batch_push_to_huggingface()
|
| 104 |
+
except Exception as e:
|
| 105 |
+
print(f"π₯ {now_str()} - μκΈ°μΉ λͺ»ν μ€λ₯ λ°μ: {e}")
|
| 106 |
+
|
| 107 |
+
if __name__ == "__main__":
|
| 108 |
+
main()
|