Spaces:
Sleeping
Sleeping
Factor Studios
commited on
Update http_storage.py
Browse files- http_storage.py +6 -7
http_storage.py
CHANGED
|
@@ -32,7 +32,7 @@ class HTTPGPUStorage:
|
|
| 32 |
return
|
| 33 |
|
| 34 |
self.base_url = base_url.rstrip('/')
|
| 35 |
-
self.api_base = f"{self.base_url}
|
| 36 |
self.session_token = None
|
| 37 |
self.session_id = None
|
| 38 |
self.lock = threading.Lock()
|
|
@@ -177,7 +177,7 @@ class HTTPGPUStorage:
|
|
| 177 |
|
| 178 |
response = self._make_request(
|
| 179 |
'POST',
|
| 180 |
-
f'/vram/blocks/{tensor_id}',
|
| 181 |
json=request_data
|
| 182 |
)
|
| 183 |
|
|
@@ -209,7 +209,7 @@ class HTTPGPUStorage:
|
|
| 209 |
logging.warning(f"Tensor {tensor_id} not registered in VRAM")
|
| 210 |
# Still try to load it in case it exists on server
|
| 211 |
|
| 212 |
-
response = self._make_request('GET', f'/vram/blocks/{tensor_id}')
|
| 213 |
|
| 214 |
if response and response.get('status') == 'success':
|
| 215 |
data = response.get('data')
|
|
@@ -396,7 +396,7 @@ class HTTPGPUStorage:
|
|
| 396 |
|
| 397 |
response = self._make_request(
|
| 398 |
'POST',
|
| 399 |
-
f'/models/{model_name}/inference',
|
| 400 |
json=request_data
|
| 401 |
)
|
| 402 |
|
|
@@ -416,9 +416,8 @@ class HTTPGPUStorage:
|
|
| 416 |
|
| 417 |
def ping(self) -> bool:
|
| 418 |
"""Ping the server to check connection status."""
|
| 419 |
-
try:
|
| 420 |
-
response
|
| 421 |
-
return response and response.get('status') == 'ok'
|
| 422 |
except Exception as e:
|
| 423 |
logging.error(f"Ping failed: {e}")
|
| 424 |
return False
|
|
|
|
| 32 |
return
|
| 33 |
|
| 34 |
self.base_url = base_url.rstrip('/')
|
| 35 |
+
self.api_base = f"{self.base_url}"
|
| 36 |
self.session_token = None
|
| 37 |
self.session_id = None
|
| 38 |
self.lock = threading.Lock()
|
|
|
|
| 177 |
|
| 178 |
response = self._make_request(
|
| 179 |
'POST',
|
| 180 |
+
f'/api/v1/vram/blocks/{tensor_id}',
|
| 181 |
json=request_data
|
| 182 |
)
|
| 183 |
|
|
|
|
| 209 |
logging.warning(f"Tensor {tensor_id} not registered in VRAM")
|
| 210 |
# Still try to load it in case it exists on server
|
| 211 |
|
| 212 |
+
response = self._make_request('GET', f'/api/v1/vram/blocks/{tensor_id}')
|
| 213 |
|
| 214 |
if response and response.get('status') == 'success':
|
| 215 |
data = response.get('data')
|
|
|
|
| 396 |
|
| 397 |
response = self._make_request(
|
| 398 |
'POST',
|
| 399 |
+
f'/api/v1/models/{model_name}/inference',
|
| 400 |
json=request_data
|
| 401 |
)
|
| 402 |
|
|
|
|
| 416 |
|
| 417 |
def ping(self) -> bool:
|
| 418 |
"""Ping the server to check connection status."""
|
| 419 |
+
try: response = self._make_request("GET", f"/api/v1/sessions")
|
| 420 |
+
return response and response.get("session_token") is not None
|
|
|
|
| 421 |
except Exception as e:
|
| 422 |
logging.error(f"Ping failed: {e}")
|
| 423 |
return False
|