ocr-job-code / llm_ocr /server.py
florentgbelidji's picture
Upload folder using huggingface_hub
fac50ab verified
"""vLLM server management and async inference client."""
from __future__ import annotations
import asyncio
import logging
import os
import signal
import subprocess
import threading
import time
from typing import TYPE_CHECKING, Any, Awaitable, Dict, List, Sequence
import requests
from openai import AsyncOpenAI
from .document import encode_image
if TYPE_CHECKING:
from PIL import Image
LOGGER = logging.getLogger(__name__)
def _stream_output(pipe, prefix: str) -> None:
"""Stream subprocess output to stdout with prefix."""
try:
for line in iter(pipe.readline, ""):
print(f"[{prefix}] {line.rstrip()}", flush=True)
finally:
pipe.close()
def launch_vllm() -> subprocess.Popen:
"""Launch vLLM server as subprocess."""
model_id = os.environ.get("MODEL_ID", "deepseek-ai/DeepSeek-OCR")
served_name = os.environ.get("SERVED_MODEL_NAME", "deepseek-ocr")
port = os.environ.get("PORT", "8080")
host = os.environ.get("HOST", "0.0.0.0")
cmd: List[str] = [
"vllm",
"serve",
"--model",
model_id,
"--served-model-name",
served_name,
"--tensor-parallel-size",
os.environ.get("TENSOR_PARALLEL_SIZE", "1"),
"--max-model-len",
os.environ.get("MAX_MODEL_LEN", "4096"),
"--gpu-memory-utilization",
os.environ.get("GPU_MEMORY_UTILIZATION", "0.90"),
"--port",
port,
"--host",
host,
"--trust-remote-code",
"--enable-chunked-prefill",
"--no-enable-prefix-caching",
"--mm-processor-cache-gb",
os.environ.get("MM_PROCESSOR_CACHE_GB", "0"),
"--logits-processors",
os.environ.get(
"LOGITS_PROCESSORS",
"vllm.model_executor.models.deepseek_ocr:NGramPerReqLogitsProcessor",
),
]
extra_args = os.environ.get("EXTRA_VLLM_ARGS")
if extra_args:
cmd.extend(extra_args.split())
LOGGER.info("Launching vLLM server: %s", " ".join(cmd))
process = subprocess.Popen(
cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True, bufsize=1
)
# Start output streaming threads
threads = []
for name, pipe in [("STDOUT", process.stdout), ("STDERR", process.stderr)]:
if pipe:
t = threading.Thread(
target=_stream_output, args=(pipe, f"vLLM {name}"), daemon=True
)
t.start()
threads.append(t)
process._log_threads = threads # type: ignore
return process
def shutdown_server(server_process: subprocess.Popen) -> None:
"""Gracefully shutdown vLLM server."""
LOGGER.info("Shutting down vLLM server")
server_process.send_signal(signal.SIGTERM)
try:
server_process.wait(timeout=30)
except subprocess.TimeoutExpired:
LOGGER.warning("Server did not exit in time, sending SIGKILL")
server_process.kill()
for thread in getattr(server_process, "_log_threads", []):
thread.join(timeout=1)
def _format_duration(seconds: float) -> str:
"""Format duration as mm:ss."""
minutes = int(seconds // 60)
secs = int(seconds % 60)
return f"{minutes:02d}:{secs:02d}"
def wait_for_server(url: str, timeout_s: int = None, interval_s: int = 5) -> bool:
"""Wait for server health endpoint to respond."""
if timeout_s is None:
timeout_s = int(os.environ.get("VLLM_STARTUP_TIMEOUT", "600")) # 10 min default
start_time = time.time()
LOGGER.info("⏳ Waiting for vLLM server to start...")
deadline = time.time() + timeout_s
while time.time() < deadline:
try:
if requests.get(url, timeout=5).ok:
elapsed = time.time() - start_time
LOGGER.info("✅ vLLM server ready in %s", _format_duration(elapsed))
return True
except Exception:
pass
time.sleep(interval_s)
elapsed = time.time() - start_time
LOGGER.error("❌ vLLM server failed to start after %s", _format_duration(elapsed))
return False
def should_launch_server() -> bool:
"""Check if server should be auto-launched."""
return os.environ.get("SKIP_SERVER_LAUNCH", "").lower() not in {"1", "true", "yes"}
def base_url_from_env() -> str:
"""Get vLLM base URL from environment."""
port = os.environ.get("PORT", "8080")
return os.environ.get("BASE_URL", f"http://127.0.0.1:{port}")
def _prepare_payload(
image: "Image.Image",
model_name: str,
prompt: str,
max_tokens: int,
temperature: float,
) -> Dict[str, Any]:
"""Prepare OpenAI-compatible chat completion payload."""
return {
"model": model_name,
"messages": [
{
"role": "user",
"content": [
{"type": "text", "text": prompt},
{
"type": "image_url",
"image_url": {
"url": f"data:image/png;base64,{encode_image(image)}"
},
},
],
}
],
"max_tokens": max_tokens,
"temperature": temperature,
"extra_body": {
"skip_special_tokens": False,
"vllm_xargs": {
"ngram_size": 30,
"window_size": 90,
"whitelist_token_ids": "[128821,128822]",
},
},
}
class DeepSeekClient:
"""Async batch inference client for DeepSeek OCR via vLLM."""
def __init__(
self,
base_url: str,
model_name: str,
max_tokens: int,
temperature: float,
*,
request_timeout: int = 120,
max_retries: int = 3,
retry_backoff_seconds: float = 2.0,
max_retry_wait_seconds: float = 60.0,
) -> None:
self.base_url = base_url.rstrip("/")
self.model_name = model_name
self.default_max_tokens = max_tokens
self.default_temperature = temperature
self.default_request_timeout = request_timeout
self.max_retries = max(0, max_retries)
self.retry_backoff_seconds = max(0.0, retry_backoff_seconds)
self.max_retry_wait_seconds = max_retry_wait_seconds
self._client = AsyncOpenAI(api_key="vllm", base_url=f"{self.base_url}/v1")
async def _async_completion(self, payload: Dict[str, Any], timeout: int) -> str:
"""Execute single async completion request."""
try:
response = await self._client.chat.completions.create(
model=payload["model"],
messages=payload["messages"],
max_tokens=payload["max_tokens"],
temperature=payload["temperature"],
timeout=timeout,
extra_body=payload.get("extra_body"),
)
except Exception as exc:
LOGGER.error("DeepSeek request failed: %s", exc)
raise
if not response.choices:
return ""
return getattr(response.choices[0].message, "content", "") or ""
def infer(self, requests_data: Sequence[Dict[str, Any]]) -> List[str]:
"""Run batch inference synchronously.
Args:
requests_data: List of dicts with keys: image (PIL.Image), prompt (str),
optional: max_tokens, temperature, request_timeout
Returns:
List of response strings, one per request
"""
if not requests_data:
return []
payloads = []
timeouts = []
for req in requests_data:
payloads.append(
_prepare_payload(
image=req["image"],
model_name=self.model_name,
prompt=req.get("prompt", ""),
max_tokens=req.get("max_tokens", self.default_max_tokens),
temperature=req.get("temperature", self.default_temperature),
)
)
timeouts.append(req.get("request_timeout") or self.default_request_timeout)
return self._run_async(self._async_infer_batch(payloads, timeouts))
async def _async_infer_batch(
self, payloads: Sequence[Dict[str, Any]], timeouts: Sequence[int]
) -> List[str]:
"""Run batch of async completions concurrently."""
tasks = [
asyncio.create_task(self._async_completion(p, t))
for p, t in zip(payloads, timeouts)
]
return await asyncio.gather(*tasks)
@staticmethod
def _run_async(coro: Awaitable[Any]) -> Any:
"""Run async coroutine in new event loop."""
loop = asyncio.new_event_loop()
try:
asyncio.set_event_loop(loop)
result = loop.run_until_complete(coro)
loop.run_until_complete(loop.shutdown_asyncgens())
return result
finally:
asyncio.set_event_loop(None)
loop.close()