burtenshaw's picture
burtenshaw HF Staff
Upload folder using huggingface_hub
09f6e69 verified
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree.
"""
HTTP server wrapper for Environment instances.
This module provides utilities to wrap any Environment subclass and expose it
over HTTP and WebSocket endpoints that EnvClient can consume.
"""
from __future__ import annotations
import asyncio
import inspect
import json
import os
import time
import uuid
from concurrent.futures import ThreadPoolExecutor
from typing import Any, Callable, Dict, Optional, Type
from fastapi import (
Body,
FastAPI,
HTTPException,
Request,
WebSocket,
WebSocketDisconnect,
status,
)
from pydantic import ValidationError
from .interfaces import Environment
from .route_config import (
GetEndpointConfig,
register_get_endpoints,
)
from .serialization import deserialize_action, serialize_observation
from .types import (
Action,
Observation,
ResetRequest,
ResetResponse,
State,
StepRequest,
StepResponse,
EnvironmentMetadata,
SchemaResponse,
HealthResponse,
HealthStatus,
ServerMode,
WSErrorCode,
WSResetMessage,
WSStepMessage,
WSStateMessage,
WSCloseMessage,
WSObservationResponse,
WSStateResponse,
WSErrorResponse,
ConcurrencyConfig,
ServerCapacityStatus,
SessionInfo,
)
from .mcp_types import (
JsonRpcErrorCode,
JsonRpcRequest,
JsonRpcResponse,
McpMethod,
WSMCPMessage,
WSMCPResponse,
)
def _make_json_serializable(obj: Any) -> Any:
"""
Convert an object to a JSON-serializable form.
Handles Pydantic models, dataclasses, and other common types.
Args:
obj: The object to convert
Returns:
A JSON-serializable representation of the object
"""
if obj is None:
return None
if isinstance(obj, (str, int, float, bool)):
return obj
if isinstance(obj, (list, tuple)):
return [_make_json_serializable(item) for item in obj]
if isinstance(obj, dict):
return {k: _make_json_serializable(v) for k, v in obj.items()}
if hasattr(obj, "model_dump"):
# Pydantic model
return obj.model_dump()
if hasattr(obj, "__dict__"):
# Object with __dict__
return {k: _make_json_serializable(v) for k, v in obj.__dict__.items()}
# Fallback to string representation
return str(obj)
from .exceptions import (
ConcurrencyConfigurationError,
SessionCapacityError,
EnvironmentFactoryError,
)
class HTTPEnvServer:
"""
HTTP server wrapper for Environment instances.
This class wraps an Environment and exposes its reset(), step(), and state
methods as HTTP and WebSocket endpoints compatible with EnvClient.
The server expects:
- Action deserialization: Converts JSON dict to Action subclass
- Observation serialization: Converts Observation subclass to JSON dict
Example:
>>> from core.env_server import HTTPEnvServer
>>> from envs.coding_env.server import CodeExecutionEnvironment
>>> from envs.coding_env.models import CodeAction, CodeObservation
>>>
>>> # Pass environment class (factory pattern)
>>> server = HTTPEnvServer(
... env=CodeExecutionEnvironment,
... action_cls=CodeAction,
... observation_cls=CodeObservation,
... max_concurrent_envs=4,
... )
>>>
>>> # Register routes with FastAPI
>>> from fastapi import FastAPI
>>> app = FastAPI()
>>> server.register_routes(app)
"""
def __init__(
self,
env: Callable[[], Environment],
action_cls: Type[Action],
observation_cls: Type[Observation],
max_concurrent_envs: Optional[int] = None,
concurrency_config: Optional[ConcurrencyConfig] = None,
):
"""
Initialize HTTP server wrapper.
Args:
env: Environment factory (callable) that creates new instances.
Will be called to create a new environment for each WebSocket session.
action_cls: The Action subclass this environment expects
observation_cls: The Observation subclass this environment returns
max_concurrent_envs: Maximum number of concurrent WebSocket sessions.
Mutually exclusive with concurrency_config.
concurrency_config: Optional ConcurrencyConfig for advanced concurrency settings.
Mutually exclusive with max_concurrent_envs.
Raises:
ValueError: If both max_concurrent_envs and concurrency_config are provided.
ConcurrencyConfigurationError: If max_concurrent_envs > 1 for an
environment that is not marked as SUPPORTS_CONCURRENT_SESSIONS.
"""
# Validate that env is callable
if not callable(env):
raise TypeError(
f"env must be a callable (class or factory function), got {type(env)}. "
f"Pass the environment class (e.g., MyEnvironment) not an instance (e.g., MyEnvironment())."
)
self._env_factory: Callable[[], Environment] = env
# Handle concurrency configuration
if max_concurrent_envs is not None and concurrency_config is not None:
raise ValueError(
"Cannot specify both 'max_concurrent_envs' and 'concurrency_config'. "
"Please use only one method to configure concurrency."
)
if concurrency_config is not None:
self._concurrency_config = concurrency_config
elif max_concurrent_envs is not None:
self._concurrency_config = ConcurrencyConfig(
max_concurrent_envs=max_concurrent_envs,
session_timeout=None,
)
else:
# Default configuration
self._concurrency_config = ConcurrencyConfig(
max_concurrent_envs=1,
session_timeout=None,
)
self._max_concurrent_envs = self._concurrency_config.max_concurrent_envs
# Validate concurrency configuration
self._validate_concurrency_safety()
self.action_cls = action_cls
self.observation_cls = observation_cls
# Session management for WebSocket connections
self._sessions: Dict[str, Environment] = {}
self._session_executors: Dict[str, ThreadPoolExecutor] = {}
self._session_info: Dict[str, SessionInfo] = {}
self._session_lock = asyncio.Lock()
# Create thread pool for running sync code in async context
# This is needed for environments using sync libraries (e.g., Playwright)
self._executor = ThreadPoolExecutor(max_workers=32)
def _validate_concurrency_safety(self) -> None:
"""
Validate that the environment supports the configured concurrency level.
Raises:
ConcurrencyConfigurationError: If max_concurrent_envs > 1 for an
environment that is not marked as SUPPORTS_CONCURRENT_SESSIONS.
"""
if self._max_concurrent_envs <= 1:
return
if inspect.isclass(self._env_factory):
env_cls = self._env_factory
else:
_temp_env = self._env_factory()
env_cls = type(_temp_env)
_temp_env.close()
del _temp_env
if not getattr(env_cls, "SUPPORTS_CONCURRENT_SESSIONS", False):
raise ConcurrencyConfigurationError(
environment_name=env_cls.__name__,
max_concurrent_envs=self._max_concurrent_envs,
)
def get_capacity_status(self) -> ServerCapacityStatus:
"""
Get the current capacity status of the server.
Returns:
ServerCapacityStatus with current session counts and availability.
"""
return ServerCapacityStatus.from_counts(
active=len(self._sessions),
max_sessions=self._max_concurrent_envs,
)
async def _run_sync_in_thread_pool(
self, func: Callable[..., Observation], *args, **kwargs
) -> Observation:
"""Run a synchronous function in the thread pool executor."""
loop = asyncio.get_event_loop()
return await loop.run_in_executor(self._executor, lambda: func(*args, **kwargs))
def _get_valid_kwargs(
self,
sig: inspect.Signature,
kwargs: Dict[str, Any],
skip_params: Optional[set[str]] = None,
) -> Dict[str, Any]:
"""Filter kwargs to only include parameters accepted by the function signature."""
if skip_params is None:
skip_params = set()
valid_kwargs = {}
has_kwargs = any(
p.kind == inspect.Parameter.VAR_KEYWORD for p in sig.parameters.values()
)
for k, v in kwargs.items():
if k in sig.parameters or has_kwargs:
if k not in skip_params:
valid_kwargs[k] = v
return valid_kwargs
async def _create_session(self) -> tuple[str, Environment]:
"""
Create a new WebSocket session with its own environment instance.
Returns:
Tuple of (session_id, environment)
Raises:
SessionCapacityError: If max concurrent sessions reached
EnvironmentFactoryError: If the factory fails to create an environment
"""
async with self._session_lock:
if len(self._sessions) >= self._max_concurrent_envs:
raise SessionCapacityError(
active_sessions=len(self._sessions),
max_sessions=self._max_concurrent_envs,
)
session_id = str(uuid.uuid4())
current_time = time.time()
# Create executor and reserve slot so capacity is not exceeded while
# we create the env outside the lock (avoids blocking other sessions)
executor = ThreadPoolExecutor(max_workers=1)
self._session_executors[session_id] = executor
self._sessions[session_id] = None # placeholder until env is ready
try:
# Create environment in the executor thread (outside lock)
loop = asyncio.get_event_loop()
env = await loop.run_in_executor(executor, self._env_factory)
except Exception as e:
async with self._session_lock:
executor.shutdown(wait=False)
self._session_executors.pop(session_id, None)
self._sessions.pop(session_id, None)
factory_name = getattr(
self._env_factory, "__name__", str(self._env_factory)
)
raise EnvironmentFactoryError(factory_name) from e
async with self._session_lock:
self._sessions[session_id] = env
self._session_info[session_id] = SessionInfo(
session_id=session_id,
created_at=current_time,
last_activity_at=current_time,
step_count=0,
environment_type=type(env).__name__,
)
return session_id, env
async def _destroy_session(self, session_id: str) -> None:
"""
Destroy a WebSocket session and cleanup resources.
Args:
session_id: The session ID to destroy
"""
async with self._session_lock:
env = self._sessions.pop(session_id, None)
executor = self._session_executors.pop(session_id, None)
self._session_info.pop(session_id, None)
# Run close() in the same executor where the env was created
# This is required for thread-sensitive libraries like Playwright/greenlet
if env is not None:
if executor is not None:
try:
loop = asyncio.get_event_loop()
await loop.run_in_executor(executor, env.close)
except Exception:
# If executor close fails, try direct close as fallback
try:
env.close()
except Exception:
pass # Best effort cleanup
else:
try:
env.close()
except Exception:
pass # Best effort cleanup
# Shutdown executor after close is done
if executor is not None:
executor.shutdown(wait=False)
def _update_session_activity(
self, session_id: str, increment_step: bool = False
) -> None:
"""
Update session activity timestamp and optionally increment step count.
Args:
session_id: The session ID to update
increment_step: If True, increment the step count
"""
if session_id in self._session_info:
self._session_info[session_id].last_activity_at = time.time()
if increment_step:
self._session_info[session_id].step_count += 1
def get_session_info(self, session_id: str) -> Optional[SessionInfo]:
"""
Get information about a specific session.
Args:
session_id: The session ID to query
Returns:
SessionInfo if the session exists, None otherwise
"""
return self._session_info.get(session_id)
async def _run_in_session_executor(
self, session_id: str, func: Callable[..., Observation], *args, **kwargs
) -> Observation:
"""Run a synchronous function in the session's thread pool executor."""
executor = self._session_executors.get(session_id, self._executor)
loop = asyncio.get_event_loop()
return await loop.run_in_executor(executor, lambda: func(*args, **kwargs))
@property
def active_sessions(self) -> int:
"""Return the number of active WebSocket sessions."""
return len(self._sessions)
@property
def max_concurrent_envs(self) -> int:
"""Return the maximum number of concurrent environments."""
return self._max_concurrent_envs
@property
def is_concurrency_safe(self) -> bool:
"""Return whether the environment is marked as concurrency safe."""
import inspect
if inspect.isclass(self._env_factory):
return getattr(self._env_factory, "SUPPORTS_CONCURRENT_SESSIONS", False)
else:
_temp_env = self._env_factory()
result = getattr(_temp_env, "SUPPORTS_CONCURRENT_SESSIONS", False)
_temp_env.close()
del _temp_env
return result
@property
def concurrency_config(self) -> ConcurrencyConfig:
"""Return the concurrency configuration."""
return self._concurrency_config
def register_routes(
self, app: FastAPI, mode: ServerMode | str = ServerMode.SIMULATION
) -> None:
"""
Register HTTP routes on a FastAPI application.
Args:
app: FastAPI application instance
mode: Server mode - either SIMULATION or PRODUCTION (or string equivalents).
In production mode, simulation control endpoints (/reset, /step, /state)
are NOT registered. Only safe endpoints (/health, /schema, /metadata, /ws)
are available. Defaults to SIMULATION for backwards compatibility.
Raises:
ValueError: If mode is not a valid ServerMode or string equivalent.
"""
# Convert string to ServerMode enum for backwards compatibility
if isinstance(mode, str):
try:
mode = ServerMode(mode.lower())
except ValueError:
valid_modes = [m.value for m in ServerMode]
raise ValueError(
f"Invalid mode: '{mode}'. Must be one of: {valid_modes}"
)
# Helper function to handle reset endpoint
async def reset_handler(
request: ResetRequest = Body(default_factory=ResetRequest),
) -> ResetResponse:
"""Reset endpoint - returns initial observation."""
_env = self._env_factory()
try:
kwargs = request.model_dump(exclude_unset=True)
is_async = _env.reset_async.__func__ is not Environment.reset_async
if is_async:
sig = inspect.signature(_env.reset_async)
else:
sig = inspect.signature(_env.reset)
valid_kwargs = self._get_valid_kwargs(sig, kwargs)
if is_async:
observation = await _env.reset_async(**valid_kwargs)
else:
observation = await self._run_sync_in_thread_pool(
_env.reset, **valid_kwargs
)
return ResetResponse(**serialize_observation(observation))
finally:
_env.close()
# Helper function to handle step endpoint
async def step_handler(request: StepRequest) -> StepResponse:
"""Step endpoint - executes action and returns observation."""
action_data = request.action
try:
action = deserialize_action(action_data, self.action_cls)
except ValidationError as e:
raise HTTPException(
status_code=status.HTTP_422_UNPROCESSABLE_CONTENT, detail=e.errors()
)
_env = self._env_factory()
try:
kwargs = request.model_dump(exclude_unset=True, exclude={"action"})
is_async = _env.step_async.__func__ is not Environment.step_async
if is_async:
sig = inspect.signature(_env.step_async)
else:
sig = inspect.signature(_env.step)
valid_kwargs = self._get_valid_kwargs(
sig, kwargs, skip_params={"action"}
)
if is_async:
observation = await _env.step_async(action, **valid_kwargs)
else:
observation = await self._run_sync_in_thread_pool(
_env.step, action, **valid_kwargs
)
return StepResponse(**serialize_observation(observation))
finally:
_env.close()
# Helper function to handle MCP endpoint
async def mcp_handler(
request: JsonRpcRequest, session_env: Optional[Environment] = None
) -> JsonRpcResponse:
"""
Handle MCP JSON-RPC requests.
Supports tools/list and tools/call methods in JSON-RPC 2.0 format.
"""
method = request.method
request_id = request.id
# Use provided session environment or create temporary one
if session_env is not None:
_env = session_env
should_close = False
else:
_env = self._env_factory()
should_close = True
try:
if method == McpMethod.TOOLS_LIST:
# Check if environment is MCP-enabled
if not hasattr(_env, "mcp_client"):
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
"Environment does not support MCP",
request_id=request_id,
)
# Use async context manager for MCP client
async with _env.mcp_client:
tools = await _env.mcp_client.list_tools()
return JsonRpcResponse.success(
result={
"tools": [
t.model_dump() if hasattr(t, "model_dump") else dict(t)
for t in tools
]
},
request_id=request_id,
)
elif method == McpMethod.TOOLS_CALL:
params = request.params
tool_name = params.get("name")
arguments = params.get("arguments", {})
if not hasattr(_env, "mcp_client"):
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
"Environment does not support MCP",
request_id=request_id,
)
if not tool_name:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INVALID_REQUEST,
"Missing 'name' in params",
request_id=request_id,
)
# Use async context manager for MCP client
async with _env.mcp_client:
result = await _env.mcp_client.call_tool(
name=tool_name, arguments=arguments
)
# Ensure result is JSON serializable
serializable_result = _make_json_serializable(result)
return JsonRpcResponse.success(
result=serializable_result,
request_id=request_id,
)
else:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.METHOD_NOT_FOUND,
f"Method not found: {method}",
request_id=request_id,
)
except Exception as e:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
str(e),
request_id=request_id,
)
finally:
if should_close:
_env.close()
# Register MCP WebSocket endpoint (available in both production and simulation modes)
@app.websocket("/mcp")
async def mcp_websocket_endpoint(websocket: WebSocket):
"""
WebSocket endpoint for MCP JSON-RPC requests.
Each WebSocket connection gets its own environment instance for MCP operations.
Message Protocol:
- Client sends: JSON-RPC 2.0 request (tools/list, tools/call)
- Server responds: JSON-RPC 2.0 response (result or error)
"""
await websocket.accept()
session_id = None
session_env = None
try:
# Create session with dedicated environment
session_id, session_env = await self._create_session()
while True:
# Receive message from client
raw_message = await websocket.receive_text()
try:
jsonrpc_dict = json.loads(raw_message)
jsonrpc_request = JsonRpcRequest(**jsonrpc_dict)
except json.JSONDecodeError as e:
error_resp = JsonRpcResponse.error_response(
JsonRpcErrorCode.PARSE_ERROR,
f"Parse error: {e}",
)
await websocket.send_text(error_resp.model_dump_json())
continue
except ValidationError as e:
error_resp = JsonRpcResponse.error_response(
JsonRpcErrorCode.INVALID_REQUEST,
f"Invalid request: {e}",
)
await websocket.send_text(error_resp.model_dump_json())
continue
try:
# Call mcp_handler with session environment
response = await mcp_handler(
jsonrpc_request, session_env=session_env
)
await websocket.send_text(response.model_dump_json())
except Exception as e:
error_resp = JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
str(e),
request_id=jsonrpc_request.id,
)
await websocket.send_text(error_resp.model_dump_json())
except WebSocketDisconnect:
pass
except SessionCapacityError as e:
error_resp = JsonRpcResponse.error_response(
JsonRpcErrorCode.SERVER_ERROR,
str(e),
data={
"active_sessions": e.active_sessions,
"max_sessions": e.max_sessions,
},
)
await websocket.send_text(error_resp.model_dump_json())
except EnvironmentFactoryError as e:
error_resp = JsonRpcResponse.error_response(
JsonRpcErrorCode.SERVER_ERROR,
str(e),
data={"factory_name": e.factory_name},
)
await websocket.send_text(error_resp.model_dump_json())
except Exception as e:
error_resp = JsonRpcResponse.error_response(
JsonRpcErrorCode.SERVER_ERROR,
str(e),
)
await websocket.send_text(error_resp.model_dump_json())
finally:
if session_id:
await self._destroy_session(session_id)
try:
await websocket.close()
except RuntimeError:
pass
# Register simulation control routes only in simulation mode
if mode == ServerMode.SIMULATION:
@app.post(
"/reset",
response_model=ResetResponse,
tags=["Environment Control"],
summary="Reset the environment",
description="""
Reset the environment to its initial state and return the first observation.
You can optionally provide a seed for reproducibility and an episode_id for tracking.
""",
responses={
200: {
"description": "Environment reset successfully",
"content": {
"application/json": {
"example": {
"observation": {"status": "ready", "data": {}},
"reward": None,
"done": False,
}
}
},
}
},
)
async def reset(
request: ResetRequest = Body(default_factory=ResetRequest),
) -> ResetResponse:
return await reset_handler(request)
@app.post(
"/step",
response_model=StepResponse,
tags=["Environment Control"],
summary="Execute an action in the environment",
description="""
Execute an action in the environment and receive the resulting observation.
The action must conform to the environment's action schema, which can be
retrieved from the `/schema` endpoint. If the action is invalid,
the endpoint will return HTTP 422 with detailed validation errors.
The response includes:
- **observation**: The environment's response to the action
- **reward**: Optional reward signal (float or None)
- **done**: Boolean indicating if the episode has terminated
""",
responses={
200: {
"description": "Action executed successfully",
"content": {
"application/json": {
"example": {
"observation": {"status": "success", "data": {}},
"reward": 1.0,
"done": False,
}
}
},
},
422: {
"description": "Validation error - invalid action format or values",
"content": {
"application/json": {
"example": {
"detail": [
{
"type": "string_too_short",
"loc": ["body", "action", "message"],
"msg": "String should have at least 1 character",
"input": "",
}
]
}
}
},
},
500: {
"description": "Internal server error during action execution"
},
},
)
async def step(request: StepRequest) -> StepResponse:
return await step_handler(request)
def get_state_handler() -> State:
_env = self._env_factory()
try:
return _env.state
finally:
_env.close()
def get_metadata_handler() -> EnvironmentMetadata:
_env = self._env_factory()
try:
return _env.get_metadata()
finally:
_env.close()
# Build list of GET endpoints based on mode
get_endpoints = [
GetEndpointConfig(
path="/metadata",
handler=get_metadata_handler,
response_model=EnvironmentMetadata,
tag="Environment Info",
summary="Get environment metadata",
description="""
Get metadata about this environment.
Returns information about the environment including name, description,
version, author, and documentation links.
""",
),
GetEndpointConfig(
path="/health",
handler=lambda: HealthResponse(status=HealthStatus.HEALTHY),
response_model=HealthResponse,
tag="Health",
summary="Health check",
description="Check if the environment server is running and healthy.",
),
]
# Only register /state endpoint in simulation mode
if mode == ServerMode.SIMULATION:
get_endpoints.insert(
0,
GetEndpointConfig(
path="/state",
handler=get_state_handler,
response_model=State,
tag="State Management",
summary="Get current environment state",
description="""
Retrieve the current internal state of the environment.
The structure of the state object is defined by the environment's State model.
""",
),
)
register_get_endpoints(app, get_endpoints)
# Register combined schema endpoint
@app.get(
"/schema",
response_model=SchemaResponse,
tags=["Schema"],
summary="Get all JSON schemas",
description="""
Get JSON schemas for actions, observations, and state in a single response.
Returns a combined schema object containing:
- **action**: JSON schema for actions accepted by this environment
- **observation**: JSON schema for observations returned by this environment
- **state**: JSON schema for environment state objects
This is more efficient than calling individual schema endpoints and provides
all schema information needed to interact with the environment.
""",
responses={
200: {
"description": "Combined schemas retrieved successfully",
"content": {
"application/json": {
"example": {
"action": {
"type": "object",
"properties": {"message": {"type": "string"}},
},
"observation": {
"type": "object",
"properties": {"response": {"type": "string"}},
},
"state": {
"type": "object",
"properties": {"step_count": {"type": "integer"}},
},
}
}
},
}
},
)
async def get_schemas() -> SchemaResponse:
"""Return all schemas in one response."""
return SchemaResponse(
action=self.action_cls.model_json_schema(),
observation=self.observation_cls.model_json_schema(),
state=State.model_json_schema(),
)
# Register MCP endpoint for production mode (direct MCP access)
@app.post("/mcp")
async def mcp_endpoint(request_raw: Request) -> Dict[str, Any]:
"""
MCP JSON-RPC endpoint for production mode.
Bypasses step() overhead and provides direct access to MCP tools.
Supports tools/list and tools/call methods.
"""
# Parse JSON manually to handle parse errors gracefully
try:
body = await request_raw.body()
request_dict = json.loads(body)
request = JsonRpcRequest(**request_dict)
except json.JSONDecodeError:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.PARSE_ERROR
).model_dump()
except ValidationError as e:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INVALID_REQUEST,
f"Invalid request: {e}",
).model_dump()
except Exception:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.PARSE_ERROR
).model_dump()
method = request.method
params = request.params
request_id = request.id
# Create a temporary environment for MCP access
_env = self._env_factory()
try:
# Check if environment supports MCP
if not hasattr(_env, "mcp_client") and not hasattr(_env, "mcp_server"):
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
"Environment does not support MCP",
request_id=request_id,
).model_dump()
if method == McpMethod.TOOLS_LIST:
# List tools from MCP server
if hasattr(_env, "mcp_client") and _env.mcp_client:
async with _env.mcp_client:
tools = await _env.mcp_client.list_tools()
return JsonRpcResponse.success(
result={
"tools": [
t.model_dump()
if hasattr(t, "model_dump")
else dict(t)
for t in tools
]
},
request_id=request_id,
).model_dump()
elif hasattr(_env, "mcp_server") and _env.mcp_server:
# Use server directly
tools = []
if hasattr(_env.mcp_server, "_tool_manager"):
tool_manager = _env.mcp_server._tool_manager
if hasattr(tool_manager, "_tools"):
for tool_name, tool in tool_manager._tools.items():
tool_dict = {
"name": tool.name,
"description": tool.description or "",
"inputSchema": tool.parameters or {},
}
tools.append(tool_dict)
return JsonRpcResponse.success(
result={"tools": tools},
request_id=request_id,
).model_dump()
else:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
"MCP server not available",
request_id=request_id,
).model_dump()
elif method == McpMethod.TOOLS_CALL:
tool_name = params.get("name")
arguments = params.get("arguments", {})
if not tool_name:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INVALID_PARAMS,
"Invalid params - 'name' is required",
request_id=request_id,
).model_dump()
# Call tool via MCP
if hasattr(_env, "mcp_client") and _env.mcp_client:
async with _env.mcp_client:
result = await _env.mcp_client.call_tool(
name=tool_name, arguments=arguments
)
elif hasattr(_env, "mcp_server") and hasattr(
_env.mcp_server, "_tool_manager"
):
# Call tool directly on FastMCP server
tool_manager = _env.mcp_server._tool_manager
if tool_name in tool_manager._tools:
tool = tool_manager._tools[tool_name]
result = tool.fn(**arguments)
else:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INVALID_PARAMS,
f"Tool not found: {tool_name}",
request_id=request_id,
).model_dump()
else:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
"MCP server not available",
request_id=request_id,
).model_dump()
# Make result JSON serializable
serializable_result = _make_json_serializable(result)
return JsonRpcResponse.success(
result=serializable_result,
request_id=request_id,
).model_dump()
else:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.METHOD_NOT_FOUND,
f"Method not found: {method}",
request_id=request_id,
).model_dump()
except Exception as e:
return JsonRpcResponse.error_response(
JsonRpcErrorCode.INTERNAL_ERROR,
str(e),
request_id=request_id,
).model_dump()
finally:
_env.close()
# Register WebSocket endpoint for persistent sessions
@app.websocket("/ws")
async def websocket_endpoint(websocket: WebSocket):
"""
WebSocket endpoint for persistent environment sessions.
Each WebSocket connection gets its own environment instance.
Message Protocol:
- Client sends: WSResetMessage | WSStepMessage | WSStateMessage | WSCloseMessage
- Server responds: WSObservationResponse | WSStateResponse | WSErrorResponse
"""
await websocket.accept()
session_id = None
session_env = None
try:
# Create session with dedicated environment
session_id, session_env = await self._create_session()
while True:
# Receive message from client
raw_message = await websocket.receive_text()
try:
message_dict = json.loads(raw_message)
except json.JSONDecodeError as e:
error_resp = WSErrorResponse(
data={
"message": f"Invalid JSON: {e}",
"code": WSErrorCode.INVALID_JSON,
}
)
await websocket.send_text(error_resp.model_dump_json())
continue
msg_type = message_dict.get("type", "")
try:
match msg_type:
case "reset":
msg = WSResetMessage(**message_dict)
is_async = (
session_env.reset_async.__func__
is not Environment.reset_async
)
if is_async:
sig = inspect.signature(session_env.reset_async)
valid_kwargs = self._get_valid_kwargs(sig, msg.data)
observation = await session_env.reset_async(
**valid_kwargs
)
else:
sig = inspect.signature(session_env.reset)
valid_kwargs = self._get_valid_kwargs(sig, msg.data)
observation = await self._run_in_session_executor(
session_id, session_env.reset, **valid_kwargs
)
self._update_session_activity(session_id)
response = WSObservationResponse(
data=serialize_observation(observation),
)
case "step":
msg = WSStepMessage(**message_dict)
action = deserialize_action(msg.data, self.action_cls)
is_async = (
session_env.step_async.__func__
is not Environment.step_async
)
if is_async:
observation = await session_env.step_async(action)
else:
observation = await self._run_in_session_executor(
session_id, session_env.step, action
)
self._update_session_activity(
session_id, increment_step=True
)
response = WSObservationResponse(
data=serialize_observation(observation)
)
case "state":
msg = WSStateMessage(**message_dict)
state = session_env.state
if hasattr(state, "model_dump"):
state_data = state.model_dump()
else:
state_data = dict(state) if state else {}
response = WSStateResponse(data=state_data)
case "close":
msg = WSCloseMessage(**message_dict)
break
case "mcp":
msg = WSMCPMessage(**message_dict)
try:
rpc_request = JsonRpcRequest(**msg.data)
except (ValidationError, Exception) as e:
rpc_response = JsonRpcResponse.error_response(
JsonRpcErrorCode.INVALID_REQUEST,
f"Invalid request: {e}",
)
else:
rpc_response = await mcp_handler(
rpc_request,
session_env=session_env,
)
response = WSMCPResponse(data=rpc_response.model_dump())
case _:
response = WSErrorResponse(
data={
"message": f"Unknown message type: {msg_type}",
"code": WSErrorCode.UNKNOWN_TYPE,
}
)
await websocket.send_text(response.model_dump_json())
except ValidationError as e:
error_resp = WSErrorResponse(
data={
"message": "Invalid message",
"code": WSErrorCode.VALIDATION_ERROR,
"errors": e.errors(),
}
)
await websocket.send_text(error_resp.model_dump_json())
except Exception as e:
error_resp = WSErrorResponse(
data={
"message": str(e),
"code": WSErrorCode.EXECUTION_ERROR,
}
)
await websocket.send_text(error_resp.model_dump_json())
except WebSocketDisconnect:
pass
except SessionCapacityError as e:
error_resp = WSErrorResponse(
data={
"message": str(e),
"code": WSErrorCode.CAPACITY_REACHED,
"active_sessions": e.active_sessions,
"max_sessions": e.max_sessions,
}
)
await websocket.send_text(error_resp.model_dump_json())
except EnvironmentFactoryError as e:
error_resp = WSErrorResponse(
data={
"message": str(e),
"code": WSErrorCode.FACTORY_ERROR,
"factory_name": e.factory_name,
}
)
await websocket.send_text(error_resp.model_dump_json())
except Exception as e:
error_resp = WSErrorResponse(
data={"message": str(e), "code": WSErrorCode.SESSION_ERROR}
)
await websocket.send_text(error_resp.model_dump_json())
finally:
if session_id:
await self._destroy_session(session_id)
try:
await websocket.close()
except RuntimeError:
pass
def create_app(
env: Callable[[], Environment],
action_cls: Type[Action],
observation_cls: Type[Observation],
env_name: Optional[str] = None,
max_concurrent_envs: Optional[int] = None,
concurrency_config: Optional[ConcurrencyConfig] = None,
gradio_builder: Optional[Callable[..., Any]] = None,
) -> FastAPI:
"""
Create a FastAPI application with or without web interface.
This function creates a FastAPI app with the web interface enabled by default,
including README integration for better user experience.
Args:
env: Environment factory (callable) that creates new instances
action_cls: The Action subclass this environment expects
observation_cls: The Observation subclass this environment returns
env_name: Optional environment name for README loading
max_concurrent_envs: Maximum concurrent WebSocket sessions.
Mutually exclusive with concurrency_config.
concurrency_config: Optional ConcurrencyConfig for advanced concurrency settings.
Mutually exclusive with max_concurrent_envs.
gradio_builder: Optional callable to build a custom Gradio UI at /web.
Signature: (web_manager, action_fields, metadata, is_chat_env, title,
quick_start_md) -> gr.Blocks. When None, the default Gradio app is used.
See docs/customizing-web-ui.md.
Returns:
FastAPI application instance with or without web interface and README integration
"""
# Check if web interface should be enabled
# This can be controlled via environment variable or build argument
enable_web = os.getenv("ENABLE_WEB_INTERFACE", "false").lower() in (
"true",
"1",
"yes",
)
if enable_web:
# Gradio-based web UI (gradio is a core dependency)
from .web_interface import create_web_interface_app
return create_web_interface_app(
env,
action_cls,
observation_cls,
env_name,
max_concurrent_envs,
concurrency_config,
gradio_builder=gradio_builder,
)
else:
# Use standard FastAPI app without web interface
return create_fastapi_app(
env, action_cls, observation_cls, max_concurrent_envs, concurrency_config
)
def create_fastapi_app(
env: Callable[[], Environment],
action_cls: Type[Action],
observation_cls: Type[Observation],
max_concurrent_envs: Optional[int] = None,
concurrency_config: Optional[ConcurrencyConfig] = None,
) -> FastAPI:
"""
Create a FastAPI application with comprehensive documentation.
Args:
env: Environment factory (callable) that creates new instances
action_cls: The Action subclass this environment expects
observation_cls: The Observation subclass this environment returns
max_concurrent_envs: Maximum concurrent WebSocket sessions.
Mutually exclusive with concurrency_config.
concurrency_config: Optional ConcurrencyConfig for advanced concurrency settings.
Mutually exclusive with max_concurrent_envs.
Returns:
FastAPI application instance
"""
try:
from fastapi import FastAPI
except ImportError:
raise ImportError(
"FastAPI is required. Install with: pip install fastapi uvicorn"
)
app = FastAPI(
title="OpenEnv Environment HTTP API",
version="1.0.0",
description="""
# OpenEnv Environment HTTP API
HTTP API for interacting with OpenEnv environments through a standardized interface.
## Features
* **Environment Reset**: Initialize or restart episodes
* **Action Execution**: Send actions and receive observations
* **State Inspection**: Query current environment state
* **Schema Access**: Retrieve JSON schemas for actions and observations
## Workflow
1. Call `/reset` to start a new episode and get initial observation
2. Call `/step` repeatedly with actions to interact with environment
3. Episode ends when observation returns `done: true`
4. Call `/state` anytime to inspect current environment state
## Documentation
* **Swagger UI**: Available at `/docs`
* **ReDoc**: Available at `/redoc`
* **OpenAPI Schema**: Available at `/openapi.json`
""",
openapi_tags=[
{
"name": "Environment Control",
"description": "Core operations for environment interaction (reset, step)",
},
{
"name": "State Management",
"description": "Operations for inspecting environment state",
},
{
"name": "Environment Info",
"description": "Information about the environment",
},
{
"name": "Schema",
"description": "JSON Schema endpoints for actions, observations, and state",
},
{"name": "Health", "description": "Service health and status checks"},
],
docs_url="/docs",
redoc_url="/redoc",
openapi_url="/openapi.json",
contact={
"name": "OpenEnv Team",
"url": "https://github.com/meta-pytorch/OpenEnv",
},
license_info={
"name": "BSD-3-Clause",
"url": "https://github.com/meta-pytorch/OpenEnv/blob/main/LICENSE",
},
)
server = HTTPEnvServer(
env,
action_cls,
observation_cls,
max_concurrent_envs,
concurrency_config=concurrency_config,
)
server.register_routes(app)
return app