Eddy3D-GAN / api.py
google-labs-jules[bot]
⚡ Bolt: [performance improvement] Optimize math normalization in ONNX inference output
931073c
"""
Eddy3D GAN Wind Prediction API
Serves the GAN surrogate model for urban wind flow prediction.
Accepts 512x512 input images or raw float arrays and returns predicted wind speeds.
"""
import os
import io
import base64
import gzip
import logging
import time
from contextlib import asynccontextmanager
from typing import Optional
import concurrent.futures
import numpy as np
from PIL import Image
from fastapi import FastAPI, Request, HTTPException
from fastapi.middleware.cors import CORSMiddleware
from fastapi.responses import JSONResponse
from pydantic import BaseModel
import onnxruntime as rt
from slowapi import Limiter, _rate_limit_exceeded_handler
from slowapi.util import get_remote_address
from slowapi.errors import RateLimitExceeded
# ---------------------------------------------------------------------------
# Configuration
# ---------------------------------------------------------------------------
MODEL_PATH = os.environ.get("MODEL_PATH", "model.onnx")
PORT = int(os.environ.get("PORT", "8000"))
LOG_LEVEL = os.environ.get("LOG_LEVEL", "INFO").upper()
ALLOWED_ORIGINS = os.environ.get("ALLOWED_ORIGINS", "*").split(",")
RATE_LIMIT_REQUESTS = os.environ.get("RATE_LIMIT_REQUESTS", "30")
RATE_LIMIT_WINDOW = os.environ.get("RATE_LIMIT_WINDOW", "minute")
MAX_UPLOAD_BYTES = 10 * 1024 * 1024 # 10 MB
logging.basicConfig(level=LOG_LEVEL, format="%(asctime)s %(levelname)s %(message)s")
logger = logging.getLogger(__name__)
# ---------------------------------------------------------------------------
# Turbo colormap (256 entries, RGB in [0, 1])
# Used for reverse-mapping predicted pixel colours to wind speed values.
# ---------------------------------------------------------------------------
TURBO_COLORMAP = np.array([
[0.18995, 0.07176, 0.23217], [0.19483, 0.08339, 0.26149],
[0.19956, 0.09498, 0.29024], [0.20415, 0.10652, 0.31844],
[0.20860, 0.11802, 0.34607], [0.21291, 0.12947, 0.37314],
[0.21708, 0.14087, 0.39964], [0.22111, 0.15223, 0.42558],
[0.22500, 0.16354, 0.45096], [0.22875, 0.17481, 0.47578],
[0.23236, 0.18603, 0.50004], [0.23582, 0.19720, 0.52373],
[0.23915, 0.20833, 0.54686], [0.24234, 0.21941, 0.56942],
[0.24539, 0.23044, 0.59142], [0.24830, 0.24143, 0.61286],
[0.25107, 0.25237, 0.63374], [0.25369, 0.26327, 0.65406],
[0.25618, 0.27412, 0.67381], [0.25853, 0.28492, 0.69300],
[0.26074, 0.29568, 0.71162], [0.26280, 0.30639, 0.72968],
[0.26473, 0.31706, 0.74718], [0.26652, 0.32768, 0.76412],
[0.26816, 0.33825, 0.78050], [0.26967, 0.34878, 0.79631],
[0.27103, 0.35926, 0.81156], [0.27226, 0.36970, 0.82624],
[0.27334, 0.38008, 0.84037], [0.27429, 0.39043, 0.85393],
[0.27509, 0.40072, 0.86692], [0.27576, 0.41097, 0.87936],
[0.27628, 0.42118, 0.89123], [0.27667, 0.43134, 0.90254],
[0.27691, 0.44145, 0.91328], [0.27701, 0.45152, 0.92347],
[0.27698, 0.46153, 0.93309], [0.27680, 0.47151, 0.94214],
[0.27648, 0.48144, 0.95064], [0.27603, 0.49132, 0.95857],
[0.27543, 0.50115, 0.96594], [0.27469, 0.51094, 0.97275],
[0.27381, 0.52069, 0.97899], [0.27273, 0.53040, 0.98461],
[0.27106, 0.54015, 0.98930], [0.26878, 0.54995, 0.99303],
[0.26592, 0.55979, 0.99583], [0.26252, 0.56967, 0.99773],
[0.25862, 0.57958, 0.99876], [0.25425, 0.58950, 0.99896],
[0.24946, 0.59943, 0.99835], [0.24427, 0.60937, 0.99697],
[0.23874, 0.61931, 0.99485], [0.23288, 0.62923, 0.99202],
[0.22676, 0.63913, 0.98851], [0.22039, 0.64901, 0.98436],
[0.21382, 0.65886, 0.97959], [0.20708, 0.66866, 0.97423],
[0.20021, 0.67842, 0.96833], [0.19326, 0.68812, 0.96190],
[0.18625, 0.69775, 0.95498], [0.17923, 0.70732, 0.94761],
[0.17223, 0.71680, 0.93981], [0.16529, 0.72620, 0.93161],
[0.15844, 0.73551, 0.92305], [0.15173, 0.74472, 0.91416],
[0.14519, 0.75381, 0.90496], [0.13886, 0.76279, 0.89550],
[0.13278, 0.77165, 0.88580], [0.12698, 0.78037, 0.87590],
[0.12151, 0.78896, 0.86581], [0.11639, 0.79740, 0.85559],
[0.11167, 0.80569, 0.84525], [0.10738, 0.81381, 0.83484],
[0.10357, 0.82177, 0.82437], [0.10026, 0.82955, 0.81389],
[0.09750, 0.83714, 0.80342], [0.09532, 0.84455, 0.79299],
[0.09377, 0.85175, 0.78264], [0.09287, 0.85875, 0.77240],
[0.09267, 0.86554, 0.76230], [0.09320, 0.87211, 0.75237],
[0.09451, 0.87844, 0.74265], [0.09662, 0.88454, 0.73316],
[0.09958, 0.89040, 0.72393], [0.10342, 0.89600, 0.71500],
[0.10815, 0.90142, 0.70599], [0.11374, 0.90673, 0.69651],
[0.12014, 0.91193, 0.68660], [0.12733, 0.91701, 0.67627],
[0.13526, 0.92197, 0.66556], [0.14391, 0.92680, 0.65448],
[0.15323, 0.93151, 0.64308], [0.16319, 0.93609, 0.63137],
[0.17377, 0.94053, 0.61938], [0.18491, 0.94484, 0.60713],
[0.19659, 0.94901, 0.59466], [0.20877, 0.95304, 0.58199],
[0.22142, 0.95692, 0.56914], [0.23449, 0.96065, 0.55614],
[0.24797, 0.96423, 0.54303], [0.26180, 0.96765, 0.52981],
[0.27597, 0.97092, 0.51653], [0.29042, 0.97403, 0.50321],
[0.30513, 0.97697, 0.48987], [0.32006, 0.97974, 0.47654],
[0.33517, 0.98234, 0.46325], [0.35043, 0.98477, 0.45002],
[0.36581, 0.98702, 0.43688], [0.38127, 0.98909, 0.42386],
[0.39678, 0.99098, 0.41098], [0.41229, 0.99268, 0.39826],
[0.42778, 0.99419, 0.38575], [0.44321, 0.99551, 0.37345],
[0.45854, 0.99663, 0.36140], [0.47375, 0.99755, 0.34963],
[0.48879, 0.99828, 0.33816], [0.50362, 0.99879, 0.32701],
[0.51822, 0.99910, 0.31622], [0.53255, 0.99919, 0.30581],
[0.54658, 0.99907, 0.29581], [0.56026, 0.99873, 0.28623],
[0.57357, 0.99817, 0.27712], [0.58646, 0.99739, 0.26849],
[0.59891, 0.99638, 0.26038], [0.61088, 0.99514, 0.25280],
[0.62233, 0.99366, 0.24579], [0.63323, 0.99195, 0.23937],
[0.64362, 0.98999, 0.23356], [0.65394, 0.98775, 0.22835],
[0.66428, 0.98524, 0.22370], [0.67462, 0.98246, 0.21960],
[0.68494, 0.97941, 0.21602], [0.69525, 0.97610, 0.21294],
[0.70553, 0.97255, 0.21032], [0.71577, 0.96875, 0.20815],
[0.72596, 0.96470, 0.20640], [0.73610, 0.96043, 0.20504],
[0.74617, 0.95593, 0.20406], [0.75617, 0.95121, 0.20343],
[0.76608, 0.94627, 0.20311], [0.77591, 0.94113, 0.20310],
[0.78563, 0.93579, 0.20336], [0.79524, 0.93025, 0.20386],
[0.80473, 0.92452, 0.20459], [0.81410, 0.91861, 0.20552],
[0.82333, 0.91253, 0.20663], [0.83241, 0.90627, 0.20788],
[0.84133, 0.89986, 0.20926], [0.85010, 0.89328, 0.21074],
[0.85868, 0.88655, 0.21230], [0.86709, 0.87968, 0.21391],
[0.87530, 0.87267, 0.21555], [0.88331, 0.86553, 0.21719],
[0.89112, 0.85826, 0.21880], [0.89870, 0.85087, 0.22038],
[0.90605, 0.84337, 0.22188], [0.91317, 0.83576, 0.22328],
[0.92004, 0.82806, 0.22456], [0.92666, 0.82025, 0.22570],
[0.93301, 0.81236, 0.22667], [0.93909, 0.80439, 0.22744],
[0.94489, 0.79634, 0.22800], [0.95039, 0.78823, 0.22831],
[0.95560, 0.78005, 0.22836], [0.96049, 0.77181, 0.22811],
[0.96507, 0.76352, 0.22754], [0.96931, 0.75519, 0.22663],
[0.97323, 0.74682, 0.22536], [0.97679, 0.73842, 0.22369],
[0.98000, 0.73000, 0.22161], [0.98289, 0.72140, 0.21918],
[0.98549, 0.71250, 0.21650], [0.98781, 0.70330, 0.21358],
[0.98986, 0.69382, 0.21043], [0.99163, 0.68408, 0.20706],
[0.99314, 0.67408, 0.20348], [0.99438, 0.66386, 0.19971],
[0.99535, 0.65341, 0.19577], [0.99607, 0.64277, 0.19165],
[0.99654, 0.63193, 0.18738], [0.99675, 0.62093, 0.18297],
[0.99672, 0.60977, 0.17842], [0.99644, 0.59846, 0.17376],
[0.99593, 0.58703, 0.16899], [0.99517, 0.57549, 0.16412],
[0.99419, 0.56386, 0.15918], [0.99297, 0.55214, 0.15417],
[0.99153, 0.54036, 0.14910], [0.98987, 0.52854, 0.14398],
[0.98799, 0.51667, 0.13883], [0.98590, 0.50479, 0.13367],
[0.98360, 0.49291, 0.12849], [0.98108, 0.48104, 0.12332],
[0.97837, 0.46920, 0.11817], [0.97545, 0.45740, 0.11305],
[0.97234, 0.44565, 0.10797], [0.96904, 0.43399, 0.10294],
[0.96555, 0.42241, 0.09798], [0.96187, 0.41093, 0.09310],
[0.95801, 0.39958, 0.08831], [0.95398, 0.38836, 0.08362],
[0.94977, 0.37729, 0.07905], [0.94538, 0.36638, 0.07461],
[0.94084, 0.35566, 0.07031], [0.93612, 0.34513, 0.06616],
[0.93125, 0.33482, 0.06218], [0.92623, 0.32473, 0.05837],
[0.92105, 0.31489, 0.05475], [0.91572, 0.30530, 0.05134],
[0.91024, 0.29599, 0.04814], [0.90463, 0.28696, 0.04516],
[0.89888, 0.27824, 0.04243], [0.89298, 0.26981, 0.03993],
[0.88691, 0.26152, 0.03753], [0.88066, 0.25334, 0.03521],
[0.87422, 0.24526, 0.03297], [0.86760, 0.23730, 0.03082],
[0.86079, 0.22945, 0.02875], [0.85380, 0.22170, 0.02677],
[0.84662, 0.21407, 0.02487], [0.83926, 0.20654, 0.02305],
[0.83172, 0.19912, 0.02131], [0.82399, 0.19182, 0.01966],
[0.81608, 0.18462, 0.01809], [0.80799, 0.17753, 0.01660],
[0.79971, 0.17055, 0.01520], [0.79125, 0.16368, 0.01387],
[0.78260, 0.15693, 0.01264], [0.77377, 0.15028, 0.01148],
[0.76476, 0.14374, 0.01041], [0.75556, 0.13731, 0.00942],
[0.74617, 0.13098, 0.00851], [0.73661, 0.12477, 0.00769],
[0.72686, 0.11867, 0.00695], [0.71692, 0.11268, 0.00629],
[0.70680, 0.10680, 0.00571], [0.69650, 0.10102, 0.00522],
[0.68602, 0.09536, 0.00481], [0.67535, 0.08980, 0.00449],
[0.66449, 0.08436, 0.00424], [0.65345, 0.07902, 0.00408],
[0.64223, 0.07380, 0.00401], [0.63082, 0.06868, 0.00401],
[0.61923, 0.06367, 0.00410], [0.60746, 0.05878, 0.00427],
[0.59550, 0.05399, 0.00453], [0.58336, 0.04931, 0.00486],
[0.57103, 0.04474, 0.00529], [0.55852, 0.04028, 0.00579],
[0.54583, 0.03593, 0.00638], [0.53295, 0.03169, 0.00705],
[0.51989, 0.02756, 0.00780], [0.50664, 0.02354, 0.00863],
[0.49321, 0.01963, 0.00955], [0.47960, 0.01583, 0.01055],
], dtype=np.float32) # shape (256, 3)
# ---------------------------------------------------------------------------
# Pre-computed 1D lookup table: 24-bit RGB -> windspeed float32
# Built once at import time (~0.3s, 64 MB). Replaces the per-request
# 3D index lookup and float multiplication with a single 1D array view lookup.
# ---------------------------------------------------------------------------
def _build_colormap_lut_1d() -> np.ndarray:
"""Build a (16777216,) float32 LUT mapping every 24-bit RGB value to
its corresponding wind speed (0-15 m/s) based on the nearest Turbo-colormap index."""
turbo_f32 = (TURBO_COLORMAP * 255.0).astype(np.float32) # (256, 3)
half_l2 = 0.5 * np.sum(turbo_f32 ** 2, axis=1) # (256,)
# ⚡ Bolt Optimization: Write directly to a 1D array to save reshaping overhead.
# We pre-allocate loop arrays `pixels` and `scores` to perform calculations
# in-place using np.matmul and np.subtract, significantly reducing memory allocations
# and saving ~2-3 seconds of initialization time.
lut_1d = np.empty(16777216, dtype=np.uint8)
# ⚡ Bolt Optimization: Factor out matrix computations that are constant across the R loop.
# The G and B channels iterate over the exact same 65536 combinations for every R value.
# We can pre-calculate `G * C_g + B * C_b - 0.5 * ||C||^2` outside the loop, saving ~4 seconds of API startup time.
gb = np.mgrid[0:256, 0:256].astype(np.float32).reshape(2, -1).T # (65536, 2)
gb_scores = np.matmul(gb, turbo_f32[:, 1:].T)
np.subtract(gb_scores, half_l2, out=gb_scores)
# Extract the R channel of the colormap for the loop
turbo_r = turbo_f32[:, 0]
import concurrent.futures
# ⚡ Bolt Optimization: Completely eliminate the python for-loop and repeated matmul calls
# by using thread pool and processing chunks of the data in parallel. Since numpy
# releases the GIL for core operations like matmul, this uses all CPU cores
# and reduces the startup time by ~30% (~2-3 seconds saved).
def process_chunk(start_r, end_r):
scores_local = np.empty((65536, 256), dtype=np.float32)
for r in range(start_r, end_r):
# Inside the inner loop, we just need to add the pre-calculated GB scores to the current R score
np.add(gb_scores, r * turbo_r, out=scores_local)
lut_1d[r*65536:(r+1)*65536] = np.argmax(scores_local, axis=1)
num_threads = os.cpu_count() or 4
# Ensure num_threads divides 256 evenly for simplicity, max 8 to control memory
num_threads = min(8, num_threads)
while 256 % num_threads != 0:
num_threads -= 1
chunk_size = 256 // num_threads
with concurrent.futures.ThreadPoolExecutor(max_workers=num_threads) as executor:
futures = []
for i in range(num_threads):
futures.append(executor.submit(process_chunk, i*chunk_size, (i+1)*chunk_size))
# Ensure any exceptions in worker threads are raised
for f in futures:
f.result()
# Convert the 1D uint8 LUT into a flat 1D float32 array
# This completely eliminates index multiplication and float cast at inference time
return (lut_1d * (15.0 / 256)).astype(np.float32)
COLORMAP_LUT_1D_FLOAT = _build_colormap_lut_1d()
def _color_to_windspeed(denorm: np.ndarray) -> np.ndarray:
"""Map denormalised uint8 output to wind speed via pre-computed LUT.
Args:
denorm: (H, W, 3) uint8 array (the image produced by _run_inference_from_array).
Returns:
Flat float32 array of wind speed values (0–15 m/s), length H*W.
"""
# ⚡ Bolt Optimization: Zero-pad the RGB array to RGBA (4 bytes) and view it
# as a 32-bit integer array. This allows for lightning-fast 1D lookup
# and completely avoids multi-dimensional array slicing and index math.
# We allocate an empty 1D `<u4` array directly and fill the channels via an
# overlaid `uint8` view to avoid the cost of `np.zeros` and `view('<u4').ravel()`.
idx = np.empty(denorm.shape[0] * denorm.shape[1], dtype='<u4')
view = idx.view(np.uint8).reshape(denorm.shape[0], denorm.shape[1], 4)
view[:, :, 0] = denorm[:, :, 2] # B
view[:, :, 1] = denorm[:, :, 1] # G
view[:, :, 2] = denorm[:, :, 0] # R
view[:, :, 3] = 0 # A (Padding)
return COLORMAP_LUT_1D_FLOAT[idx]
# ---------------------------------------------------------------------------
# Rate limiting
# ---------------------------------------------------------------------------
_rate_limit_str = f"{RATE_LIMIT_REQUESTS}/{RATE_LIMIT_WINDOW}"
_rate_limiting_enabled = int(RATE_LIMIT_REQUESTS) > 0
limiter = Limiter(
key_func=get_remote_address,
default_limits=[_rate_limit_str] if _rate_limiting_enabled else [],
enabled=_rate_limiting_enabled,
)
# ---------------------------------------------------------------------------
# Application lifecycle
# ---------------------------------------------------------------------------
# ⚡ Bolt Optimization: A global thread pool to avoid per-request thread churn
# when parallelizing slow I/O bound tasks like gzip compression and PNG saving.
_thread_pool = concurrent.futures.ThreadPoolExecutor(max_workers=8)
@asynccontextmanager
async def lifespan(app: FastAPI):
"""Load ONNX model at startup, release at shutdown."""
logger.info("Loading ONNX model from %s ...", MODEL_PATH)
if not os.path.exists(MODEL_PATH):
logger.error("Model file not found at %s", MODEL_PATH)
raise FileNotFoundError(f"Model file not found at {MODEL_PATH}")
opts = rt.SessionOptions()
opts.graph_optimization_level = rt.GraphOptimizationLevel.ORT_ENABLE_ALL
opts.intra_op_num_threads = 2
opts.inter_op_num_threads = 1
opts.execution_mode = rt.ExecutionMode.ORT_SEQUENTIAL
session = rt.InferenceSession(MODEL_PATH, opts)
input_name = session.get_inputs()[0].name
input_shape = session.get_inputs()[0].shape
output_name = session.get_outputs()[0].name
output_shape = session.get_outputs()[0].shape
logger.info("Model loaded. Input: %s %s Output: %s %s",
input_name, input_shape, output_name, output_shape)
app.state.session = session
app.state.input_name = input_name
app.state.model_loaded = True
yield
app.state.model_loaded = False
logger.info("Shutting down.")
app = FastAPI(
title="Eddy3D GAN Wind Prediction API",
version="2.0.0",
lifespan=lifespan,
)
app.state.limiter = limiter
app.add_exception_handler(RateLimitExceeded, _rate_limit_exceeded_handler)
app.add_middleware(
CORSMiddleware,
allow_origins=ALLOWED_ORIGINS,
allow_credentials=True,
allow_methods=["*"],
allow_headers=["*"],
)
None
def _run_inference_from_array(data: np.ndarray) -> np.ndarray:
"""Run model from a pre-normalised float32 array.
Args:
data: shape (1, 3, 512, 512), float32, values already in [-1, 1].
Returns:
denorm_output: shape (512, 512, 3), values in [0, 255], uint8
"""
session = app.state.session
inputs = {app.state.input_name: data}
outputs = session.run(None, inputs)
raw = outputs[0][0] # (3, 512, 512)
# ⚡ Bolt Optimization: Use in-place array operations on the ONNX output
# directly to avoid allocating any new float32 arrays in memory during
# math operations. The ONNX runtime returns a standard mutable numpy array.
# The final .transpose() returns a memory view.
# We apply (raw * 127.5) + 127.5 instead of (raw + 1.0) * 127.5.
# This evaluates as Fused Multiply-Add logic reducing float addition overhead.
np.multiply(raw, 127.5, out=raw)
np.add(raw, 127.5, out=raw)
np.clip(raw, 0, 255, out=raw)
return raw.transpose((1, 2, 0)).astype(np.uint8)
None
class PredictRequest(BaseModel):
"""Request body for /predict.
The data_b64 field is a base64-encoded, gzip-compressed flat array of
786432 float32 values (3 * 512 * 512) in channel-first order (R, G, B),
with values already normalised to [-1, 1].
"""
data_b64: str
# ---------------------------------------------------------------------------
# Endpoints
# ---------------------------------------------------------------------------
@app.get("/")
async def root():
return {
"service": "Eddy3D GAN Wind Prediction API",
"status": "ok",
"endpoints": ["/health", "/predict"],
}
@app.get("/health")
async def health():
"""Health check (not rate limited)."""
return {
"status": "healthy",
"model_loaded": getattr(app.state, "model_loaded", False),
}
None
@app.post("/predict")
@limiter.limit(_rate_limit_str)
def predict(request: Request, body: PredictRequest):
"""Run GAN inference from a gzip-compressed, base64-encoded float32 array.
Returns JSON:
wind_speeds_b64: base64-encoded, gzip-compressed float32 array (length 262144)
image_base64: base64-encoded PNG of the output image
width: 512
height: 512
"""
t0 = time.perf_counter()
try:
compressed_bytes = base64.b64decode(body.data_b64)
raw_bytes = gzip.decompress(compressed_bytes)
arr = np.frombuffer(raw_bytes, dtype=np.float32)
expected = 3 * 512 * 512
if arr.size != expected:
raise HTTPException(
status_code=400,
detail=f"Expected {expected} floats (3*512*512), got {arr.size}.",
)
# ⚡ Bolt Optimization: Directly reshape array into model input layout (1, 3, 512, 512)
# to avoid calling np.expand_dims and .astype(..., copy=False), which implicitly forces array
# copying even for matching dtypes. np.frombuffer output is already np.float32.
arr = arr.reshape((1, 3, 512, 512))
denorm = _run_inference_from_array(arr)
wind_speeds_arr = _color_to_windspeed(denorm)
# ⚡ Bolt Optimization: Use a ThreadPoolExecutor to run the two slow I/O bound
# tasks (gzip compression and PNG saving) concurrently. Both `gzip.compress` (zlib)
# and PIL's `Image.save` release the Python GIL internally, meaning they can
# execute truly in parallel. This halves the payload preparation time (~90ms -> ~45ms).
def _compress_wind_speeds():
# Use memoryview instead of .tobytes() to avoid allocating
# and copying a new bytes object for the full array before compression.
wind_speeds_bytes = memoryview(wind_speeds_arr)
# Optimization: use compresslevel=1. The default is 9 which is extremely slow
# for a 1MB payload (~500ms -> ~25ms) but only ~2% worse compression.
compressed = gzip.compress(wind_speeds_bytes, compresslevel=1)
return base64.b64encode(compressed).decode("ascii")
def _encode_image():
output_image = Image.fromarray(denorm, "RGB")
buf = io.BytesIO()
# Optimization: use compress_level=1 for PNG saving.
# This saves ~5-10ms per image generation with virtually identical size.
output_image.save(buf, format="PNG", compress_level=1)
return base64.b64encode(buf.getvalue()).decode("ascii")
future_wind = _thread_pool.submit(_compress_wind_speeds)
future_img = _thread_pool.submit(_encode_image)
wind_speeds_b64 = future_wind.result()
image_b64 = future_img.result()
except HTTPException:
raise
except Exception as e:
logger.exception("Inference failed")
raise HTTPException(status_code=500, detail=f"Inference error: {e}")
elapsed = time.perf_counter() - t0
logger.info("Binary inference completed in %.2f s", elapsed)
return JSONResponse({
"wind_speeds_b64": wind_speeds_b64,
"image_base64": image_b64,
"width": 512,
"height": 512,
})