fitleech commited on
Commit
bca414f
·
verified ·
1 Parent(s): be38fe3

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. __pycache__/miner.cpython-310.pyc +0 -0
  2. chute_config.yml +19 -0
  3. miner.py +424 -0
  4. weights.onnx +3 -0
__pycache__/miner.cpython-310.pyc ADDED
Binary file (13.4 kB). View file
 
chute_config.yml ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Image:
2
+ from_base: parachutes/python:3.12
3
+ run_command:
4
+ - pip install --upgrade setuptools wheel
5
+ - pip install 'numpy>=1.23' 'onnxruntime-gpu[cuda,cudnn]>=1.16' 'opencv-python>=4.7' 'pillow>=9.5' 'huggingface_hub>=0.19.4' 'pydantic>=2.0' 'pyyaml>=6.0' 'aiohttp>=3.9'
6
+ - pip install torch torchvision
7
+ set_workdir: /app
8
+
9
+ NodeSelector:
10
+ gpu_count: 1
11
+ min_vram_gb_per_gpu: 16
12
+ max_hourly_price_per_gpu: 1
13
+
14
+ Chute:
15
+ timeout_seconds: 900
16
+ concurrency: 4
17
+ max_instances: 5
18
+ scaling_threshold: 0.5
19
+ shutdown_after_seconds: 288000
miner.py ADDED
@@ -0,0 +1,424 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from pathlib import Path
2
+ import math
3
+
4
+ import cv2
5
+ import numpy as np
6
+ import onnxruntime as ort
7
+ from numpy import ndarray
8
+ from pydantic import BaseModel
9
+
10
+
11
+ class BoundingBox(BaseModel):
12
+ x1: int
13
+ y1: int
14
+ x2: int
15
+ y2: int
16
+ cls_id: int
17
+ conf: float
18
+
19
+
20
+ class TVFrameResult(BaseModel):
21
+ frame_id: int
22
+ boxes: list[BoundingBox]
23
+ keypoints: list[tuple[int, int]]
24
+
25
+
26
+ SIZE = 1280
27
+
28
+
29
+ class Miner:
30
+ def __init__(self, path_hf_repo: Path) -> None:
31
+ model_path = path_hf_repo / "weights.onnx"
32
+ cn_path = model_path.with_name("class_names.txt")
33
+ if cn_path.is_file():
34
+ lines = cn_path.read_text(encoding="utf-8").splitlines()
35
+ self.class_names = [
36
+ ln.strip()
37
+ for ln in lines
38
+ if ln.strip() and not ln.strip().startswith("#")
39
+ ]
40
+ else:
41
+ self.class_names = ["numberplate"]
42
+ print("ORT version:", ort.__version__)
43
+
44
+ try:
45
+ ort.preload_dlls()
46
+ print("onnxruntime.preload_dlls() success")
47
+ except Exception as e:
48
+ print(f"preload_dlls failed: {e}")
49
+
50
+ print("ORT available providers BEFORE session:", ort.get_available_providers())
51
+
52
+ try:
53
+ import torch
54
+ if torch.cuda.is_available():
55
+ print(f"GPU: {torch.cuda.get_device_name(0)}")
56
+ print(f"GPU memory: {torch.cuda.get_device_properties(0).total_mem / 1e9:.1f} GB")
57
+ else:
58
+ print("GPU: CUDA not available via torch")
59
+ except Exception as e:
60
+ print(f"GPU detection failed: {e}")
61
+
62
+ sess_options = ort.SessionOptions()
63
+ sess_options.graph_optimization_level = ort.GraphOptimizationLevel.ORT_ENABLE_ALL
64
+
65
+ try:
66
+ self.session = ort.InferenceSession(
67
+ str(model_path),
68
+ sess_options=sess_options,
69
+ providers=["CUDAExecutionProvider", "CPUExecutionProvider"],
70
+ )
71
+ print("Created ORT session with preferred CUDA provider list")
72
+ except Exception as e:
73
+ print(f"CUDA session creation failed, falling back to CPU: {e}")
74
+ self.session = ort.InferenceSession(
75
+ str(model_path),
76
+ sess_options=sess_options,
77
+ providers=["CPUExecutionProvider"],
78
+ )
79
+
80
+ print("ORT session providers:", self.session.get_providers())
81
+
82
+ for inp in self.session.get_inputs():
83
+ print("INPUT:", inp.name, inp.shape, inp.type)
84
+ for out in self.session.get_outputs():
85
+ print("OUTPUT:", out.name, out.shape, out.type)
86
+
87
+ self.input_name = self.session.get_inputs()[0].name
88
+ self.output_names = [o.name for o in self.session.get_outputs()]
89
+ self.input_shape = self.session.get_inputs()[0].shape
90
+
91
+ self.input_height = self._safe_dim(self.input_shape[2], default=SIZE)
92
+ self.input_width = self._safe_dim(self.input_shape[3], default=SIZE)
93
+
94
+ # Primary pass: alfred001 tuning (optimized for hermestech weights)
95
+ self.conf_thres = 0.23
96
+ self.iou_thres = 0.66
97
+ self.sigma = 0.465
98
+ self.max_det = 300
99
+
100
+ # Conditional tile-pass (trimmed for latency: no hflip, tighter sparse)
101
+ self.sparse_threshold = 3 # fire tiles only if primary returns < this
102
+ self.tile_conf = 0.57
103
+ self.tile_overlap = 0.20
104
+ self.novelty_iou = 0.10
105
+ self.final_max_det = 17
106
+ self.tile_use_hflip = False # skip hflip tile pass to save ~4 forwards
107
+
108
+ self.use_tta = True
109
+
110
+ print(f"ONNX model loaded from: {model_path}")
111
+ print(f"ONNX providers: {self.session.get_providers()}")
112
+ print(f"ONNX input: name={self.input_name}, shape={self.input_shape}")
113
+
114
+ def __repr__(self) -> str:
115
+ return (
116
+ f"ONNXRuntime(session={type(self.session).__name__}, "
117
+ f"providers={self.session.get_providers()})"
118
+ )
119
+
120
+ @staticmethod
121
+ def _safe_dim(value, default: int) -> int:
122
+ return value if isinstance(value, int) and value > 0 else default
123
+
124
+ # ---------- image preprocessing ----------
125
+ def _letterbox(
126
+ self,
127
+ image: ndarray,
128
+ new_shape: tuple[int, int],
129
+ color=(114, 114, 114),
130
+ ) -> tuple[ndarray, float, tuple[float, float]]:
131
+ h, w = image.shape[:2]
132
+ new_w, new_h = new_shape
133
+ ratio = min(new_w / w, new_h / h)
134
+ resized_w = int(round(w * ratio))
135
+ resized_h = int(round(h * ratio))
136
+ if (resized_w, resized_h) != (w, h):
137
+ interp = cv2.INTER_CUBIC if ratio > 1.0 else cv2.INTER_LINEAR
138
+ image = cv2.resize(image, (resized_w, resized_h), interpolation=interp)
139
+ dw = (new_w - resized_w) / 2.0
140
+ dh = (new_h - resized_h) / 2.0
141
+ left = int(round(dw - 0.1))
142
+ right = int(round(dw + 0.1))
143
+ top = int(round(dh - 0.1))
144
+ bottom = int(round(dh + 0.1))
145
+ padded = cv2.copyMakeBorder(
146
+ image, top, bottom, left, right,
147
+ borderType=cv2.BORDER_CONSTANT, value=color,
148
+ )
149
+ return padded, ratio, (dw, dh)
150
+
151
+ def _preprocess(self, image: ndarray):
152
+ img, ratio, pad = self._letterbox(image, (self.input_width, self.input_height))
153
+ img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB).astype(np.float32) / 255.0
154
+ img = np.transpose(img, (2, 0, 1))[None, ...]
155
+ return np.ascontiguousarray(img, dtype=np.float32), ratio, pad
156
+
157
+ @staticmethod
158
+ def _clip_boxes(boxes: np.ndarray, image_size: tuple[int, int]) -> np.ndarray:
159
+ w, h = image_size
160
+ boxes[:, 0] = np.clip(boxes[:, 0], 0, w - 1)
161
+ boxes[:, 1] = np.clip(boxes[:, 1], 0, h - 1)
162
+ boxes[:, 2] = np.clip(boxes[:, 2], 0, w - 1)
163
+ boxes[:, 3] = np.clip(boxes[:, 3], 0, h - 1)
164
+ return boxes
165
+
166
+ # ---------- NMS primitives ----------
167
+ @staticmethod
168
+ def _hard_nms(boxes: np.ndarray, scores: np.ndarray, iou_thresh: float) -> np.ndarray:
169
+ N = len(boxes)
170
+ if N == 0:
171
+ return np.array([], dtype=np.intp)
172
+ boxes = np.asarray(boxes, dtype=np.float32)
173
+ scores = np.asarray(scores, dtype=np.float32)
174
+ order = np.argsort(-scores)
175
+ keep: list[int] = []
176
+ while len(order):
177
+ i = int(order[0])
178
+ keep.append(i)
179
+ if len(order) == 1:
180
+ break
181
+ rest = order[1:]
182
+ xx1 = np.maximum(boxes[i, 0], boxes[rest, 0])
183
+ yy1 = np.maximum(boxes[i, 1], boxes[rest, 1])
184
+ xx2 = np.minimum(boxes[i, 2], boxes[rest, 2])
185
+ yy2 = np.minimum(boxes[i, 3], boxes[rest, 3])
186
+ inter = np.maximum(0.0, xx2 - xx1) * np.maximum(0.0, yy2 - yy1)
187
+ area_i = (boxes[i, 2] - boxes[i, 0]) * (boxes[i, 3] - boxes[i, 1])
188
+ area_r = (boxes[rest, 2] - boxes[rest, 0]) * (boxes[rest, 3] - boxes[rest, 1])
189
+ iou = inter / (area_i + area_r - inter + 1e-7)
190
+ order = rest[iou <= iou_thresh]
191
+ return np.array(keep, dtype=np.intp)
192
+
193
+ def _soft_nms(
194
+ self,
195
+ boxes: np.ndarray,
196
+ scores: np.ndarray,
197
+ sigma: float,
198
+ score_thresh: float = 0.01,
199
+ ) -> tuple[np.ndarray, np.ndarray]:
200
+ N = len(boxes)
201
+ if N == 0:
202
+ return np.array([], dtype=np.intp), np.array([], dtype=np.float32)
203
+ boxes = boxes.astype(np.float32, copy=True)
204
+ scores = scores.astype(np.float32, copy=True)
205
+ order = np.arange(N)
206
+ for i in range(N):
207
+ max_pos = i + int(np.argmax(scores[i:]))
208
+ boxes[[i, max_pos]] = boxes[[max_pos, i]]
209
+ scores[[i, max_pos]] = scores[[max_pos, i]]
210
+ order[[i, max_pos]] = order[[max_pos, i]]
211
+ if i + 1 >= N:
212
+ break
213
+ xx1 = np.maximum(boxes[i, 0], boxes[i + 1:, 0])
214
+ yy1 = np.maximum(boxes[i, 1], boxes[i + 1:, 1])
215
+ xx2 = np.minimum(boxes[i, 2], boxes[i + 1:, 2])
216
+ yy2 = np.minimum(boxes[i, 3], boxes[i + 1:, 3])
217
+ inter = np.maximum(0.0, xx2 - xx1) * np.maximum(0.0, yy2 - yy1)
218
+ area_i = float(
219
+ (boxes[i, 2] - boxes[i, 0]) * (boxes[i, 3] - boxes[i, 1])
220
+ )
221
+ areas_j = (
222
+ np.maximum(0.0, boxes[i + 1:, 2] - boxes[i + 1:, 0])
223
+ * np.maximum(0.0, boxes[i + 1:, 3] - boxes[i + 1:, 1])
224
+ )
225
+ iou = inter / (area_i + areas_j - inter + 1e-7)
226
+ scores[i + 1:] *= np.exp(-(iou ** 2) / sigma)
227
+ mask = scores > score_thresh
228
+ return order[mask], scores[mask]
229
+
230
+ @staticmethod
231
+ def _box_iou_one_to_many(box: np.ndarray, boxes: np.ndarray) -> np.ndarray:
232
+ if len(boxes) == 0:
233
+ return np.zeros(0, dtype=np.float32)
234
+ xx1 = np.maximum(box[0], boxes[:, 0])
235
+ yy1 = np.maximum(box[1], boxes[:, 1])
236
+ xx2 = np.minimum(box[2], boxes[:, 2])
237
+ yy2 = np.minimum(box[3], boxes[:, 3])
238
+ inter = np.maximum(0.0, xx2 - xx1) * np.maximum(0.0, yy2 - yy1)
239
+ area_a = max(0.0, (box[2] - box[0]) * (box[3] - box[1]))
240
+ area_b = np.maximum(0.0, boxes[:, 2] - boxes[:, 0]) * np.maximum(0.0, boxes[:, 3] - boxes[:, 1])
241
+ return inter / (area_a + area_b - inter + 1e-7)
242
+
243
+ # ---------- raw-dets helper ----------
244
+ def _raw_dets(self, image: ndarray, conf: float) -> np.ndarray:
245
+ """Run a single forward pass and return [N, 5] dets in ORIGINAL image coords."""
246
+ x, ratio, (dw, dh) = self._preprocess(image)
247
+ out = self.session.run(self.output_names, {self.input_name: x})[0]
248
+ if out.ndim == 3:
249
+ out = out[0]
250
+ if out.shape[1] < 5:
251
+ return np.zeros((0, 5), dtype=np.float32)
252
+ boxes = out[:, :4].astype(np.float32)
253
+ scores = out[:, 4].astype(np.float32)
254
+ keep = scores >= conf
255
+ boxes, scores = boxes[keep], scores[keep]
256
+ if len(boxes) == 0:
257
+ return np.zeros((0, 5), dtype=np.float32)
258
+ boxes[:, [0, 2]] -= dw
259
+ boxes[:, [1, 3]] -= dh
260
+ boxes /= ratio
261
+ oh, ow = image.shape[:2]
262
+ boxes = self._clip_boxes(boxes, (ow, oh))
263
+ return np.concatenate([boxes, scores[:, None]], axis=1)
264
+
265
+ # ---------- primary pass: soft-NMS + hflip TTA ----------
266
+ def _primary(self, image: ndarray) -> np.ndarray:
267
+ d1 = self._raw_dets(image, self.conf_thres)
268
+ flipped = cv2.flip(image, 1)
269
+ d2 = self._raw_dets(flipped, self.conf_thres)
270
+ if len(d2):
271
+ w = image.shape[1]
272
+ x1 = w - d2[:, 2]
273
+ x2 = w - d2[:, 0]
274
+ d2 = np.stack([x1, d2[:, 1], x2, d2[:, 3], d2[:, 4]], axis=1)
275
+ all_d = np.concatenate([d1, d2], axis=0) if len(d2) else d1
276
+ if len(all_d) == 0:
277
+ return np.zeros((0, 5), dtype=np.float32)
278
+ # soft-NMS, then hard-NMS
279
+ keep_idx, scores = self._soft_nms(all_d[:, :4].copy(), all_d[:, 4].copy(), sigma=self.sigma)
280
+ if len(keep_idx) == 0:
281
+ return np.zeros((0, 5), dtype=np.float32)
282
+ merged = np.concatenate([all_d[keep_idx, :4], scores[:, None]], axis=1)
283
+ keep = self._hard_nms(merged[:, :4], merged[:, 4], self.iou_thres)
284
+ merged = merged[keep]
285
+ if len(merged) > self.max_det:
286
+ merged = merged[np.argsort(-merged[:, 4])[: self.max_det]]
287
+ return merged
288
+
289
+ # ---------- conditional tile pass ----------
290
+ def _tile_augment(self, image: ndarray, primary: np.ndarray) -> np.ndarray:
291
+ """Run 2x2 overlapping tiles + hflip, novelty-merge into primary."""
292
+ oh, ow = image.shape[:2]
293
+ tw, th = ow // 2, oh // 2
294
+ ox, oy = int(tw * self.tile_overlap), int(th * self.tile_overlap)
295
+ tiles = [
296
+ (0, 0, min(ow, tw + ox), min(oh, th + oy)),
297
+ (max(0, tw - ox), 0, ow, min(oh, th + oy)),
298
+ (0, max(0, th - oy), min(ow, tw + ox), oh),
299
+ (max(0, tw - ox), max(0, th - oy), ow, oh),
300
+ ]
301
+ collected: list[np.ndarray] = []
302
+ for x1, y1, x2, y2 in tiles:
303
+ crop = image[y1:y2, x1:x2]
304
+ if crop.size == 0:
305
+ continue
306
+ d = self._raw_dets(crop, self.tile_conf)
307
+ if len(d):
308
+ d[:, 0] += x1
309
+ d[:, 1] += y1
310
+ d[:, 2] += x1
311
+ d[:, 3] += y1
312
+ collected.append(d)
313
+
314
+ # hflip tile pass (skipped when tile_use_hflip=False — saves 4 ONNX forwards)
315
+ if self.tile_use_hflip:
316
+ flipped = cv2.flip(image, 1)
317
+ for x1, y1, x2, y2 in tiles:
318
+ fx1 = ow - x2
319
+ fx2 = ow - x1
320
+ if fx2 <= fx1:
321
+ continue
322
+ crop = flipped[y1:y2, fx1:fx2]
323
+ if crop.size == 0:
324
+ continue
325
+ d = self._raw_dets(crop, self.tile_conf)
326
+ if len(d):
327
+ d_un = d.copy()
328
+ d_un[:, 0] = (ow - (d[:, 2] + fx1))
329
+ d_un[:, 2] = (ow - (d[:, 0] + fx1))
330
+ d_un[:, 1] = d[:, 1] + y1
331
+ d_un[:, 3] = d[:, 3] + y1
332
+ collected.append(d_un)
333
+
334
+ if not collected:
335
+ return primary
336
+
337
+ tile_dets = np.concatenate(collected, axis=0)
338
+ keep = self._hard_nms(tile_dets[:, :4], tile_dets[:, 4], 0.5)
339
+ tile_dets = tile_dets[keep]
340
+
341
+ # Novelty: drop tile boxes that overlap any primary box at IoU >= novelty_iou
342
+ if len(primary) > 0 and len(tile_dets) > 0:
343
+ mask = np.ones(len(tile_dets), dtype=bool)
344
+ for i in range(len(tile_dets)):
345
+ ious = self._box_iou_one_to_many(tile_dets[i, :4], primary[:, :4])
346
+ if len(ious) and np.max(ious) >= self.novelty_iou:
347
+ mask[i] = False
348
+ tile_dets = tile_dets[mask]
349
+
350
+ if len(tile_dets) == 0:
351
+ return primary
352
+
353
+ # Sanity filter: min/max size, aspect ratio
354
+ w = tile_dets[:, 2] - tile_dets[:, 0]
355
+ h = tile_dets[:, 3] - tile_dets[:, 1]
356
+ area = w * h
357
+ ar = np.maximum(w / np.maximum(h, 1e-6), h / np.maximum(w, 1e-6))
358
+ img_area = float(ow * oh)
359
+ ok = (w >= 7) & (h >= 7) & (area >= 85) & (area <= 0.5 * img_area) & (ar <= 10.0)
360
+ tile_dets = tile_dets[ok]
361
+ if len(tile_dets) == 0:
362
+ return primary
363
+
364
+ merged = np.concatenate([primary, tile_dets], axis=0)
365
+ keep = self._hard_nms(merged[:, :4], merged[:, 4], self.iou_thres)
366
+ merged = merged[keep]
367
+ if len(merged) > self.final_max_det:
368
+ merged = merged[np.argsort(-merged[:, 4])[: self.final_max_det]]
369
+ return merged
370
+
371
+ # ---------- single-image predict ----------
372
+ def _predict_single(self, image: ndarray) -> list[BoundingBox]:
373
+ if image is None or not isinstance(image, np.ndarray) or image.ndim != 3:
374
+ return []
375
+ if image.shape[0] <= 0 or image.shape[1] <= 0 or image.shape[2] != 3:
376
+ return []
377
+ if image.dtype != np.uint8:
378
+ image = image.astype(np.uint8)
379
+
380
+ primary = self._primary(image)
381
+ if len(primary) < self.sparse_threshold:
382
+ dets = self._tile_augment(image, primary)
383
+ else:
384
+ dets = primary
385
+
386
+ results: list[BoundingBox] = []
387
+ for row in dets:
388
+ x1, y1, x2, y2, conf = row.tolist()
389
+ if x2 <= x1 or y2 <= y1:
390
+ continue
391
+ results.append(
392
+ BoundingBox(
393
+ x1=int(math.floor(x1)),
394
+ y1=int(math.floor(y1)),
395
+ x2=int(math.ceil(x2)),
396
+ y2=int(math.ceil(y2)),
397
+ cls_id=0,
398
+ conf=float(conf),
399
+ )
400
+ )
401
+ return results
402
+
403
+ # ---------- chute entrypoint ----------
404
+ def predict_batch(
405
+ self,
406
+ batch_images: list[ndarray],
407
+ offset: int,
408
+ n_keypoints: int,
409
+ ) -> list[TVFrameResult]:
410
+ results: list[TVFrameResult] = []
411
+ for frame_number_in_batch, image in enumerate(batch_images):
412
+ try:
413
+ boxes = self._predict_single(image)
414
+ except Exception as e:
415
+ print(f"Inference failed for frame {offset + frame_number_in_batch}: {e}")
416
+ boxes = []
417
+ results.append(
418
+ TVFrameResult(
419
+ frame_id=offset + frame_number_in_batch,
420
+ boxes=boxes,
421
+ keypoints=[(0, 0) for _ in range(max(0, int(n_keypoints)))],
422
+ )
423
+ )
424
+ return results
weights.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac86065b9d2b73ee302124f3fdaae9144d09d756ab8290900146e98375bba724
3
+ size 19892467