| | |
| | """ |
| | HuggingFace Jobs script: Collect teacher outputs with metadata tracking. |
| | |
| | Saves for each image: |
| | - Full SAM 3D Body outputs (.npz) |
| | - Metadata: num_humans, image_width, image_height, processing_time |
| | """ |
| | import argparse |
| | import os |
| | from pathlib import Path |
| | import warnings |
| | warnings.filterwarnings('ignore') |
| | import logging |
| | import sys |
| |
|
| | |
| | logging.basicConfig( |
| | level=logging.INFO, |
| | format='[%(asctime)s] %(levelname)s: %(message)s', |
| | datefmt='%Y-%m-%d %H:%M:%S', |
| | stream=sys.stdout, |
| | force=True |
| | ) |
| | logger = logging.getLogger(__name__) |
| |
|
| | |
| | sys.stdout.reconfigure(line_buffering=True) if hasattr(sys.stdout, 'reconfigure') else None |
| |
|
| | import numpy as np |
| | import torch |
| | from datasets import load_dataset, Dataset as HFDataset, Features, Value |
| | from PIL import Image |
| | import cv2 |
| | from typing import List, Dict, Optional |
| | import time |
| | import functools |
| | import json |
| | from collections import defaultdict |
| | import subprocess |
| |
|
| | |
| | import sys |
| | sam_repo = Path(__file__).parent.parent / "sam-3d-body" |
| | if str(sam_repo) not in sys.path: |
| | sys.path.insert(0, str(sam_repo)) |
| | from sam_3d_body import load_sam_3d_body, SAM3DBodyEstimator |
| |
|
| | |
| | os.environ['PYOPENGL_PLATFORM'] = 'osmesa' |
| |
|
| |
|
| | class GazeEstimator: |
| | """Gaze estimation using L2CS-Net""" |
| | |
| | def __init__(self, device='cuda'): |
| | self.device = device |
| | logger.info("Installing L2CS-Net...") |
| | |
| | try: |
| | subprocess.run( |
| | ['pip', 'install', '-q', 'git+https://github.com/edavalosanaya/L2CS-Net.git@main'], |
| | check=True, |
| | capture_output=True |
| | ) |
| | logger.info("✓ L2CS-Net installed") |
| | except Exception as e: |
| | print(f"Warning: L2CS-Net installation failed: {e}") |
| | |
| | print("Loading L2CS-Net gaze estimator...") |
| | try: |
| | from l2cs import Pipeline |
| | |
| | self.pipeline = Pipeline( |
| | weights='L2CSNet_gaze360.pkl', |
| | arch='ResNet50', |
| | device=device |
| | ) |
| | self.enabled = True |
| | print("✓ L2CS-Net gaze estimator loaded") |
| | except Exception as e: |
| | print(f"Warning: Could not load L2CS-Net: {e}") |
| | print("Gaze estimation will be disabled") |
| | self.enabled = False |
| | |
| | def estimate_gaze(self, bbox, detections=None, image_bgr=None): |
| | """ |
| | Estimate gaze direction for a bbox using optional precomputed detections. |
| | |
| | Args: |
| | bbox: [x1, y1, x2, y2] |
| | detections: cached L2CS pipeline outputs for the full image |
| | image_bgr: optional BGR image (used only when detections missing) |
| | """ |
| | if not self.enabled: |
| | return None |
| | |
| | try: |
| | if detections is None and image_bgr is not None: |
| | detections = self.pipeline.step(image_bgr) |
| | if not detections: |
| | return None |
| | |
| | x1, y1, x2, y2 = bbox |
| | bbox_center = np.array([(x1 + x2) / 2, (y1 + y2) / 2]) |
| | best_result = None |
| | min_dist = float('inf') |
| | |
| | for result in detections: |
| | face_bbox = result.get('bbox') |
| | if face_bbox is None: |
| | continue |
| | fx1, fy1, fx2, fy2 = face_bbox |
| | face_center = np.array([(fx1 + fx2) / 2, (fy1 + fy2) / 2]) |
| | dist = np.linalg.norm(bbox_center - face_center) |
| | if dist < min_dist: |
| | min_dist = dist |
| | best_result = result |
| | |
| | if best_result is not None: |
| | pitch = float(best_result.get('pitch', 0)) |
| | yaw = float(best_result.get('yaw', 0)) |
| | return {'pitch': pitch, 'yaw': yaw} |
| | return None |
| | except Exception as e: |
| | print(f"Gaze estimation error: {e}") |
| | return None |
| | |
| | def run_pipeline(self, image_bgr): |
| | """Run L2CS pipeline once per image and reuse detections.""" |
| | if not self.enabled: |
| | return None |
| | try: |
| | return self.pipeline.step(image_bgr) |
| | except Exception as e: |
| | print(f"Warning: L2CS pipeline failed: {e}") |
| | return None |
| |
|
| |
|
| | class FaceEmbedder: |
| | """Face embedding extraction using InsightFace ArcFace (ResNet100-IR)""" |
| | |
| | def __init__(self, device='cuda'): |
| | self.device = device |
| | print("Installing InsightFace...") |
| | |
| | try: |
| | subprocess.run( |
| | ['pip', 'install', '-q', 'insightface', 'onnxruntime-gpu' if device.type == 'cuda' else 'onnxruntime'], |
| | check=True, |
| | capture_output=True |
| | ) |
| | print("✓ InsightFace installed") |
| | except Exception as e: |
| | print(f"Warning: InsightFace installation failed: {e}") |
| | |
| | print("Loading InsightFace ArcFace (ResNet100-IR)...") |
| | try: |
| | import insightface |
| | from insightface.app import FaceAnalysis |
| | |
| | |
| | |
| | self.app = FaceAnalysis( |
| | name='buffalo_l', |
| | providers=['CUDAExecutionProvider'] if device.type == 'cuda' else ['CPUExecutionProvider'] |
| | ) |
| | self.app.prepare(ctx_id=0 if device.type == 'cuda' else -1, det_size=(640, 640)) |
| | self.enabled = True |
| | print("✓ InsightFace ArcFace loaded (ResNet100-IR)") |
| | print(" Model: buffalo_l (ResNet100 + ArcFace head)") |
| | print(" Robust to: occlusion, lighting, blur, pose variations, aging") |
| | except Exception as e: |
| | print(f"Warning: Could not load InsightFace: {e}") |
| | print("Face embeddings will be disabled") |
| | self.enabled = False |
| | |
| | def extract_embedding(self, image, bbox=None, keypoints_2d=None): |
| | """ |
| | Extract 512-dimensional ArcFace embedding from face. |
| | |
| | Args: |
| | image: PIL Image or BGR numpy array |
| | bbox: [x1, y1, x2, y2] in pixel coordinates (optional, for cropping) |
| | keypoints_2d: Face keypoints for alignment (optional) |
| | |
| | Returns: |
| | dict with 'embedding' (512-dim vector), 'det_score' (confidence), or None if failed |
| | """ |
| | if not self.enabled: |
| | return None |
| | |
| | try: |
| | |
| | if isinstance(image, Image.Image): |
| | image_np = np.array(image) |
| | if image_np.shape[2] == 3: |
| | image_bgr = cv2.cvtColor(image_np, cv2.COLOR_RGB2BGR) |
| | else: |
| | image_bgr = image_np |
| | else: |
| | image_bgr = image |
| | |
| | |
| | if bbox is not None: |
| | x1, y1, x2, y2 = map(int, bbox) |
| | |
| | pad = 20 |
| | h, w = image_bgr.shape[:2] |
| | x1 = max(0, x1 - pad) |
| | y1 = max(0, y1 - pad) |
| | x2 = min(w, x2 + pad) |
| | y2 = min(h, y2 + pad) |
| | image_bgr = image_bgr[y1:y2, x1:x2] |
| | |
| | |
| | faces = self.app.get(image_bgr) |
| | |
| | if len(faces) == 0: |
| | return None |
| | |
| | |
| | face = max(faces, key=lambda x: x.det_score) |
| | |
| | |
| | embedding = face.embedding |
| | det_score = float(face.det_score) |
| | |
| | |
| | embedding_norm = embedding / np.linalg.norm(embedding) |
| | |
| | return { |
| | 'embedding': embedding_norm.astype(np.float32).tolist(), |
| | 'det_score': det_score, |
| | 'embedding_dim': len(embedding) |
| | } |
| | |
| | except Exception as e: |
| | print(f"Face embedding extraction error: {e}") |
| | return None |
| |
|
| |
|
| | class NSFWClassifier: |
| | """NSFW classification using EraX-NSFW-V1.0 YOLO model""" |
| | |
| | def __init__(self, device='cuda'): |
| | self.device = device |
| | print("Loading EraX-NSFW YOLO model...") |
| | try: |
| | |
| | from huggingface_hub import snapshot_download |
| | snapshot_download(repo_id="erax-ai/EraX-NSFW-V1.0", local_dir="./", force_download=False) |
| | |
| | from ultralytics import YOLO |
| | |
| | self.model = YOLO('erax_nsfw_yolo11m.pt') |
| | self.enabled = True |
| | print("✓ EraX-NSFW classifier loaded (YOLO11m)") |
| | except Exception as e: |
| | print(f"Warning: Could not load EraX-NSFW: {e}") |
| | print("NSFW classification will be disabled") |
| | self.enabled = False |
| | |
| | def classify_crop(self, image_pil, bbox): |
| | """ |
| | Classify NSFW content in a crop defined by bbox. |
| | |
| | Args: |
| | image_pil: PIL Image |
| | bbox: [x1, y1, x2, y2] in pixel coordinates |
| | |
| | Returns: |
| | dict with class scores, or None if failed |
| | """ |
| | if not self.enabled: |
| | return None |
| | |
| | try: |
| | |
| | x1, y1, x2, y2 = bbox |
| | |
| | |
| | crop = image_pil.crop((x1, y1, x2, y2)) |
| | |
| | |
| | crop_np = np.array(crop) |
| | |
| | |
| | results = self.model(crop_np, conf=0.2, iou=0.3, verbose=False) |
| | |
| | detections = [] |
| | if len(results) > 0 and len(results[0].boxes) > 0: |
| | boxes = results[0].boxes |
| | for box in boxes: |
| | class_id = int(box.cls.item()) |
| | confidence = box.conf.item() |
| | |
| | |
| | class_names = ['anus', 'make_love', 'nipple', 'penis', 'vagina'] |
| | class_name = class_names[class_id] if class_id < len(class_names) else f'class_{class_id}' |
| | |
| | |
| | dx1, dy1, dx2, dy2 = box.xyxy[0].tolist() |
| | abs_bbox = [x1 + dx1, y1 + dy1, x1 + dx2, y1 + dy2] |
| | |
| | detections.append({ |
| | 'class': class_name, |
| | 'confidence': confidence, |
| | 'bbox': abs_bbox |
| | }) |
| | |
| | if detections: |
| | return detections |
| | else: |
| | |
| | return [{'class': 'safe', 'confidence': 1.0, 'bbox': [x1, y1, x2, y2]}] |
| | |
| | except Exception as e: |
| | print(f"! NSFW classification failed: {e}") |
| | return None |
| |
|
| |
|
| | def compute_face_orientation(vertices, keypoints_3d): |
| | """ |
| | Compute face orientation vector from 3D mesh vertices and keypoints. |
| | Uses nose→head_top vector as face direction. |
| | |
| | Args: |
| | vertices: (N, 3) array of 3D vertices |
| | keypoints_3d: (70, 3) array of 3D keypoints (MHR70 format) |
| | |
| | Returns: |
| | (3,) normalized face orientation vector [x, y, z] or None |
| | """ |
| | if vertices is None or keypoints_3d is None: |
| | return None |
| | |
| | try: |
| | |
| | |
| | |
| | nose_3d = keypoints_3d[0] |
| | left_eye_3d = keypoints_3d[1] |
| | right_eye_3d = keypoints_3d[2] |
| | |
| | |
| | if (np.linalg.norm(nose_3d) < 1e-6 or |
| | np.linalg.norm(left_eye_3d) < 1e-6 or |
| | np.linalg.norm(right_eye_3d) < 1e-6): |
| | return None |
| | |
| | |
| | head_top_idx = np.argmax(vertices[:, 1]) |
| | head_top_3d = vertices[head_top_idx] |
| | |
| | |
| | face_orientation = head_top_3d - nose_3d |
| | |
| | |
| | norm = np.linalg.norm(face_orientation) |
| | if norm > 1e-6: |
| | face_orientation = face_orientation / norm |
| | return face_orientation.astype(np.float32) |
| | |
| | return None |
| | |
| | except Exception as e: |
| | print(f"Face orientation computation failed: {e}") |
| | return None |
| |
|
| |
|
| | def compute_bbox_from_keypoints(keypoints_2d, indices): |
| | """ |
| | Compute bounding box from a set of 2D keypoints. |
| | |
| | Args: |
| | keypoints_2d: (70, 2) array of 2D keypoints |
| | indices: list of keypoint indices to include |
| | |
| | Returns: |
| | [x1, y1, x2, y2] or None if no valid keypoints |
| | """ |
| | if keypoints_2d is None or len(keypoints_2d) < max(indices) + 1: |
| | return None |
| | |
| | valid_points = [] |
| | for idx in indices: |
| | kp = keypoints_2d[idx] |
| | if kp[0] >= 0 and kp[1] >= 0: |
| | valid_points.append(kp) |
| | |
| | if len(valid_points) < 2: |
| | return None |
| | |
| | points = np.array(valid_points) |
| | x1, y1 = points.min(axis=0) |
| | x2, y2 = points.max(axis=0) |
| | |
| | |
| | width = x2 - x1 |
| | height = y2 - y1 |
| | padding_x = width * 0.1 |
| | padding_y = height * 0.1 |
| | |
| | x1 = max(0, x1 - padding_x) |
| | y1 = max(0, y1 - padding_y) |
| | x2 = x2 + padding_x |
| | y2 = y2 + padding_y |
| | |
| | return [float(x1), float(y1), float(x2), float(y2)] |
| |
|
| |
|
| | def process_batch(batch, teacher, nsfw_classifier, gaze_estimator, face_embedder, faces, out_dir): |
| | """ |
| | Process a batch of samples using dataset.map() with batched NSFW inference |
| | |
| | Args: |
| | batch: dict with 'image' list and optional 'image_path' list |
| | ... (other args) |
| | |
| | Returns: |
| | dict with 'metadata' list |
| | """ |
| | images = batch['image'] |
| | image_paths = batch.get('image_path', [f'img_{i:06d}' for i in range(len(images))]) |
| | |
| | |
| | humans_data_list = [] |
| | outputs_list = [] |
| | image_rgbs = [] |
| | image_bgrs = [] |
| | gaze_detections = [] |
| | |
| | for img_idx, image_pil in enumerate(images): |
| | img_width, img_height = image_pil.size |
| | image_rgb = np.array(image_pil.convert('RGB')) |
| | image_rgbs.append(image_rgb) |
| | image_bgr = cv2.cvtColor(image_rgb, cv2.COLOR_RGB2BGR) |
| | image_bgrs.append(image_bgr) |
| | |
| | detections = gaze_estimator.run_pipeline(image_bgr) if gaze_estimator is not None else None |
| | gaze_detections.append(detections) |
| | |
| | with torch.inference_mode(): |
| | outputs = teacher.process_one_image(image_bgr) |
| | outputs_list.append(outputs) |
| | |
| | if not outputs: |
| | humans_data_list.append([]) |
| | continue |
| | |
| | humans_data = [] |
| | for human_idx, pred in enumerate(outputs): |
| | vertices = pred.get('pred_vertices') |
| | cam_t = pred.get('pred_cam_t') |
| | focal_length = pred.get('focal_length') |
| | kpts2d = pred.get('pred_keypoints_2d') |
| | kpts3d = pred.get('pred_keypoints_3d') |
| | bbox = pred.get('bbox', None) |
| | |
| | |
| | has_face = False |
| | if kpts2d is not None and kpts3d is not None and len(kpts2d) >= 3 and len(kpts3d) >= 3: |
| | nose_2d = kpts2d[0] |
| | left_eye_2d = kpts2d[1] |
| | right_eye_2d = kpts2d[2] |
| | nose_3d = kpts3d[0] |
| | left_eye_3d = kpts3d[1] |
| | right_eye_3d = kpts3d[2] |
| | |
| | keypoints_valid_3d = (np.linalg.norm(nose_3d) > 1e-6 and |
| | np.linalg.norm(left_eye_3d) > 1e-6 and |
| | np.linalg.norm(right_eye_3d) > 1e-6) |
| | |
| | keypoints_in_image = True |
| | if keypoints_valid_3d: |
| | for kp in [nose_2d, left_eye_2d, right_eye_2d]: |
| | if (kp[0] < 0 or kp[0] >= img_width or |
| | kp[1] < 0 or kp[1] >= img_height): |
| | keypoints_in_image = False |
| | break |
| | |
| | has_face = keypoints_valid_3d and keypoints_in_image |
| | |
| | face_orientation = None |
| | if has_face: |
| | face_orientation = compute_face_orientation(vertices, kpts3d) |
| | |
| | gaze_direction = None |
| | if has_face and bbox is not None and gaze_estimator is not None: |
| | try: |
| | gaze_direction = gaze_estimator.estimate_gaze(bbox, detections=gaze_detections[img_idx]) |
| | except Exception as e: |
| | gaze_direction = None |
| | |
| | face_embedding = None |
| | if has_face and bbox is not None and face_embedder is not None: |
| | try: |
| | face_embedding = face_embedder.extract_embedding(image_bgrs[img_idx], bbox, kpts2d) |
| | except Exception as e: |
| | face_embedding = None |
| | |
| | |
| | left_hand_bbox = None |
| | right_hand_bbox = None |
| | left_foot_bbox = None |
| | right_foot_bbox = None |
| | |
| | if kpts2d is not None: |
| | left_hand_indices = list(range(42, 62)) |
| | left_hand_bbox = compute_bbox_from_keypoints(kpts2d, left_hand_indices) |
| | |
| | right_hand_indices = list(range(21, 41)) |
| | right_hand_bbox = compute_bbox_from_keypoints(kpts2d, right_hand_indices) |
| | |
| | left_foot_indices = [15, 16, 17] |
| | left_foot_bbox = compute_bbox_from_keypoints(kpts2d, left_foot_indices) |
| | |
| | right_foot_indices = [18, 19, 20] |
| | right_foot_bbox = compute_bbox_from_keypoints(kpts2d, right_foot_indices) |
| | |
| | humans_data.append({ |
| | 'human_idx': human_idx, |
| | 'bbox': bbox.tolist() if bbox is not None else None, |
| | 'left_hand_bbox': left_hand_bbox, |
| | 'right_hand_bbox': right_hand_bbox, |
| | 'left_foot_bbox': left_foot_bbox, |
| | 'right_foot_bbox': right_foot_bbox, |
| | 'has_face': has_face, |
| | 'face_orientation': face_orientation.tolist() if face_orientation is not None else None, |
| | 'gaze_direction': gaze_direction, |
| | 'face_embedding': face_embedding, |
| | 'has_mesh': vertices is not None, |
| | 'nsfw_scores': None |
| | }) |
| | |
| | humans_data_list.append(humans_data) |
| | |
| | |
| | crops = [] |
| | crop_info = [] |
| | |
| | for img_idx, humans_data in enumerate(humans_data_list): |
| | image_pil = images[img_idx] |
| | for human_idx, human in enumerate(humans_data): |
| | bbox = human['bbox'] |
| | if bbox is not None: |
| | x1, y1, x2, y2 = bbox |
| | ix1, iy1, ix2, iy2 = map(lambda v: max(0, int(round(v))), [x1, y1, x2, y2]) |
| | ix1, iy1 = min(ix1, image_rgbs[img_idx].shape[1]-1), min(iy1, image_rgbs[img_idx].shape[0]-1) |
| | ix2, iy2 = max(ix1+1, min(ix2, image_rgbs[img_idx].shape[1])), max(iy1+1, min(iy2, image_rgbs[img_idx].shape[0])) |
| | crop_np = np.ascontiguousarray(image_rgbs[img_idx][iy1:iy2, ix1:ix2]) |
| | crops.append(crop_np) |
| | crop_info.append((img_idx, human_idx)) |
| | |
| | if crops and nsfw_classifier is not None and nsfw_classifier.enabled: |
| | try: |
| | results = nsfw_classifier.model(crops, conf=0.2, iou=0.3, verbose=False) |
| | |
| | for crop_idx, result in enumerate(results): |
| | img_idx, human_idx = crop_info[crop_idx] |
| | bbox = humans_data_list[img_idx][human_idx]['bbox'] |
| | x1, y1, x2, y2 = bbox |
| | |
| | detections = [] |
| | if result.boxes: |
| | for box in result.boxes: |
| | class_id = int(box.cls.item()) |
| | confidence = box.conf.item() |
| | class_names = ['anus', 'make_love', 'nipple', 'penis', 'vagina'] |
| | class_name = class_names[class_id] if class_id < len(class_names) else f'class_{class_id}' |
| | |
| | dx1, dy1, dx2, dy2 = box.xyxy[0].tolist() |
| | abs_bbox = [x1 + dx1, y1 + dy1, x1 + dx2, y1 + dy2] |
| | |
| | detections.append({ |
| | 'class': class_name, |
| | 'confidence': confidence, |
| | 'bbox': abs_bbox |
| | }) |
| | |
| | if detections: |
| | humans_data_list[img_idx][human_idx]['nsfw_scores'] = detections |
| | else: |
| | humans_data_list[img_idx][human_idx]['nsfw_scores'] = [{'class': 'safe', 'confidence': 1.0, 'bbox': [x1, y1, x2, y2]}] |
| | except Exception as e: |
| | print(f"! Batched NSFW failed: {e}") |
| | |
| | for img_idx, humans_data in enumerate(humans_data_list): |
| | for human in humans_data: |
| | if human['bbox'] is not None: |
| | x1, y1, x2, y2 = human['bbox'] |
| | human['nsfw_scores'] = [{'class': 'safe', 'confidence': 1.0, 'bbox': [x1, y1, x2, y2]}] |
| | |
| | |
| | metadatas = [] |
| | for img_idx, (humans_data, image_path) in enumerate(zip(humans_data_list, image_paths)): |
| | image_pil = images[img_idx] |
| | image_id = Path(image_path).stem if image_path else f'img_{img_idx:06d}' |
| | img_width, img_height = image_pil.size |
| | |
| | out_path = out_dir / f"{image_id}.npz" |
| | if out_path.exists(): |
| | metadatas.append(None) |
| | continue |
| | |
| | if not humans_data: |
| | metadata = { |
| | 'image_id': image_id, |
| | 'num_humans': 0, |
| | 'image_width': img_width, |
| | 'image_height': img_height, |
| | 'processing_time_ms': 0, |
| | 'status': 'no_detection', |
| | 'humans': [] |
| | } |
| | else: |
| | num_humans = len(humans_data) |
| | |
| | |
| | pred = outputs_list[img_idx][0] |
| | vertices = pred.get('pred_vertices') |
| | cam_t = pred.get('pred_cam_t') |
| | focal_length = pred.get('focal_length') |
| | kpts2d = pred.get('pred_keypoints_2d') |
| | kpts3d = pred.get('pred_keypoints_3d') |
| | bbox_0 = pred.get('bbox', None) |
| | |
| | np.savez_compressed( |
| | out_path, |
| | vertices=vertices.astype(np.float32) if vertices is not None else None, |
| | faces=faces.astype(np.int32), |
| | cam_t=cam_t.astype(np.float32) if cam_t is not None else None, |
| | focal_length=np.array([focal_length], dtype=np.float32) if focal_length is not None else None, |
| | keypoints_2d=kpts2d.astype(np.float32) if kpts2d is not None else None, |
| | keypoints_3d=kpts3d.astype(np.float32) if kpts3d is not None else None, |
| | bbox=np.array(bbox_0, dtype=np.float32) if bbox_0 is not None else None, |
| | image_id=image_id, |
| | num_humans=num_humans, |
| | image_width=img_width, |
| | image_height=img_height, |
| | humans_metadata=json.dumps(humans_data) |
| | ) |
| | |
| | metadata = { |
| | 'image_id': image_id, |
| | 'num_humans': num_humans, |
| | 'image_width': img_width, |
| | 'image_height': img_height, |
| | 'processing_time_ms': 0, |
| | 'status': 'success', |
| | 'npz_size_bytes': out_path.stat().st_size, |
| | 'humans': humans_data |
| | } |
| | |
| | metadatas.append(metadata) |
| | |
| | return {'metadata': metadatas} |
| |
|
| |
|
| | def main(): |
| | logger.info("="*60) |
| | logger.info("SAM 3D Body Metadata Collection with Face Features") |
| | logger.info("="*60) |
| | sys.stdout.flush() |
| | |
| | ap = argparse.ArgumentParser() |
| | ap.add_argument('--input-dataset', type=str, required=True) |
| | ap.add_argument('--output-dataset', type=str, required=True) |
| | ap.add_argument('--split', type=str, default='train') |
| | ap.add_argument('--checkpoint', type=str, default='checkpoints/sam-3d-body-dinov3/model.ckpt') |
| | ap.add_argument('--mhr-path', type=str, default='checkpoints/sam-3d-body-dinov3/assets/mhr_model.pt') |
| | ap.add_argument('--limit', type=int, default=0) |
| | ap.add_argument('--shard-index', type=int, default=0) |
| | ap.add_argument('--num-shards', type=int, default=1) |
| | args = ap.parse_args() |
| |
|
| | logger.info(f"Arguments: {vars(args)}") |
| | sys.stdout.flush() |
| |
|
| | device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') |
| | logger.info(f"Using device: {device}") |
| | if torch.cuda.is_available(): |
| | logger.info(f" GPU: {torch.cuda.get_device_name(0)}") |
| | logger.info(f" CUDA version: {torch.version.cuda}") |
| | sys.stdout.flush() |
| |
|
| | |
| | logger.info("Loading gaze estimator...") |
| | sys.stdout.flush() |
| | gaze_estimator = GazeEstimator(device=device) |
| | |
| | |
| | logger.info("Loading face embedder (InsightFace ArcFace)...") |
| | sys.stdout.flush() |
| | face_embedder = FaceEmbedder(device=device) |
| | |
| | |
| | logger.info("Loading NSFW classifier...") |
| | sys.stdout.flush() |
| | nsfw_classifier = NSFWClassifier(device=device) |
| |
|
| | |
| | logger.info("Loading SAM 3D Body teacher...") |
| | sys.stdout.flush() |
| | start_load = time.time() |
| | model, model_cfg = load_sam_3d_body(args.checkpoint, device=device, mhr_path=args.mhr_path) |
| | model.eval() |
| | |
| | teacher = SAM3DBodyEstimator( |
| | sam_3d_body_model=model, |
| | model_cfg=model_cfg, |
| | human_detector=None, |
| | human_segmentor=None, |
| | fov_estimator=None, |
| | ) |
| | logger.info(f"✓ Model loaded in {time.time() - start_load:.1f}s") |
| | sys.stdout.flush() |
| |
|
| | |
| | logger.info(f"Loading dataset {args.input_dataset}...") |
| | sys.stdout.flush() |
| | start_ds = time.time() |
| | ds = load_dataset(args.input_dataset, split=args.split, streaming=True) |
| | |
| | if args.num_shards > 1: |
| | ds = ds.shard(num_shards=args.num_shards, index=args.shard_index) |
| | logger.info(f"Using shard {args.shard_index+1}/{args.num_shards} (~{100/args.num_shards:.1f}% of dataset)") |
| | |
| | if args.limit and args.limit > 0: |
| | ds = ds.take(args.limit) |
| | logger.info(f"✓ Dataset ready in {time.time() - start_ds:.1f}s") |
| | sys.stdout.flush() |
| |
|
| | |
| | out_dir = Path('teacher_labels') |
| | out_dir.mkdir(exist_ok=True) |
| | faces = teacher.faces |
| | logger.info(f"Mesh topology: {faces.shape[0]} faces") |
| | sys.stdout.flush() |
| |
|
| | |
| | batch_size = 4 |
| | logger.info(f"Processing with batch_size={batch_size} using dataset.map()") |
| | sys.stdout.flush() |
| | |
| | process_batch_partial = functools.partial( |
| | process_batch, |
| | teacher=teacher, |
| | nsfw_classifier=nsfw_classifier, |
| | gaze_estimator=gaze_estimator, |
| | face_embedder=face_embedder, |
| | faces=faces, |
| | out_dir=out_dir |
| | ) |
| | |
| | processed_ds = ds.map( |
| | process_batch_partial, |
| | batched=True, |
| | batch_size=batch_size, |
| | remove_columns=ds.column_names |
| | ) |
| | |
| | |
| | metadata_records = [] |
| | batch_count = 0 |
| | start_process = time.time() |
| | |
| | for batch_result in processed_ds: |
| | metadata_records.extend(batch_result['metadata']) |
| | batch_count += 1 |
| | |
| | if batch_count % 10 == 0: |
| | elapsed = time.time() - start_process |
| | processed = sum(1 for m in metadata_records if m and m['status'] == 'success') |
| | speed = processed / elapsed if elapsed > 0 else 0 |
| | logger.info(f"[{batch_count} batches] success={processed}, speed={speed:.2f} img/s") |
| | sys.stdout.flush() |
| | |
| | total_time = time.time() - start_process |
| | processed = sum(1 for m in metadata_records if m and m['status'] == 'success') |
| | no_detection = sum(1 for m in metadata_records if m and m['status'] == 'no_detection') |
| | failed = sum(1 for m in metadata_records if m and m['status'] == 'error') |
| | |
| | logger.info("="*60) |
| | logger.info(f"✓ Processing complete!") |
| | logger.info(f" Processed: {processed} images in {total_time:.1f}s ({processed/total_time:.2f} img/s)") |
| | logger.info(f" No detection: {no_detection}, Failed: {failed}") |
| | logger.info("="*60) |
| | sys.stdout.flush() |
| | |
| | |
| | if metadata_records: |
| | successful = [m for m in metadata_records if m['status'] == 'success'] |
| | if successful: |
| | total_humans = sum(m['num_humans'] for m in successful) |
| | avg_humans = total_humans / len(successful) |
| | avg_width = sum(m['image_width'] for m in successful) / len(successful) |
| | avg_height = sum(m['image_height'] for m in successful) / len(successful) |
| | avg_time = sum(m['processing_time_ms'] for m in successful) / len(successful) |
| | |
| | |
| | nsfw_stats = defaultdict(list) |
| | for m in successful: |
| | for human in m.get('humans', []): |
| | nsfw_list = human.get('nsfw_scores', []) |
| | for detection in nsfw_list: |
| | label = detection['class'] |
| | score = detection['confidence'] |
| | nsfw_stats[label].append(score) |
| | |
| | print(f"\nMetadata Statistics:") |
| | print(f" Total humans detected: {total_humans}") |
| | print(f" Avg humans per image: {avg_humans:.2f}") |
| | print(f" Avg image size: {avg_width:.0f}x{avg_height:.0f}") |
| | print(f" Avg processing time: {avg_time:.0f}ms") |
| | |
| | if nsfw_stats: |
| | print(f"\nNSFW Classification Statistics:") |
| | for label, scores in nsfw_stats.items(): |
| | avg_score = sum(scores) / len(scores) |
| | max_score = max(scores) |
| | print(f" {label}: avg={avg_score:.3f}, max={max_score:.3f}, n={len(scores)}") |
| | |
| | |
| | face_orientation_count = sum(1 for m in successful for h in m.get('humans', []) if h.get('face_orientation')) |
| | gaze_count = sum(1 for m in successful for h in m.get('humans', []) if h.get('gaze_direction')) |
| | face_embedding_count = sum(1 for m in successful for h in m.get('humans', []) if h.get('face_embedding')) |
| | print(f"\nFace Orientation & Gaze Statistics:") |
| | print(f" Face orientations computed: {face_orientation_count}/{total_humans}") |
| | print(f" Gaze directions estimated: {gaze_count}/{total_humans}") |
| | print(f" Face embeddings extracted: {face_embedding_count}/{total_humans}") |
| | |
| | |
| | if face_embedding_count > 0: |
| | det_scores = [h['face_embedding']['det_score'] for m in successful |
| | for h in m.get('humans', []) if h.get('face_embedding')] |
| | avg_det_score = sum(det_scores) / len(det_scores) |
| | min_det_score = min(det_scores) |
| | print(f"\nFace Embedding Quality (InsightFace ArcFace):") |
| | print(f" Model: ResNet100-IR + ArcFace head (512-dim)") |
| | print(f" Avg detection confidence: {avg_det_score:.3f}") |
| | print(f" Min detection confidence: {min_det_score:.3f}") |
| |
|
| | |
| | |
| | metadata_path = Path('metadata.json') |
| | with open(metadata_path, 'w') as f: |
| | json.dump(metadata_records, f, indent=2) |
| | print(f"Saved metadata to {metadata_path}") |
| | |
| | |
| | print(f"\nUploading labels to {args.output_dataset}...") |
| | |
| | label_files = sorted(out_dir.glob('*.npz')) |
| | data = {'image_id': [], 'label_data': []} |
| | |
| | for npz_path in label_files: |
| | data['image_id'].append(npz_path.stem) |
| | with open(npz_path, 'rb') as f: |
| | data['label_data'].append(f.read()) |
| | |
| | features = Features({ |
| | 'image_id': Value('string'), |
| | 'label_data': Value('binary'), |
| | }) |
| | |
| | label_ds = HFDataset.from_dict(data, features=features) |
| | label_ds.push_to_hub( |
| | args.output_dataset, |
| | split=args.split, |
| | token=os.environ.get('HF_TOKEN'), |
| | private=True, |
| | ) |
| | logger.info(f"✓ Uploaded {len(label_files)} labels to {args.output_dataset}") |
| | sys.stdout.flush() |
| | |
| | |
| | from huggingface_hub import HfApi |
| | api = HfApi(token=os.environ.get('HF_TOKEN')) |
| | api.upload_file( |
| | path_or_fileobj=str(metadata_path), |
| | path_in_repo=f'metadata_shard{args.shard_index}.json', |
| | repo_id=args.output_dataset, |
| | repo_type='dataset' |
| | ) |
| | logger.info(f"✓ Uploaded metadata to {args.output_dataset}/metadata_shard{args.shard_index}.json") |
| | sys.stdout.flush() |
| |
|
| |
|
| | if __name__ == '__main__': |
| | main() |
| |
|