{ "manifest_version": 1, "updated_at": "2026-04-10", "min_app_version": "1.0", "categories": [ { "id": "llm", "name": "Large Language Models", "icon": "bubble.left.and.text.bubble.right", "order": 0 }, { "id": "segmentation", "name": "Segmentation", "icon": "person.and.background.dotted", "order": 1 }, { "id": "enhancement", "name": "Image Enhancement", "icon": "wand.and.stars", "order": 2 }, { "id": "detection", "name": "Object Detection", "icon": "viewfinder", "order": 3 }, { "id": "depth", "name": "Depth & Geometry", "icon": "cube.transparent", "order": 4 }, { "id": "vision_language", "name": "Vision-Language", "icon": "text.viewfinder", "order": 5 }, { "id": "face", "name": "Face Processing", "icon": "face.smiling", "order": 6 }, { "id": "generation", "name": "Image Generation", "icon": "sparkles", "order": 7 }, { "id": "video", "name": "Video Processing", "icon": "film", "order": 8 }, { "id": "audio", "name": "Audio Processing", "icon": "waveform.circle", "order": 9 }, { "id": "speech", "name": "Speech & Music", "icon": "music.note", "order": 10 } ], "models": [ { "id": "gemma4_e2b", "name": "Gemma 4 E2B", "subtitle": "Google DeepMind, 2025", "category_id": "llm", "description_md": "Google's latest on-device multimodal LLM. 2.3B effective parameters with Per-Layer Embeddings. Text + image input, streaming text output. Runs on Apple Neural Engine at ~31 tok/s decode. Supports multi-turn conversations, image understanding, and reasoning.", "thumbnail_url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/thumbnails/gemma4.jpg", "demo": { "template": "chat", "config": { "max_tokens": 1024, "multimodal": true } }, "files": [ { "name": "gemma4-e2b-coreml.zip", "url": "https://huggingface.co/mlboydaisuke/gemma-4-E2B-coreml/resolve/main/gemma4-e2b-coreml.zip", "archive": "zip", "size_bytes": 2700000000, "sha256": "TODO", "compute_units": "cpuAndNeuralEngine", "kind": "model" } ], "requirements": { "min_ios": "18.0", "min_ram_mb": 1500, "device_capabilities": [ "arm64" ] }, "license": { "name": "Gemma", "url": "https://ai.google.dev/gemma/terms" }, "upstream": { "name": "google/gemma-4-e2b", "url": "https://huggingface.co/google/gemma-4-e2b", "year": 2025 } }, { "id": "rmbg_1_4", "name": "RMBG-1.4", "subtitle": "BRIA AI, 2023", "category_id": "segmentation", "description_md": "High-quality background removal. Outputs foreground with alpha mask. 1024×1024 input.", "demo": { "template": "image_in_out", "config": { "input_size": 1024, "output_type": "mask" } }, "files": [ { "name": "RMBG_1_4.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/rmbg/RMBG_1_4.mlpackage.zip", "archive": "zip", "size_bytes": 38771210, "sha256": "a80dbb5f04c922a8fa698c38592e4e52af4e62471d70bc7c59c28a3355a1da95", "compute_units": "cpuOnly", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 300 }, "license": { "name": "Apache-2.0", "url": "https://huggingface.co/briaai/RMBG-1.4" }, "upstream": { "name": "briaai/RMBG-1.4", "url": "https://huggingface.co/briaai/RMBG-1.4", "year": 2023 } }, { "id": "ddcolor", "name": "DDColor Tiny", "subtitle": "Image Colorization, 2023", "category_id": "enhancement", "description_md": "Automatic grayscale image colorization via dual decoders. 512×512 input.", "demo": { "template": "image_in_out", "config": { "input_size": 512, "output_type": "lab_ab" } }, "files": [ { "name": "DDColor_Tiny.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/ddcolor/DDColor_Tiny.mlpackage.zip", "archive": "zip", "size_bytes": 212344570, "sha256": "bfecea37d66005f602efe13978360b8e4707923234c3d1d00beeb4e36cb1b02c", "compute_units": "all", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 400 }, "license": { "name": "Apache-2.0", "url": "https://github.com/piddnad/DDColor" }, "upstream": { "name": "piddnad/DDColor", "url": "https://github.com/piddnad/DDColor", "year": 2023 } }, { "id": "sinsr", "name": "SinSR", "subtitle": "Single-Step Super-Resolution, 2024", "category_id": "enhancement", "description_md": "4× super-resolution via single-step diffusion. 256→1024. Swin Transformer denoiser (FP32).", "demo": { "template": "image_in_out", "config": { "input_size": 256, "output_type": "sinsr" } }, "files": [ { "name": "SinSR_Encoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/sinsr/SinSR_Encoder.mlpackage.zip", "archive": "zip", "size_bytes": 41246338, "sha256": "fdec09d17561ec1bb5a2e829683d48c2b45e76b876285619a6e29a3523b8b7e2", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "SinSR_Denoiser.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/sinsr/SinSR_Denoiser.mlpackage.zip", "archive": "zip", "size_bytes": 440014511, "sha256": "b31374c2d539b2cdd81499d6062c801ca00e405f5a67507cd609d14e2d6d4beb", "compute_units": "cpuOnly", "kind": "model" }, { "name": "SinSR_Decoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/sinsr/SinSR_Decoder.mlpackage.zip", "archive": "zip", "size_bytes": 60880285, "sha256": "b8b9a7b52d6b240cf9fb3352b286ea83eb984fd73f5dd81c9f034f0016a5cb8c", "compute_units": "cpuAndGPU", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 600 }, "license": { "name": "Apache-2.0", "url": "https://github.com/wyf0912/SinSR" }, "upstream": { "name": "wyf0912/SinSR", "url": "https://github.com/wyf0912/SinSR", "year": 2024 } }, { "id": "yolo26s", "name": "YOLO26s", "subtitle": "NMS-Free Detection, 2026", "category_id": "detection", "description_md": "NMS-free object detection. 640×640 input, 80 COCO classes.", "demo": { "template": "image_detection", "config": { "input_size": 640, "confidence_threshold": 0.25 } }, "files": [ { "name": "yolo26s.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/yolo26/yolo26s.mlpackage.zip", "archive": "zip", "size_bytes": 17697581, "sha256": "0ec02fb0cf2dbd6e09601cbbc00a9734156ea4c2a52b0da23a984337074c6fd4", "compute_units": "all", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 300 }, "license": { "name": "AGPL-3.0", "url": "https://github.com/ultralytics/ultralytics" }, "upstream": { "name": "ultralytics/ultralytics", "url": "https://github.com/ultralytics/ultralytics", "year": 2026 } }, { "id": "yolo11s", "name": "YOLO11s", "subtitle": "Object Detection, 2024", "category_id": "detection", "description_md": "YOLO11 small detection with Vision framework NMS. 640×640 input.", "demo": { "template": "image_detection", "config": { "input_size": 640, "confidence_threshold": 0.25 } }, "files": [ { "name": "yolo11s.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/yolov9/yolo11s.mlpackage.zip", "archive": "zip", "size_bytes": 17580204, "sha256": "79e82aacc3ad20fc1eb990df6979fae9b927d4b06f33bd20ec0e1c0dcb7d1f6b", "compute_units": "all", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 300 }, "license": { "name": "AGPL-3.0", "url": "https://github.com/ultralytics/ultralytics" }, "upstream": { "name": "ultralytics/ultralytics", "url": "https://github.com/ultralytics/ultralytics", "year": 2024 } }, { "id": "yolov10n", "name": "YOLOv10n", "subtitle": "Object Detection, 2024", "category_id": "detection", "description_md": "YOLOv10 nano. 640×640 input. Dual-assignment strategy.", "demo": { "template": "image_detection", "config": { "input_size": 640, "confidence_threshold": 0.25 } }, "files": [ { "name": "YOLOv10N.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/yolov10/YOLOv10N.mlpackage.zip", "archive": "zip", "size_bytes": 4309168, "sha256": "9a687144a6b0b764f508c8f544fe46b6674629b8f09a1e99d8ca69b0be899891", "compute_units": "all", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 300 }, "license": { "name": "AGPL-3.0", "url": "https://github.com/THU-MIG/yolov10" }, "upstream": { "name": "THU-MIG/yolov10", "url": "https://github.com/THU-MIG/yolov10", "year": 2024 } }, { "id": "yoloworld", "name": "YOLO-World", "subtitle": "Open-Vocabulary Detection, 2024", "category_id": "detection", "description_md": "Open-vocabulary detection. Type any text query. YOLO-World V2-S + CLIP ViT-B/32.", "demo": { "template": "open_vocab_detection", "config": { "input_size": 640 } }, "files": [ { "name": "yoloworld_detector.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/yoloworld/yoloworld_detector.mlpackage.zip", "archive": "zip", "size_bytes": 23710620, "sha256": "611d299ae74c83f90a5cc9f4585709859d5db735baa8ade721e0c2d99cd5af92", "compute_units": "all", "kind": "model" }, { "name": "clip_text_encoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/yoloworld/clip_text_encoder.mlpackage.zip", "archive": "zip", "size_bytes": 116681932, "sha256": "45770a743297e8c2a57cc330d4f5c80f47734263680895b33b593b50dd2c382b", "compute_units": "cpuOnly", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 600 }, "license": { "name": "GPL-3.0", "url": "https://github.com/AILab-CVC/YOLO-World" }, "upstream": { "name": "AILab-CVC/YOLO-World", "url": "https://github.com/AILab-CVC/YOLO-World", "year": 2024 } }, { "id": "moge2_vitb_normal_504", "name": "MoGe-2 ViT-B (504×504)", "subtitle": "Microsoft, CVPR 2025", "category_id": "depth", "description_md": "Monocular geometry from a single image. Metric depth, surface normals, confidence mask. DINOv2 ViT-B/14 backbone.", "demo": { "template": "depth_visualization", "config": { "input_size": 504, "output_keys": [ "depth", "normal", "mask", "metric_scale" ], "depth_unit": "meters" } }, "files": [ { "name": "MoGe2_ViTB_Normal_504.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/moge2/MoGe2_ViTB_Normal_504.mlpackage.zip", "archive": "zip", "size_bytes": 193312088, "sha256": "f60cfb4804707a489d99e24453188cd31ddcabb299bbf6da4507edc9cecbf9e7", "compute_units": "all", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 600 }, "license": { "name": "MIT", "url": "https://github.com/microsoft/MoGe/blob/main/LICENSE" }, "upstream": { "name": "microsoft/MoGe", "url": "https://github.com/microsoft/MoGe", "year": 2025 } }, { "id": "siglip", "name": "SigLIP", "subtitle": "Zero-Shot Classification, 2023", "category_id": "vision_language", "description_md": "Zero-shot image classification. Dual encoder (image + text). 224×224 input.", "demo": { "template": "zero_shot_classify", "config": { "input_size": 224, "image_encoder": "SigLIP_ImageEncoder.mlpackage.zip", "text_encoder": "SigLIP_TextEncoder.mlpackage.zip", "vocab_file": "siglip_vocab.json", "prompt_template": "{}", "logit_scale": 117.33 } }, "files": [ { "name": "SigLIP_ImageEncoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/siglip/SigLIP_ImageEncoder.mlpackage.zip", "archive": "zip", "size_bytes": 170352400, "sha256": "98f6abf5f4aa145199f4ae22305f9c1d5929eee6b126daad84783b2b2090ee24", "compute_units": "cpuOnly", "kind": "model" }, { "name": "SigLIP_TextEncoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/siglip/SigLIP_TextEncoder.mlpackage.zip", "archive": "zip", "size_bytes": 203975769, "sha256": "9dead2d58705838aef7ad83c3bf4036698c78d872ca1cdd04f2c4a6272009ccf", "compute_units": "cpuOnly", "kind": "model" }, { "name": "siglip_vocab.json", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/siglip/siglip_vocab.json", "size_bytes": 673754, "sha256": "b94b3a58e04f619936b3890804dff7c478522c07515ff748cf127c5443ee5229", "kind": "vocab" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 800 }, "license": { "name": "Apache-2.0", "url": "https://github.com/google-research/big_vision" }, "upstream": { "name": "google-research/big_vision", "url": "https://github.com/google-research/big_vision", "year": 2023 } }, { "id": "florence2", "name": "Florence-2", "subtitle": "Microsoft, 2024", "category_id": "vision_language", "description_md": "Vision-language captioning, OCR, and VQA. Three-stage encoder-decoder. 768×768 input.", "demo": { "template": "image_to_text", "config": { "image_size": 768, "max_tokens": 256, "vision_encoder": "Florence2VisionEncoder.mlpackage.zip", "text_encoder": "Florence2TextEncoder.mlpackage.zip", "decoder": "Florence2Decoder.mlpackage.zip", "vocab_file": "florence2_vocab.json", "tasks": { "caption": [ 0, 2264, 473, 5, 2274, 6190, 116, 2 ], "detailed_caption": [ 0, 2264, 473, 5, 31962, 2274, 6190, 116, 2 ], "ocr": [ 0, 2264, 473, 5, 71307, 116, 2 ] } } }, "files": [ { "name": "Florence2VisionEncoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/florence2/Florence2VisionEncoder.mlpackage.zip", "archive": "zip", "size_bytes": 81198683, "sha256": "9422f189c21220a0f9966eb9d780856772feb55597dcc579fc4e3c88990d0046", "compute_units": "cpuOnly", "kind": "model" }, { "name": "Florence2TextEncoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/florence2/Florence2TextEncoder.mlpackage.zip", "archive": "zip", "size_bytes": 72742890, "sha256": "f985deeef0408ea8aac33ac4f5c6d9635cd9c64c98b53f85031db6e27f3bfd92", "compute_units": "cpuOnly", "kind": "model" }, { "name": "Florence2Decoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/florence2/Florence2Decoder.mlpackage.zip", "archive": "zip", "size_bytes": 85329746, "sha256": "fe85a6faab5281272bcd79dabfbf87d60ba1a78dd9455e2bf71c67a134d61dc5", "compute_units": "cpuOnly", "kind": "model" }, { "name": "florence2_vocab.json", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/florence2/florence2_vocab.json", "size_bytes": 999352, "sha256": "861fee9af5520403f6dbb4940d6af6627f1481b71cdc4a870f1f61344e57e645", "kind": "vocab" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 1200 }, "license": { "name": "MIT", "url": "https://huggingface.co/microsoft/Florence-2-base" }, "upstream": { "name": "microsoft/Florence-2", "url": "https://huggingface.co/microsoft/Florence-2-base", "year": 2024 } }, { "id": "face3d", "name": "3DDFA V2", "subtitle": "3D Face Reconstruction, 2020", "category_id": "face", "description_md": "Single-image 3D face reconstruction. Predicts 6 DoF pose + expression parameters.", "demo": { "template": "face_3d", "config": { "input_size": 120 } }, "files": [ { "name": "3DDFA_V2.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/face3d/3DDFA_V2.mlpackage.zip", "archive": "zip", "size_bytes": 6083375, "sha256": "0f715dc220c046f558e3b8fc65246df9a2eec77182830a16628783430cdacdc8", "compute_units": "all", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 200 }, "license": { "name": "MIT", "url": "https://github.com/cleardusk/3DDFA_V2" }, "upstream": { "name": "cleardusk/3DDFA_V2", "url": "https://github.com/cleardusk/3DDFA_V2", "year": 2020 } }, { "id": "hypersd", "name": "Hyper-SD (1-Step)", "subtitle": "ByteDance, 2024", "category_id": "generation", "description_md": "Single-step text-to-image from SD1.5 via TCD distillation. 512×512. Chunked UNet (6-bit).", "demo": { "template": "text_to_image", "config": { "image_size": 512, "latent_size": 64, "latent_channels": 4, "steps": 1, "guidance_scale": 1.0, "text_encoder": "HyperSDTextEncoder.mlpackage.zip", "unet_chunk1": "HyperSDUnetChunk1.mlpackage.zip", "unet_chunk2": "HyperSDUnetChunk2.mlpackage.zip", "vae_decoder": "HyperSDVAEDecoder.mlpackage.zip", "vocab_file": "vocab.json", "merges_file": "merges.txt" } }, "files": [ { "name": "HyperSDTextEncoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/hypersd/HyperSDTextEncoder.mlpackage.zip", "archive": "zip", "size_bytes": 226397794, "sha256": "201b0fcc3573811aac6a4e8545c695bc4fb2f7710ea0d60c227919d87b37687e", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "HyperSDUnetChunk1.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/hypersd/HyperSDUnetChunk1.mlpackage.zip", "archive": "zip", "size_bytes": 324819653, "sha256": "279da11b8231aeeb9045f6ceabebb3a68c20a1b86ecc81aa6914b77ce76d5203", "compute_units": "cpuAndNeuralEngine", "kind": "model" }, { "name": "HyperSDUnetChunk2.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/hypersd/HyperSDUnetChunk2.mlpackage.zip", "archive": "zip", "size_bytes": 304530429, "sha256": "0a700d11a105da589bb3e5666e38b9c72fa283149951b253fc11722e70e72faa", "compute_units": "cpuAndNeuralEngine", "kind": "model" }, { "name": "HyperSDVAEDecoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/hypersd/HyperSDVAEDecoder.mlpackage.zip", "archive": "zip", "size_bytes": 91282754, "sha256": "1260371542d845a2261ed2de36c5fe3e9ccb740a6ceb59b1990705d125e8cf66", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "vocab.json", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/hypersd/vocab.json", "size_bytes": 1059962, "sha256": "e089ad92ba36837a0d31433e555c8f45fe601ab5c221d4f607ded32d9f7a4349", "kind": "vocab" }, { "name": "merges.txt", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/hypersd/merges.txt", "size_bytes": 524619, "sha256": "9fd691f7c8039210e0fced15865466c65820d09b63988b0174bfe25de299051a", "kind": "vocab" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 1000 }, "license": { "name": "OpenRAIL-M", "url": "https://huggingface.co/ByteDance/Hyper-SD" }, "upstream": { "name": "ByteDance/Hyper-SD", "url": "https://huggingface.co/ByteDance/Hyper-SD", "year": 2024 } }, { "id": "matanyone", "name": "MatAnyone", "subtitle": "Video Matting, 2025", "category_id": "video", "description_md": "Temporally consistent video matting. 5-model pipeline with memory propagation.", "demo": { "template": "video_matting", "config": { "frame_size": 512, "encoder": "MatAnyone_encoder.mlpackage.zip", "mask_encoder": "MatAnyone_mask_encoder.mlpackage.zip", "read_first": "MatAnyone_read_first.mlpackage.zip", "read": "MatAnyone_read.mlpackage.zip", "decoder": "MatAnyone_decoder.mlpackage.zip" } }, "files": [ { "name": "MatAnyone_encoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/matanyone/MatAnyone_encoder.mlpackage.zip", "archive": "zip", "size_bytes": 17306121, "sha256": "97ffd6bc4611f9a3351dc890fc00954ba48171e517e66a39f7a5f1f38110dfda", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "MatAnyone_mask_encoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/matanyone/MatAnyone_mask_encoder.mlpackage.zip", "archive": "zip", "size_bytes": 16819866, "sha256": "ba67559188ffc64d8e46418c051c6a55815d4482def17519fa518daac7d5a911", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "MatAnyone_read_first.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/matanyone/MatAnyone_read_first.mlpackage.zip", "archive": "zip", "size_bytes": 21991849, "sha256": "34daf7227dbcec7373a3fef175259fa7ec631ed8cb91d5595ca57ee9b22df7bb", "compute_units": "cpuOnly", "kind": "model" }, { "name": "MatAnyone_read.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/matanyone/MatAnyone_read.mlpackage.zip", "archive": "zip", "size_bytes": 22135429, "sha256": "052e52c0ffb7ff9ede448128950cd4c1c9a96589b6900c82b5104d99addb7fa5", "compute_units": "cpuOnly", "kind": "model" }, { "name": "MatAnyone_decoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/matanyone/MatAnyone_decoder.mlpackage.zip", "archive": "zip", "size_bytes": 8807630, "sha256": "67136aa67000e604838fe9aa7de151c514ef84f0b83f1da0f043cf70652d28eb", "compute_units": "cpuAndGPU", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 800 }, "license": { "name": "MIT", "url": "https://github.com/pq-yang/MatAnyone" }, "upstream": { "name": "pq-yang/MatAnyone", "url": "https://github.com/pq-yang/MatAnyone", "year": 2025 } }, { "id": "demucs", "name": "HTDemucs", "subtitle": "Audio Source Separation", "category_id": "audio", "description_md": "Split music into 4 stems: drums, bass, vocals, other. 44.1 kHz stereo, FP32.", "demo": { "template": "audio_in_out", "config": { "sample_rate": 44100, "segment_length": 343980, "output_stems": [ "drums", "bass", "vocals", "other" ] } }, "files": [ { "name": "HTDemucs_SourceSeparation_F32.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/demucs/HTDemucs_SourceSeparation_F32.mlpackage.zip", "archive": "zip", "size_bytes": 79076395, "sha256": "0fbb941e15a5b2fa425d14fe630ed4c14b6dee72780c1f5b2b05f58803bce5f7", "compute_units": "cpuOnly", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 1000 }, "license": { "name": "MIT", "url": "https://github.com/adefossez/demucs" }, "upstream": { "name": "adefossez/demucs", "url": "https://github.com/adefossez/demucs", "year": 2021 } }, { "id": "kokoro", "name": "Kokoro-82M", "subtitle": "Multilingual TTS", "category_id": "speech", "description_md": "English + Japanese text-to-speech. 24 kHz. StyleTTS2 + iSTFTNet vocoder. Multiple voices.", "demo": { "template": "text_to_audio", "config": { "mode": "tts", "sample_rate": 24000, "vocab_file": "kokoro_vocab.json", "voices": [ "af_heart", "af_bella", "am_michael", "bf_emma", "bm_george" ] } }, "files": [ { "name": "Kokoro_Predictor.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/kokoro/Kokoro_Predictor.mlpackage.zip", "archive": "zip", "size_bytes": 72191470, "sha256": "af1d55dc842980c32b5591a70f603941f11ab60a435bed0c13a107a8ef467bed", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "Kokoro_Decoder_128.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/kokoro/Kokoro_Decoder_128.mlpackage.zip", "archive": "zip", "size_bytes": 229120589, "sha256": "cece0d072f5ba6aa3f729cf4c76b4de51823bcc65a26ab363c10441c3cd8b306", "compute_units": "all", "kind": "model" }, { "name": "Kokoro_Decoder_256.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/kokoro/Kokoro_Decoder_256.mlpackage.zip", "archive": "zip", "size_bytes": 229123438, "sha256": "36d5e16d5c5ccb500fc96f1b07a1d5ac57b791f8e09e61b78319d76949003efe", "compute_units": "all", "kind": "model" }, { "name": "Kokoro_Decoder_512.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/kokoro/Kokoro_Decoder_512.mlpackage.zip", "archive": "zip", "size_bytes": 229128735, "sha256": "0a44484c327e4fe8443b0bcf104d6964fe3f30d628c9e78aee3f31af7f2475dc", "compute_units": "all", "kind": "model" }, { "name": "kokoro_vocab.json", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/kokoro/kokoro_vocab.json", "size_bytes": 1144, "sha256": "70abefbe8a1c8865e43e0a43bbdc25b91a33e4aa053479d443ccf23e20a59e5d", "kind": "vocab" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 1000 }, "license": { "name": "Apache-2.0", "url": "https://huggingface.co/hexgrad/Kokoro-82M" }, "upstream": { "name": "hexgrad/Kokoro-82M", "url": "https://huggingface.co/hexgrad/Kokoro-82M", "year": 2024 } }, { "id": "stable_audio", "name": "Stable Audio Open", "subtitle": "Text-to-Music, 2024", "category_id": "speech", "description_md": "Text-to-music. Up to 11.9s stereo 44.1 kHz. Rectified flow DiT + T5 + Oobleck VAE.", "demo": { "template": "text_to_audio", "config": { "mode": "music", "sample_rate": 44100, "max_duration": 11.9 } }, "files": [ { "name": "StableAudioT5Encoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/stableaudio/StableAudioT5Encoder.mlpackage.zip", "archive": "zip", "size_bytes": 98538259, "sha256": "319a8ba775d309240253ced68a03a3923d0aec9a79f608044f9403bdcfe4b741", "compute_units": "cpuOnly", "kind": "model" }, { "name": "StableAudioNumberEmbedder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/stableaudio/StableAudioNumberEmbedder.mlpackage.zip", "archive": "zip", "size_bytes": 376018, "sha256": "04bdc5de00a2cf1c4a18f80c94f0d74ecfab41f3ad99f2fb7a031d6ff5af75da", "compute_units": "cpuOnly", "kind": "model" }, { "name": "StableAudioDiT.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/stableaudio/StableAudioDiT.mlpackage.zip", "archive": "zip", "size_bytes": 1265748504, "sha256": "b17da4fc4df857821d39dbdf7d3bfe7062a2272ab3e5df1284d545afb54047e4", "compute_units": "cpuOnly", "kind": "model" }, { "name": "StableAudioVAEDecoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/stableaudio/StableAudioVAEDecoder.mlpackage.zip", "archive": "zip", "size_bytes": 144960275, "sha256": "7207544cca9799cc1d6803c5e81badd0bb4527b2d3a64d5cab5700a5f19a9374", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "t5_vocab.json", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/stableaudio/t5_vocab.json", "size_bytes": 749757, "sha256": "7c9ff3ac1b3dbcaa617ee659f2df68688cfd44f1a5eb3be3fa0a2f56c749d56a", "kind": "vocab" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 1200 }, "license": { "name": "custom", "url": "https://huggingface.co/stabilityai/stable-audio-open-small" }, "upstream": { "name": "stabilityai/stable-audio-open-small", "url": "https://huggingface.co/stabilityai/stable-audio-open-small", "year": 2024 } }, { "id": "openvoice", "name": "OpenVoice V2", "subtitle": "Voice Cloning", "category_id": "audio", "description_md": "Zero-shot voice conversion. Clone a speaker from ~10s reference audio.", "demo": { "template": "audio_in_out", "config": { "sample_rate": 22050, "output_stems": [ "converted" ] } }, "files": [ { "name": "OpenVoice_SpeakerEncoder.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/openvoice/OpenVoice_SpeakerEncoder.mlpackage.zip", "archive": "zip", "size_bytes": 1519880, "sha256": "c3f2a96aaf5ecb5c5afc62b3d3dfbd47dc7ae64bc3edb7aa68befb54aef74459", "compute_units": "cpuAndGPU", "kind": "model" }, { "name": "OpenVoice_VoiceConverter.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/openvoice/OpenVoice_VoiceConverter.mlpackage.zip", "archive": "zip", "size_bytes": 59799630, "sha256": "ef3ce8a2d1564aefa13830d7d0ca43f85e0aa62d5f59622c8bc456c307ab5e05", "compute_units": "cpuAndGPU", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 500 }, "license": { "name": "MIT", "url": "https://github.com/myshell-ai/OpenVoice" }, "upstream": { "name": "myshell-ai/OpenVoice", "url": "https://github.com/myshell-ai/OpenVoice", "year": 2023 } }, { "id": "diarization", "name": "Pyannote Diarization", "subtitle": "Speaker Identification", "category_id": "audio", "description_md": "Speaker diarization: who spoke when. 16 kHz mono, 10s segments.", "demo": { "template": "audio_in_out", "config": { "sample_rate": 16000, "output_stems": [ "speaker_timeline" ] } }, "files": [ { "name": "SpeakerSegmentation.mlpackage.zip", "url": "https://huggingface.co/mlboydaisuke/coreml-zoo/resolve/main/diarization/SpeakerSegmentation.mlpackage.zip", "archive": "zip", "size_bytes": 5327137, "sha256": "dcfa2b98900f2b99029abfb593644b70418186a6ec2e94c9a79c2b3d7a84378a", "compute_units": "cpuAndGPU", "kind": "model" } ], "requirements": { "min_ios": "17.0", "min_ram_mb": 200 }, "license": { "name": "MIT", "url": "https://github.com/pyannote/pyannote-audio" }, "upstream": { "name": "pyannote/pyannote-audio", "url": "https://github.com/pyannote/pyannote-audio", "year": 2021 } } ] }