LFM2.5-VL-1.6B CoreML (ANE Native)
CoreML ML Program export of LFM2.5-VL-1.6B for Apple Silicon with a fixed 4096-token context window, including the text decoder plus split multimodal vision assets.
Swift inference package and CLI: https://github.com/mweinbach/LFM2.5-VL-1.6B-ANE-Native
Included files
CoreMLModels/Embeddings.mlpackageCoreMLModels/LMHead.mlpackageCoreMLModels/DecoderChunk{0-3}Prefill.mlpackageCoreMLModels/DecoderChunk{0-3}Decode.mlpackageCoreMLModels/VisionPatchEmbedding.mlpackageCoreMLModels/VisionEncoder.mlpackageCoreMLModels/VisionProjector.mlpackageCoreMLModels/VisionPositionEmbeddings.float16.binCoreMLModels/meta.jsontokenizer.jsontokenizer_config.jsonconfig.jsonprocessor_config.jsongeneration_config.jsonchat_template.jinja
Runtime characteristics
- 4096-token fixed context window
- chunked decoder export for prompt prefill + single-token decode
- split vision path: patch embedding + encoder + projector
- supports multi-image prompts and the model's tiling/thumbnail preprocessing flow
- Apple CoreML / ML Program target
- tokenizer and chat template included for local execution
Usage
Clone or download this repo, then point the Swift CLI at the downloaded folder.
Text-only example:
swift run ANEInferenceCLI \
--bundle-root /path/to/LFM2.5-VL-1.6B-CoreML \
--prompt "Summarize the benefits of on-device inference" \
--max-new-tokens 64
Multimodal example:
swift run ANEInferenceCLI \
--bundle-root /path/to/LFM2.5-VL-1.6B-CoreML \
--image /path/to/image.png \
--prompt "Describe the image briefly." \
--max-new-tokens 64
The matching Swift package is here: https://github.com/mweinbach/LFM2.5-VL-1.6B-ANE-Native
- Downloads last month
- 324