ποΈ LFM2-VL
Collection
LFM2-VL is our first series of vision-language models, designed for on-device deployment. β’ 10 items β’ Updated β’ 66
llm.create_chat_completion(
messages = [
{
"role": "user",
"content": [
{
"type": "text",
"text": "Describe this image in one sentence."
},
{
"type": "image_url",
"image_url": {
"url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
}
}
]
}
]
)
LFM2-VL is a new generation of vision models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.
Find more details in the original model card: https://huggingface.co/LiquidAI/LFM2-VL-3B
Example usage with llama.cpp:
full precision (F16/F16):
llama-mtmd-cli -hf LiquidAI/LFM2-VL-3B-GGUF:F16
fastest inference (Q4_0/Q8_0):
llama-mtmd-cli -hf LiquidAI/LFM2-VL-3B-GGUF:Q4_0
4-bit
8-bit
16-bit
Base model
LiquidAI/LFM2-VL-3B
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="LiquidAI/LFM2-VL-3B-GGUF", filename="", )