DeepCaption-VLA-V2.0-7B-AIO-GGUF

DeepCaption-VLA-V2.0-7B from prithivMLmods is an advanced 7B-parameter vision-language model fine-tuned from Qwen2.5-VL-7B-Instruct, enhancing the original DeepCaption-VLA with significantly improved multilingual inference for precise image captioning and Vision Language Attribution (VLA) across English, Chinese (Zh), Thai (Th), and other languages, generating attribute-rich descriptions of visual properties, object attributes, scene details, colors, environments, moods, and actions in diverse formats (wide, tall, square, irregular aspect ratios) for general, artistic, technical, abstract, or low-context images. V2.0 delivers higher captioning quality and attribution accuracy through refined training on curated datasets emphasizing semantic precision and descriptive fluency, supporting variational detail control—from concise summaries to fine-grained outputs—via structured formats like Caption:, Attributes:, and {class_name==core_theme}, making it ideal for research, dataset annotation, object detection, scene understanding, and multilingual creative applications using Transformers/Qwen2VLForConditionalGeneration. While robust across non-standard visuals, it may over-attribute ambiguous content, vary by prompt phrasing, or degrade slightly on highly synthetic domains, positioning it as an experimental upstream VLM for advanced vision-language tasks.

DeepCaption-VLA-V2.0-7B [GGUF]

File Name Quant Type File Size File Link
DeepCaption-VLA-V2.0-7B.IQ4_XS.gguf IQ4_XS 4.25 GB Download
DeepCaption-VLA-V2.0-7B.Q2_K.gguf Q2_K 3.02 GB Download
DeepCaption-VLA-V2.0-7B.Q3_K_L.gguf Q3_K_L 4.09 GB Download
DeepCaption-VLA-V2.0-7B.Q3_K_M.gguf Q3_K_M 3.81 GB Download
DeepCaption-VLA-V2.0-7B.Q3_K_S.gguf Q3_K_S 3.49 GB Download
DeepCaption-VLA-V2.0-7B.Q4_K_M.gguf Q4_K_M 4.68 GB Download
DeepCaption-VLA-V2.0-7B.Q4_K_S.gguf Q4_K_S 4.46 GB Download
DeepCaption-VLA-V2.0-7B.Q5_K_M.gguf Q5_K_M 5.44 GB Download
DeepCaption-VLA-V2.0-7B.Q5_K_S.gguf Q5_K_S 5.32 GB Download
DeepCaption-VLA-V2.0-7B.Q6_K.gguf Q6_K 6.25 GB Download
DeepCaption-VLA-V2.0-7B.Q8_0.gguf Q8_0 8.1 GB Download
DeepCaption-VLA-V2.0-7B.f16.gguf F16 15.2 GB Download
DeepCaption-VLA-V2.0-7B.mmproj-Q8_0.gguf mmproj-Q8_0 856 MB Download
DeepCaption-VLA-V2.0-7B.mmproj-f16.gguf mmproj-f16 1.35 GB Download

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
-
GGUF
Model size
8B params
Architecture
qwen2vl
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prithivMLmods/DeepCaption-VLA-V2.0-7B-AIO-GGUF

Quantized
(4)
this model