| --- |
| license: apache-2.0 |
| task_categories: |
| - video-text-to-text |
| - image-to-text |
| language: |
| - en |
| tags: |
| - colab |
| - notebook |
| - demo |
| - vlm |
| - models |
| - hf |
| - ocr |
| - reasoning |
| - code |
| size_categories: |
| - n<1K |
| --- |
| # **VLM-Video-Understanding** |
|
|
| > A minimalistic demo for image inference and video understanding using OpenCV, built on top of several popular open-source Vision-Language Models (VLMs). This repository provides Colab notebooks demonstrating how to apply these VLMs to video and image tasks using Python and Gradio. |
|
|
| ## Overview |
|
|
| This project showcases lightweight inference pipelines for the following: |
| - Video frame extraction and preprocessing |
| - Image-level inference with VLMs |
| - Real-time or pre-recorded video understanding |
| - OCR-based text extraction from video frames |
|
|
| ## Models Included |
|
|
| The repository supports a variety of open-source models and configurations, including: |
|
|
| - Aya-Vision-8B |
| - Florence-2-Base |
| - Gemma3-VL |
| - MiMo-VL-7B-RL |
| - MiMo-VL-7B-SFT |
| - Qwen2-VL |
| - Qwen2.5-VL |
| - Qwen-2VL-MessyOCR |
| - RolmOCR-Qwen2.5-VL |
| - olmOCR-Qwen2-VL |
| - typhoon-ocr-7b-Qwen2.5VL |
|
|
| Each model has a dedicated Colab notebook to help users understand how to use it with video inputs. |
|
|
| ## Technologies Used |
|
|
| - **Python** |
| - **OpenCV** – for video and image processing |
| - **Gradio** – for interactive UI |
| - **Jupyter Notebooks** – for easy experimentation |
| - **Hugging Face Transformers** – for loading VLMs |
|
|
| ## Folder Structure |
|
|
| ``` |
| |
| ├── Aya-Vision-8B/ |
| ├── Florence-2-Base/ |
| ├── Gemma3-VL/ |
| ├── MiMo-VL-7B-RL/ |
| ├── MiMo-VL-7B-SFT/ |
| ├── Qwen2-VL/ |
| ├── Qwen2.5-VL/ |
| ├── Qwen-2VL-MessyOCR/ |
| ├── RolmOCR-Qwen2.5-VL/ |
| ├── olmOCR-Qwen2-VL/ |
| ├── typhoon-ocr-7b-Qwen2.5VL/ |
| ├── LICENSE |
| └── README.md |
| |
| ```` |
|
|
| ## Getting Started |
|
|
| 1. Clone the repository: |
|
|
| ```bash |
| git clone https://github.com/PRITHIVSAKTHIUR/VLM-Video-Understanding.git |
| cd VLM-Video-Understanding |
| ```` |
|
|
| 2. Open any of the Colab notebooks and follow the instructions to run image or video inference. |
|
|
| 3. Optionally, install dependencies locally: |
|
|
| ```bash |
| pip install opencv-python gradio transformers |
| ``` |
|
|
| ## Hugging Face Dataset |
|
|
| The models and examples are supported by a dataset on Hugging Face: |
|
|
| [VLM-Video-Understanding](https://huggingface.co/datasets/prithivMLmods/VLM-Video-Understanding) |
|
|
| ## License |
|
|
| This project is licensed under the Apache-2.0 License. |