README / README.md
mikeriess's picture
Update README.md
62af628 verified
---
title: README
emoji: ๐Ÿ‘€
colorFrom: purple
colorTo: indigo
sdk: static
pinned: false
short_description: VLM assets for Llama-3.2-11B-Vision-Instruct
---
# ๐Ÿ‡ณ๐Ÿ‡ด๐Ÿ‡ฉ๐Ÿ‡ฐ Open Source Vision Language Model assets
Building on the philosophy of open source with the Llama-models ๐Ÿฆ™, this repo is an effort to support development of small VLM's in the Scandinavian languages. Aa we are only fluent in Norwegian and Danish, we have focused on these two languages. However, we encourgage the community (๐Ÿ‡ซ๐Ÿ‡ฎ๐Ÿ‡ธ๐Ÿ‡ช๐Ÿ‡ซ๐Ÿ‡ด๐Ÿ‡ฎ๐Ÿ‡ธ๐Ÿ‡ฌ๐Ÿ‡ฑSami) to help build on our work and extend the coverage.
The current models and data focus on transcription and annotiation of documents in Norwegian and Danish, going beyond the limitations of OCR.
We expect this line of work to help businesses, government institutions and citizens alike. Please se <repo> for how to run inference on the final models.
# In these collections you will find:
- ๐Ÿ’ฝ Datasets for fine-tuning VLM
- ๐Ÿ‡ณ๐Ÿ‡ด See collection: https://huggingface.co/collections/MykMaks/datasets-nb-679f081d89be13de6a9fe71b
- ๐Ÿ‡ฉ๐Ÿ‡ฐ See collection: https://huggingface.co/collections/MykMaks/datasets-da-679f07b68e587e67bba71fdd
- ๐Ÿ’พ Training code
- Approach: We trained every epoch with a different prompt, stored the adapter as a checkpoint and continued to next prompt-dataset pair.
- MM checkpoints: https://github.com/Mikeriess/llama33_resources/tree/MM-models
- V-I checkpoints: https://github.com/Mikeriess/llama33_resources/tree/v-i-models
- ๐Ÿค– Model LORA-adapter checkpoints for Llama-3.2-11B-Vision-Instruct
- The model is iteratively trained over all datasets:
- The suffix of each file denotes the order of the checkpoint, along with the dataset that it was fine-tuned on
- Prompts can be tracked in the respective experiment.json files in the MM and V-I code repositories
- ๐Ÿ’ธ Final full-precision merged models:
- See collection: ๐Ÿฆ™ https://huggingface.co/collections/MykMaks/models-679f08ab3ea3e21df62c87e8
- <b>MykMaks/llama-3.2-11B-MM-20-MykMaks_da-wit-merged</b>
- <b>MykMaks/llama-3.2-11B-V-I_39_MykMaks_NorwegianDataset-compressed-pt2-merged</b>