File size: 2,166 Bytes
7bca7d8 1aedef8 7bca7d8 1aedef8 7bca7d8 520b051 4e8a531 2bc39e2 4a50b77 1aedef8 6c2e62a 1aedef8 9c3ebda 8c4dbe0 bb48740 9b598fc 1aedef8 8c4dbe0 520b051 bb48740 22e4934 62af628 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
---
title: README
emoji: ๐
colorFrom: purple
colorTo: indigo
sdk: static
pinned: false
short_description: VLM assets for Llama-3.2-11B-Vision-Instruct
---
# ๐ณ๐ด๐ฉ๐ฐ Open Source Vision Language Model assets
Building on the philosophy of open source with the Llama-models ๐ฆ, this repo is an effort to support development of small VLM's in the Scandinavian languages. Aa we are only fluent in Norwegian and Danish, we have focused on these two languages. However, we encourgage the community (๐ซ๐ฎ๐ธ๐ช๐ซ๐ด๐ฎ๐ธ๐ฌ๐ฑSami) to help build on our work and extend the coverage.
The current models and data focus on transcription and annotiation of documents in Norwegian and Danish, going beyond the limitations of OCR.
We expect this line of work to help businesses, government institutions and citizens alike. Please se <repo> for how to run inference on the final models.
# In these collections you will find:
- ๐ฝ Datasets for fine-tuning VLM
- ๐ณ๐ด See collection: https://huggingface.co/collections/MykMaks/datasets-nb-679f081d89be13de6a9fe71b
- ๐ฉ๐ฐ See collection: https://huggingface.co/collections/MykMaks/datasets-da-679f07b68e587e67bba71fdd
- ๐พ Training code
- Approach: We trained every epoch with a different prompt, stored the adapter as a checkpoint and continued to next prompt-dataset pair.
- MM checkpoints: https://github.com/Mikeriess/llama33_resources/tree/MM-models
- V-I checkpoints: https://github.com/Mikeriess/llama33_resources/tree/v-i-models
- ๐ค Model LORA-adapter checkpoints for Llama-3.2-11B-Vision-Instruct
- The model is iteratively trained over all datasets:
- The suffix of each file denotes the order of the checkpoint, along with the dataset that it was fine-tuned on
- Prompts can be tracked in the respective experiment.json files in the MM and V-I code repositories
- ๐ธ Final full-precision merged models:
- See collection: ๐ฆ https://huggingface.co/collections/MykMaks/models-679f08ab3ea3e21df62c87e8
- <b>MykMaks/llama-3.2-11B-MM-20-MykMaks_da-wit-merged</b>
- <b>MykMaks/llama-3.2-11B-V-I_39_MykMaks_NorwegianDataset-compressed-pt2-merged</b> |