File size: 2,166 Bytes
7bca7d8
 
1aedef8
7bca7d8
 
 
 
1aedef8
7bca7d8
 
520b051
4e8a531
2bc39e2
 
4a50b77
 
1aedef8
6c2e62a
1aedef8
9c3ebda
 
8c4dbe0
bb48740
9b598fc
 
1aedef8
8c4dbe0
520b051
bb48740
22e4934
62af628
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
---
title: README
emoji: ๐Ÿ‘€
colorFrom: purple
colorTo: indigo
sdk: static
pinned: false
short_description: VLM assets for Llama-3.2-11B-Vision-Instruct
---

# ๐Ÿ‡ณ๐Ÿ‡ด๐Ÿ‡ฉ๐Ÿ‡ฐ Open Source Vision Language Model assets
Building on the philosophy of open source with the Llama-models ๐Ÿฆ™, this repo is an effort to support development of small VLM's in the Scandinavian languages. Aa we are only fluent in Norwegian and Danish, we have focused on these two languages. However, we encourgage the community (๐Ÿ‡ซ๐Ÿ‡ฎ๐Ÿ‡ธ๐Ÿ‡ช๐Ÿ‡ซ๐Ÿ‡ด๐Ÿ‡ฎ๐Ÿ‡ธ๐Ÿ‡ฌ๐Ÿ‡ฑSami) to help build on our work and extend the coverage. 

The current models and data focus on transcription and annotiation of documents in Norwegian and Danish, going beyond the limitations of OCR.

We expect this line of work to help businesses, government institutions and citizens alike. Please se <repo> for how to run inference on the final models.

# In these collections you will find:
- ๐Ÿ’ฝ Datasets for fine-tuning VLM
  - ๐Ÿ‡ณ๐Ÿ‡ด See collection: https://huggingface.co/collections/MykMaks/datasets-nb-679f081d89be13de6a9fe71b
  - ๐Ÿ‡ฉ๐Ÿ‡ฐ See collection: https://huggingface.co/collections/MykMaks/datasets-da-679f07b68e587e67bba71fdd
- ๐Ÿ’พ Training code
  - Approach: We trained every epoch with a different prompt, stored the adapter as a checkpoint and continued to next prompt-dataset pair.
  - MM checkpoints: https://github.com/Mikeriess/llama33_resources/tree/MM-models
  - V-I checkpoints: https://github.com/Mikeriess/llama33_resources/tree/v-i-models
- ๐Ÿค– Model LORA-adapter checkpoints for Llama-3.2-11B-Vision-Instruct
  - The model is iteratively trained over all datasets:
    - The suffix of each file denotes the order of the checkpoint, along with the dataset that it was fine-tuned on
    - Prompts can be tracked in the respective experiment.json files in the MM and V-I code repositories
- ๐Ÿ’ธ Final full-precision merged models:
  - See collection: ๐Ÿฆ™ https://huggingface.co/collections/MykMaks/models-679f08ab3ea3e21df62c87e8
    - <b>MykMaks/llama-3.2-11B-MM-20-MykMaks_da-wit-merged</b>
    - <b>MykMaks/llama-3.2-11B-V-I_39_MykMaks_NorwegianDataset-compressed-pt2-merged</b>