prince-canuma commited on
Commit
0154197
1 Parent(s): e8e3f49

Upload 6 files

Browse files
config.json ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aligner_config": {
3
+ "cls": "MlpProjector",
4
+ "model_type": "aligner",
5
+ "params": {
6
+ "depth": 2,
7
+ "input_dim": 1024,
8
+ "n_embed": 4096,
9
+ "projector_type": "low_high_hybrid_split_mlp_gelu"
10
+ }
11
+ },
12
+ "architectures": [
13
+ "MultiModalityCausalLM"
14
+ ],
15
+ "language_config": {
16
+ "max_position_embeddings": 16384,
17
+ "model_type": "llama",
18
+ "num_hidden_layers": 30,
19
+ "torch_dtype": "float16",
20
+ "vocab_size": 102400
21
+ },
22
+ "model_type": "multi_modality",
23
+ "quantization": {
24
+ "group_size": 64,
25
+ "bits": 4
26
+ },
27
+ "torch_dtype": "float16",
28
+ "transformers_version": "4.38.2",
29
+ "vision_config": {
30
+ "cls": "HybridVisionTower",
31
+ "model_type": "vision",
32
+ "params": {
33
+ "concat_type": "tuple",
34
+ "freeze_high": true,
35
+ "freeze_low": true,
36
+ "high_res_cfg": {
37
+ "ckpt_path": "",
38
+ "image_size": 1024,
39
+ "model_name": "sam_b_downsample",
40
+ "output_dim": 1024,
41
+ "pixel_mean": [
42
+ 0.48145466,
43
+ 0.4578275,
44
+ 0.40821073
45
+ ],
46
+ "pixel_std": [
47
+ 0.26862954,
48
+ 0.26130258,
49
+ 0.27577711
50
+ ],
51
+ "select_feature": "same",
52
+ "select_layer": -1
53
+ },
54
+ "low_res_cfg": {
55
+ "ckpt_path": "",
56
+ "image_size": 384,
57
+ "model_name": "siglip_large_patch16_384",
58
+ "output_dim": 1024,
59
+ "pixel_mean": [
60
+ 0.5,
61
+ 0.5,
62
+ 0.5
63
+ ],
64
+ "pixel_std": [
65
+ 0.5,
66
+ 0.5,
67
+ 0.5
68
+ ],
69
+ "select_feature": "same",
70
+ "select_layer": -1
71
+ }
72
+ },
73
+ "intermediate_size": 4096
74
+ }
75
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d6164c5f2b9a3999e395c7fb9b0d6db62441843ec229131b5bf50d1046b2dda
3
+ size 4150452457
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<image_placeholder>"
4
+ ],
5
+ "bos_token": {
6
+ "content": "<|begin▁of▁sentence|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "content": "<|end▁of▁sentence|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ }
19
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "100000": {
6
+ "content": "<|begin▁of▁sentence|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "100001": {
14
+ "content": "<|end▁of▁sentence|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "100002": {
22
+ "content": "ø",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": false
28
+ },
29
+ "100003": {
30
+ "content": "ö",
31
+ "lstrip": false,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": false
36
+ },
37
+ "100004": {
38
+ "content": "ú",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": false
44
+ },
45
+ "100005": {
46
+ "content": "ÿ",
47
+ "lstrip": false,
48
+ "normalized": true,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": false
52
+ },
53
+ "100006": {
54
+ "content": "õ",
55
+ "lstrip": false,
56
+ "normalized": true,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": false
60
+ },
61
+ "100007": {
62
+ "content": "÷",
63
+ "lstrip": false,
64
+ "normalized": true,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": false
68
+ },
69
+ "100008": {
70
+ "content": "û",
71
+ "lstrip": false,
72
+ "normalized": true,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": false
76
+ },
77
+ "100009": {
78
+ "content": "ý",
79
+ "lstrip": false,
80
+ "normalized": true,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": false
84
+ },
85
+ "100010": {
86
+ "content": "À",
87
+ "lstrip": false,
88
+ "normalized": true,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": false
92
+ },
93
+ "100011": {
94
+ "content": "ù",
95
+ "lstrip": false,
96
+ "normalized": true,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": false
100
+ },
101
+ "100012": {
102
+ "content": "Á",
103
+ "lstrip": false,
104
+ "normalized": true,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": false
108
+ },
109
+ "100013": {
110
+ "content": "þ",
111
+ "lstrip": false,
112
+ "normalized": true,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": false
116
+ },
117
+ "100014": {
118
+ "content": "ü",
119
+ "lstrip": false,
120
+ "normalized": true,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "100015": {
126
+ "content": "<image>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": true
132
+ }
133
+ },
134
+ "additional_special_tokens": [
135
+ "<image>"
136
+ ],
137
+ "bos_token": "<|begin▁of▁sentence|>",
138
+ "clean_up_tokenization_spaces": false,
139
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<<SYS>>\\n' + system_message + '\\n\\n'%}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ 'You are a helpful language and vision assistant. You are able to understand the visual content that the user provides, and assist the user with a variety of tasks using natural language.\n\n' + 'User: ' + content.strip() + '\n\n' }}{% elif message['role'] == 'assistant' %}{{ 'Assistant: ' + content.strip() + '\n\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'Assistant:' }}{% endif %}",
140
+ "eos_token": "<|end▁of▁sentence|>",
141
+ "model_max_length": 16384,
142
+ "pad_token": null,
143
+ "processor_class": "VLChatProcessor",
144
+ "tokenizer_class": "LlamaTokenizer",
145
+ "unk_token": null,
146
+ "use_default_system_prompt": false
147
+ }