RaushanTurganbay HF staff commited on
Commit
2b654a4
1 Parent(s): 34373d9

Upload processor

Browse files
README.md CHANGED
@@ -2,9 +2,9 @@
2
  language:
3
  - en
4
  license: llama2
5
- pipeline_tag: image-text-to-text
6
  datasets:
7
  - lmms-lab/VideoChatGPT
 
8
  ---
9
 
10
  # LLaVA-NeXT-Video Model Card
 
2
  language:
3
  - en
4
  license: llama2
 
5
  datasets:
6
  - lmms-lab/VideoChatGPT
7
+ pipeline_tag: image-text-to-text
8
  ---
9
 
10
  # LLaVA-NeXT-Video Model Card
processor_config.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' }}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] }}{% endfor %}{{' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'][0]['text'] + '<\\s> '}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
3
  "processor_class": "LlavaNextVideoProcessor"
4
  }
 
1
  {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' }}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all videos next #}{% for content in message['content'] | selectattr('type', 'equalto', 'video') %}{{ '<video>\n' }}{% endfor %}{# Render all text next #}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] }}{% endfor %}{{' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'][0]['text'] + '<\\s> '}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
3
  "processor_class": "LlavaNextVideoProcessor"
4
  }
special_tokens_map.json CHANGED
@@ -13,7 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "</s>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -50,8 +50,11 @@
50
  "clean_up_tokenization_spaces": false,
51
  "eos_token": "</s>",
52
  "legacy": true,
 
53
  "model_max_length": 1000000000000000019884624838656,
 
54
  "pad_token": "</s>",
 
55
  "padding_side": "left",
56
  "processor_class": "LlavaNextVideoProcessor",
57
  "sp_model_kwargs": {},
 
50
  "clean_up_tokenization_spaces": false,
51
  "eos_token": "</s>",
52
  "legacy": true,
53
+ "max_length": null,
54
  "model_max_length": 1000000000000000019884624838656,
55
+ "pad_to_multiple_of": null,
56
  "pad_token": "</s>",
57
+ "pad_token_type_id": 0,
58
  "padding_side": "left",
59
  "processor_class": "LlavaNextVideoProcessor",
60
  "sp_model_kwargs": {},