ybelkada commited on
Commit
af2d425
1 Parent(s): 8fd5792

Upload LlavaForVisionText2Text

Browse files
config.json CHANGED
@@ -5,6 +5,7 @@
5
  "ignore_index": -100,
6
  "image_token_index": 32000,
7
  "model_type": "llava",
 
8
  "projector_hidden_act": "gelu",
9
  "text_config": {
10
  "_name_or_path": "mistralai/Mistral-7B-v0.1",
 
5
  "ignore_index": -100,
6
  "image_token_index": 32000,
7
  "model_type": "llava",
8
+ "pad_token_id": 32001,
9
  "projector_hidden_act": "gelu",
10
  "text_config": {
11
  "_name_or_path": "mistralai/Mistral-7B-v0.1",
generation_config.json CHANGED
@@ -2,5 +2,6 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
 
5
  "transformers_version": "4.36.0.dev0"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "pad_token_id": 32001,
6
  "transformers_version": "4.36.0.dev0"
7
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb4bedb59b12072dd560621117fb220bf0c8527f302612c67882071f732305e8
3
  size 4891055040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec9698807268456e5a3fa74d657a8bffa5dc161baed731488775454880424011
3
  size 4891055040
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:562d07b8003fe4a5b3f86ee3f63a21adc3b3f3c3b3917b7cb9a1d5ad526ce543
3
- size 671131400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b288c14ec213a3acc8c3eda0611da931cd5d60a2dc6f16550972eb72b4b4fc2f
3
+ size 933799800
model.safetensors.index.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
  "metadata": {
3
- "total_size": 15477833728
4
  },
5
  "weight_map": {
 
6
  "language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
7
  "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
8
  "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 15740502016
4
  },
5
  "weight_map": {
6
+ "language_model.lm_head.weight": "model-00004-of-00004.safetensors",
7
  "language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
8
  "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
  "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",