prince-canuma commited on
Commit
87c501a
1 Parent(s): 6718dc8

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -176,13 +176,31 @@ extra_gated_fields:
176
  extra_gated_description: The information you provide will be collected, stored, processed
177
  and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
178
  extra_gated_button_content: Submit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
179
  ---
180
 
181
  # mlx-community/Meta-Llama-3-70B-Instruct-4bit
182
- This model was converted to MLX format from [`meta-llama/Meta-Llama-3-70B-Instruct`]() using mlx-lm version **0.9.0**.
183
-
184
- Model added by [Prince Canuma](https://twitter.com/Prince_Canuma).
185
-
186
  Refer to the [original model card](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) for more details on the model.
187
  ## Use with mlx
188
 
 
176
  extra_gated_description: The information you provide will be collected, stored, processed
177
  and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
178
  extra_gated_button_content: Submit
179
+ widget:
180
+ - example_title: Winter holidays
181
+ messages:
182
+ - role: system
183
+ content: You are a helpful and honest assistant. Please, respond concisely and
184
+ truthfully.
185
+ - role: user
186
+ content: Can you recommend a good destination for Winter holidays?
187
+ - example_title: Programming assistant
188
+ messages:
189
+ - role: system
190
+ content: You are a helpful and honest code and programming assistant. Please,
191
+ respond concisely and truthfully.
192
+ - role: user
193
+ content: Write a function that computes the nth fibonacci number.
194
+ inference:
195
+ parameters:
196
+ max_new_tokens: 300
197
+ stop:
198
+ - <|end_of_text|>
199
+ - <|eot_id|>
200
  ---
201
 
202
  # mlx-community/Meta-Llama-3-70B-Instruct-4bit
203
+ This model was converted to MLX format from [`meta-llama/Meta-Llama-3-70B-Instruct`]() using mlx-lm version **0.10.0**.
 
 
 
204
  Refer to the [original model card](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) for more details on the model.
205
  ## Use with mlx
206
 
config.json CHANGED
@@ -1,85 +1,31 @@
1
  {
2
- "add_cross_attention": false,
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
- "bad_words_ids": null,
9
- "begin_suppress_tokens": null,
10
  "bos_token_id": 128000,
11
- "chunk_size_feed_forward": 0,
12
- "cross_attention_hidden_size": null,
13
- "decoder_start_token_id": null,
14
- "diversity_penalty": 0.0,
15
- "do_sample": false,
16
- "early_stopping": false,
17
- "encoder_no_repeat_ngram_size": 0,
18
  "eos_token_id": 128001,
19
- "exponential_decay_length_penalty": null,
20
- "finetuning_task": null,
21
- "forced_bos_token_id": null,
22
- "forced_eos_token_id": null,
23
  "hidden_act": "silu",
24
  "hidden_size": 8192,
25
- "id2label": {
26
- "0": "LABEL_0",
27
- "1": "LABEL_1"
28
- },
29
  "initializer_range": 0.02,
30
  "intermediate_size": 28672,
31
- "is_decoder": false,
32
- "is_encoder_decoder": false,
33
- "label2id": {
34
- "LABEL_0": 0,
35
- "LABEL_1": 1
36
- },
37
- "length_penalty": 1.0,
38
- "max_length": 20,
39
  "max_position_embeddings": 8192,
40
- "min_length": 0,
41
  "model_type": "llama",
42
- "no_repeat_ngram_size": 0,
43
  "num_attention_heads": 64,
44
- "num_beam_groups": 1,
45
- "num_beams": 1,
46
  "num_hidden_layers": 80,
47
  "num_key_value_heads": 8,
48
- "num_return_sequences": 1,
49
- "output_attentions": false,
50
- "output_hidden_states": false,
51
- "output_scores": false,
52
- "pad_token_id": null,
53
- "prefix": null,
54
  "pretraining_tp": 1,
55
- "problem_type": null,
56
- "pruned_heads": {},
57
  "quantization": {
58
  "group_size": 64,
59
  "bits": 4
60
  },
61
- "remove_invalid_values": false,
62
- "repetition_penalty": 1.0,
63
- "return_dict": true,
64
- "return_dict_in_generate": false,
65
  "rms_norm_eps": 1e-05,
66
  "rope_scaling": null,
67
  "rope_theta": 500000.0,
68
- "sep_token_id": null,
69
- "suppress_tokens": null,
70
- "task_specific_params": null,
71
- "temperature": 1.0,
72
- "tf_legacy_loss": false,
73
- "tie_encoder_decoder": false,
74
  "tie_word_embeddings": false,
75
- "tokenizer_class": null,
76
- "top_k": 50,
77
- "top_p": 1.0,
78
  "torch_dtype": "bfloat16",
79
- "torchscript": false,
80
- "transformers_version": "4.40.0",
81
- "typical_p": 1.0,
82
- "use_bfloat16": false,
83
  "use_cache": true,
84
  "vocab_size": 128256
85
  }
 
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
 
7
  "bos_token_id": 128000,
 
 
 
 
 
 
 
8
  "eos_token_id": 128001,
 
 
 
 
9
  "hidden_act": "silu",
10
  "hidden_size": 8192,
 
 
 
 
11
  "initializer_range": 0.02,
12
  "intermediate_size": 28672,
 
 
 
 
 
 
 
 
13
  "max_position_embeddings": 8192,
 
14
  "model_type": "llama",
 
15
  "num_attention_heads": 64,
 
 
16
  "num_hidden_layers": 80,
17
  "num_key_value_heads": 8,
 
 
 
 
 
 
18
  "pretraining_tp": 1,
 
 
19
  "quantization": {
20
  "group_size": 64,
21
  "bits": 4
22
  },
 
 
 
 
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": null,
25
  "rope_theta": 500000.0,
 
 
 
 
 
 
26
  "tie_word_embeddings": false,
 
 
 
27
  "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.40.0.dev0",
 
 
 
29
  "use_cache": true,
30
  "vocab_size": 128256
31
  }
model-00001-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6db3fd6ec2266ed400f0a967358f72c2ce4b4df53513a4e0365f2210f1e92f4b
3
- size 5338514834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fad2c312718cadd55d3fe50fd5ae44cce500f9b61513ba92b180fc46cc34e49e
3
+ size 5272167770
model-00002-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0be6811819d56923a5aede8271ba3838b5c51998cb523103934341f02c25ef9
3
- size 5294649665
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffb4f11fa816e5532f014eac891d3ab6da9b0e9f58d951cc7cf3e10dcd9715cb
3
+ size 5294649694
model-00003-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35e0d9e2217379bef4e1a890f8ccec67527854e0fb95e6b15ed62719e4e38441
3
- size 5294649735
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c60d9e5e491175430f7c8f526821afb5ccea4028cf541a945d5683e1333b0ab3
3
+ size 5294649717
model-00004-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13973f7214cead910f4d0a45928ac338fd7a5bbefb08b2c4aae3942cfc102c33
3
- size 5294649737
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf71556de92e26d1299a55e0f07dd3103ea88e724b38cdd4e7c7240552215011
3
+ size 5294649733
model-00005-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e85e1c9358138723de2d13c4ff3166419c2337d43c5fffbc72580c83b2a2028a
3
- size 5294649745
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5bb8c45b4242c3b0850b775188732cdb0c8ac4b4cd557edfac7041030c0529c
3
+ size 5294649717
model-00006-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96946683ba5d32e811b35a3c4373564ded1dbcc039f9f04b9615cdb98a93b6ef
3
- size 5294649715
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26397f69a4bd11ee5a94926b5700228b23563bb1aca1bfd51d84712341207174
3
+ size 5294649733
model-00007-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eb5deace8a8b18ea5aecca75c1ce5b1e78e89e84db3f9e980c5bd1427d9a169
3
- size 5294649715
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32e20364eee71bde6fbfafe11b9fc65a5be8a8f6ddd1383b785114497f4a5fe8
3
+ size 5294649739
model-00008-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94722b9eaef1d6638523202d91c9614611d209257c33caa449647871aba886ff
3
- size 4092496847
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f163ab85b3a035cedd77d53960d1ed65442ca0b6e33462329e0883692c52a562
3
+ size 2648501502
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json CHANGED
@@ -7,7 +7,7 @@
7
  "single_word": false
8
  },
9
  "eos_token": {
10
- "content": "<|eot_id|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
 
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|end_of_text|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
tokenizer.json CHANGED
@@ -2348,6 +2348,7 @@
2348
  "end_of_word_suffix": null,
2349
  "fuse_unk": false,
2350
  "byte_fallback": false,
 
2351
  "vocab": {
2352
  "!": 0,
2353
  "\"": 1,
 
2348
  "end_of_word_suffix": null,
2349
  "fuse_unk": false,
2350
  "byte_fallback": false,
2351
+ "ignore_merges": false,
2352
  "vocab": {
2353
  "!": 0,
2354
  "\"": 1,
tokenizer_config.json CHANGED
@@ -2050,9 +2050,9 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}",
2054
  "clean_up_tokenization_spaces": true,
2055
- "eos_token": "<|eot_id|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|end_of_text|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"