update to new format
Browse files- README.md +1 -1
- params.json +9 -0
- CodeLlama-7b.npz → weights.npz +2 -2
README.md
CHANGED
@@ -30,7 +30,7 @@ export HF_HUB_ENABLE_HF_TRANSFER=1
|
|
30 |
huggingface-cli download --local-dir CodeLlama-7b-mlx mlx-llama/CodeLlama-7b-mlx
|
31 |
|
32 |
# Run example
|
33 |
-
python mlx-examples/llama/llama.py CodeLlama-7b-mlx/
|
34 |
```
|
35 |
|
36 |
Please, refer to the [original model card](https://github.com/facebookresearch/codellama/blob/main/MODEL_CARD.md) for details on CodeLlama.
|
|
|
30 |
huggingface-cli download --local-dir CodeLlama-7b-mlx mlx-llama/CodeLlama-7b-mlx
|
31 |
|
32 |
# Run example
|
33 |
+
python mlx-examples/llama/llama.py CodeLlama-7b-mlx/ CodeLlama-7b-mlx/tokenizer.model "def fibonacci("
|
34 |
```
|
35 |
|
36 |
Please, refer to the [original model card](https://github.com/facebookresearch/codellama/blob/main/MODEL_CARD.md) for details on CodeLlama.
|
params.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dim": 4096,
|
3 |
+
"n_layers": 32,
|
4 |
+
"n_heads": 32,
|
5 |
+
"multiple_of": 256,
|
6 |
+
"ffn_dim_multiplier": 1.0,
|
7 |
+
"norm_eps": 1e-5,
|
8 |
+
"rope_theta": 1000000
|
9 |
+
}
|
CodeLlama-7b.npz → weights.npz
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b153c305decd738a4486cfabad0e4afbd6ef77b9ee80c62e6b0584ef65e142e
|
3 |
+
size 13477181452
|