Commit
•
6e45a58
0
Parent(s):
Duplicate from MaziyarPanahi/Meta-Llama-3.1-70B-Instruct-GGUF
Browse filesCo-authored-by: Maziyar Panahi <MaziyarPanahi@users.noreply.huggingface.co>
- .gitattributes +55 -0
- Meta-Llama-3.1-70B-Instruct-GGUF_imatrix.dat +3 -0
- Meta-Llama-3.1-70B-Instruct.IQ1_M.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.IQ1_S.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.IQ2_XS.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.IQ3_XS.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.IQ4_XS.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q2_K.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q3_K_L.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q3_K_M.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q3_K_S.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q4_K_M.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q4_K_S.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q5_K_M.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q5_K_S.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00001-of-00006.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00002-of-00006.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00003-of-00006.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00004-of-00006.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00005-of-00006.gguf +3 -0
- Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00006-of-00006.gguf +3 -0
- README.md +55 -0
.gitattributes
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
Meta-Llama-3.1-70B-Instruct.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
Meta-Llama-3.1-70B-Instruct.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
Meta-Llama-3.1-70B-Instruct.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
Meta-Llama-3.1-70B-Instruct.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
Meta-Llama-3.1-70B-Instruct.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
Meta-Llama-3.1-70B-Instruct.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
Meta-Llama-3.1-70B-Instruct.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
Meta-Llama-3.1-70B-Instruct.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
Meta-Llama-3.1-70B-Instruct-GGUF_imatrix.dat filter=lfs diff=lfs merge=lfs -text
|
45 |
+
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00001-of-00006.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00002-of-00006.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00003-of-00006.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00004-of-00006.gguf filter=lfs diff=lfs merge=lfs -text
|
49 |
+
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00005-of-00006.gguf filter=lfs diff=lfs merge=lfs -text
|
50 |
+
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00006-of-00006.gguf filter=lfs diff=lfs merge=lfs -text
|
51 |
+
Meta-Llama-3.1-70B-Instruct.IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text
|
52 |
+
Meta-Llama-3.1-70B-Instruct.IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text
|
53 |
+
Meta-Llama-3.1-70B-Instruct.IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
54 |
+
Meta-Llama-3.1-70B-Instruct.IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
55 |
+
Meta-Llama-3.1-70B-Instruct.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
Meta-Llama-3.1-70B-Instruct-GGUF_imatrix.dat
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e2b854a6bf2fb589e9e98cc6a55ac9c32ccf9d0dbbb8dd0aede7c4ef36bbb06
|
3 |
+
size 24922274
|
Meta-Llama-3.1-70B-Instruct.IQ1_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d160d9a4096358a8b43e71e9c4cecba6c70a0e1f78728302942fd02086583054
|
3 |
+
size 16751196288
|
Meta-Llama-3.1-70B-Instruct.IQ1_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1dd3d85e14861fb74e12481c3d38ccad605ab023d265f9f8f1db6cf8c4a43fb
|
3 |
+
size 15343483008
|
Meta-Llama-3.1-70B-Instruct.IQ2_XS.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c9f39a2cd7c38194da02a0fdfcad7cd1d07e3596c995c12304b72cad7df1d21
|
3 |
+
size 21142108288
|
Meta-Llama-3.1-70B-Instruct.IQ3_XS.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cd535ce55b9e8e501a81c9ffd2a520a3a941c546e8ed545b4d7460b6f90bed5c
|
3 |
+
size 29307730048
|
Meta-Llama-3.1-70B-Instruct.IQ4_XS.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e9c31040a7eea6fd2d987b031afe16ea380fa689401097a53a6f7f70c2aeed7
|
3 |
+
size 37902661760
|
Meta-Llama-3.1-70B-Instruct.Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08b70ba58c9bd0da4ec7c725cee05c7795565f2a142fff33c4c5612dfe8493b1
|
3 |
+
size 26375108736
|
Meta-Llama-3.1-70B-Instruct.Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4aca3bd69ed9ad42e4c7c6acb6bddcbd1897c1b0b2ae6449c5deb35d3dcafabc
|
3 |
+
size 37140592768
|
Meta-Llama-3.1-70B-Instruct.Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edaa3775e4f3599a9e3ff576492f9da11e18c10ccbbc1b52b67eb90660a1b0a9
|
3 |
+
size 34267494528
|
Meta-Llama-3.1-70B-Instruct.Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f953fe1cdeae9b4c6d8dd806d1c1223f2eaa5d900c23cf788d1aa0718c028ccd
|
3 |
+
size 30912051328
|
Meta-Llama-3.1-70B-Instruct.Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f16ab17da4521fe3ed7c5d7beed960d3fe7b5b64421ee9650aa53d6b649ccab
|
3 |
+
size 42520393856
|
Meta-Llama-3.1-70B-Instruct.Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7416595e813087dc83b504ca47236ae27f581f18627a731abe77ba8b4a638087
|
3 |
+
size 40347220096
|
Meta-Llama-3.1-70B-Instruct.Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6973a51ab0535ff2b2ab3bc5d1e68f1ac3b7e10ad0c2a6f9a8437005518503ec
|
3 |
+
size 49949816960
|
Meta-Llama-3.1-70B-Instruct.Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ed2763f5246fc23a3a7d01ef56ace5077dccb9beebea4a24eeb46cb8347798f
|
3 |
+
size 48657447040
|
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00001-of-00006.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4e2683c351f9dbf5af4e4210c6739b754815e6be03f0efc0ddbf100ea5788aa2
|
3 |
+
size 10697125376
|
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00002-of-00006.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae12924bb9397d5260efe744e13e81dda30d2b0e043f05958f4b2818b96e644c
|
3 |
+
size 10212744800
|
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00003-of-00006.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1113b6129394a7c2ddc039c14eb31e8d430f2865c909b4ed11a1a087162b33f
|
3 |
+
size 10020101728
|
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00004-of-00006.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab5001643fa91798e2d2d7058f665e9e62d7c3cad776c02e6998eaf351a2b1a3
|
3 |
+
size 9889324640
|
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00005-of-00006.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a56c61d2023334f0b2b5bf1a22c6415248fe46e220ef91df3afb2308c5f241d
|
3 |
+
size 9889324640
|
Meta-Llama-3.1-70B-Instruct.Q6_K.gguf-00006-of-00006.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:678835a6f5bc4315fa3f4fc70570dfe579b403d5a57b350adaa25c3baf61cbef
|
3 |
+
size 7179523008
|
README.md
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
- de
|
5 |
+
- fr
|
6 |
+
- it
|
7 |
+
- pt
|
8 |
+
- hi
|
9 |
+
- es
|
10 |
+
- th
|
11 |
+
tags:
|
12 |
+
- quantized
|
13 |
+
- 2-bit
|
14 |
+
- 3-bit
|
15 |
+
- 4-bit
|
16 |
+
- 5-bit
|
17 |
+
- 6-bit
|
18 |
+
- 8-bit
|
19 |
+
- GGUF
|
20 |
+
- text-generation
|
21 |
+
- text-generation
|
22 |
+
model_name: Meta-Llama-3.1-70B-Instruct-GGUF
|
23 |
+
base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
|
24 |
+
inference: false
|
25 |
+
model_creator: meta-llama
|
26 |
+
pipeline_tag: text-generation
|
27 |
+
quantized_by: MaziyarPanahi
|
28 |
+
---
|
29 |
+
# [MaziyarPanahi/Meta-Llama-3.1-70B-Instruct-GGUF](https://huggingface.co/MaziyarPanahi/Meta-Llama-3.1-70B-Instruct-GGUF)
|
30 |
+
- Model creator: [meta-llama](https://huggingface.co/meta-llama)
|
31 |
+
- Original model: [meta-llama/Meta-Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct)
|
32 |
+
|
33 |
+
## Description
|
34 |
+
[MaziyarPanahi/Meta-Llama-3.1-70B-Instruct-GGUF](https://huggingface.co/MaziyarPanahi/Meta-Llama-3.1-70B-Instruct-GGUF) contains GGUF format model files for [meta-llama/Meta-Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct).
|
35 |
+
|
36 |
+
### About GGUF
|
37 |
+
|
38 |
+
GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
|
39 |
+
|
40 |
+
Here is an incomplete list of clients and libraries that are known to support GGUF:
|
41 |
+
|
42 |
+
* [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option.
|
43 |
+
* [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), a Python library with GPU accel, LangChain support, and OpenAI-compatible API server.
|
44 |
+
* [LM Studio](https://lmstudio.ai/), an easy-to-use and powerful local GUI for Windows and macOS (Silicon), with GPU acceleration. Linux available, in beta as of 27/11/2023.
|
45 |
+
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration.
|
46 |
+
* [KoboldCpp](https://github.com/LostRuins/koboldcpp), a fully featured web UI, with GPU accel across all platforms and GPU architectures. Especially good for story telling.
|
47 |
+
* [GPT4All](https://gpt4all.io/index.html), a free and open source local running GUI, supporting Windows, Linux and macOS with full GPU accel.
|
48 |
+
* [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), a great web UI with many interesting and unique features, including a full model library for easy model selection.
|
49 |
+
* [Faraday.dev](https://faraday.dev/), an attractive and easy to use character-based chat GUI for Windows and macOS (both Silicon and Intel), with GPU acceleration.
|
50 |
+
* [candle](https://github.com/huggingface/candle), a Rust ML framework with a focus on performance, including GPU support, and ease of use.
|
51 |
+
* [ctransformers](https://github.com/marella/ctransformers), a Python library with GPU accel, LangChain support, and OpenAI-compatible AI server. Note, as of time of writing (November 27th 2023), ctransformers has not been updated in a long time and does not support many recent models.
|
52 |
+
|
53 |
+
## Special thanks
|
54 |
+
|
55 |
+
🙏 Special thanks to [Georgi Gerganov](https://github.com/ggerganov) and the whole team working on [llama.cpp](https://github.com/ggerganov/llama.cpp/) for making all of this possible.
|