Jared Van Bortel
commited on
Commit
·
a4ff1a1
1
Parent(s):
d7e1f72
README: add note about llama.cpp version
Browse files
README.md
CHANGED
@@ -25,6 +25,8 @@ This repo contains llama.cpp-compatible files for [nomic-embed-text-v1](https://
|
|
25 |
|
26 |
llama.cpp will default to 2048 tokens of context with these files. To use the full 8192 tokens that Nomic Embed is benchmarked on, you will have to choose a context extension method. The original model uses Dynamic NTK-Aware RoPE scaling, but that is not currently available in llama.cpp. A combination of YaRN and linear scaling is an acceptable substitute.
|
27 |
|
|
|
|
|
28 |
## Example `llama.cpp` Command
|
29 |
|
30 |
Compute a single embedding:
|
|
|
25 |
|
26 |
llama.cpp will default to 2048 tokens of context with these files. To use the full 8192 tokens that Nomic Embed is benchmarked on, you will have to choose a context extension method. The original model uses Dynamic NTK-Aware RoPE scaling, but that is not currently available in llama.cpp. A combination of YaRN and linear scaling is an acceptable substitute.
|
27 |
|
28 |
+
These files were converted and quantized with llama.cpp commit [6c00a0669](https://github.com/ggerganov/llama.cpp/commit/6c00a066928b0475b865a2e3e709e2166e02d548).
|
29 |
+
|
30 |
## Example `llama.cpp` Command
|
31 |
|
32 |
Compute a single embedding:
|