IDDT
commited on
Commit
•
9298d7d
1
Parent(s):
e9a1ace
Update readme.
Browse files
README.md
CHANGED
@@ -1,3 +1,28 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Refact-1.6B-fim-GGUF
|
2 |
+
- Model creator: [Small Magellanic Cloud AI](https://huggingface.co/smallcloudai)
|
3 |
+
- Original model: [Refact-1.6B](https://huggingface.co/smallcloudai/Refact-1_6B-fim)
|
4 |
+
|
5 |
+
|
6 |
+
## Description
|
7 |
+
This repository contains quantized GGUF format model files for [Refact-1.6B](https://huggingface.co/smallcloudai/Refact-1_6B-fim).
|
8 |
+
|
9 |
+
|
10 |
+
## Prompt: fill in the middle
|
11 |
+
```
|
12 |
+
<fim_prefix>def print_hello_world():\n """<fim_suffix>\n print("Hello world!")<fim_middle>
|
13 |
+
```
|
14 |
+
|
15 |
+
|
16 |
+
## Prompt: chat (experimental)
|
17 |
+
```
|
18 |
+
<empty_output>SYSTEM You are a programming assistant
|
19 |
+
<empty_output>USER How do I sort a list in Python?
|
20 |
+
<empty_output>ASSISTANT
|
21 |
+
```
|
22 |
+
|
23 |
+
|
24 |
+
## Example `llama.cpp` command
|
25 |
+
```shell
|
26 |
+
./main -m refact-1_6b-Q4_K_M.gguf -c 4096 -n -1 -p '<fim_prefix>{prefix}<fim_suffix>{suffix}<fim_middle>'
|
27 |
+
```
|
28 |
+
For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md)
|