Text Generation
GGUF
English
code
Inference Endpoints
IDDT commited on
Commit
9298d7d
1 Parent(s): e9a1ace

Update readme.

Browse files
Files changed (1) hide show
  1. README.md +28 -3
README.md CHANGED
@@ -1,3 +1,28 @@
1
- ---
2
- license: bigscience-openrail-m
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Refact-1.6B-fim-GGUF
2
+ - Model creator: [Small Magellanic Cloud AI](https://huggingface.co/smallcloudai)
3
+ - Original model: [Refact-1.6B](https://huggingface.co/smallcloudai/Refact-1_6B-fim)
4
+
5
+
6
+ ## Description
7
+ This repository contains quantized GGUF format model files for [Refact-1.6B](https://huggingface.co/smallcloudai/Refact-1_6B-fim).
8
+
9
+
10
+ ## Prompt: fill in the middle
11
+ ```
12
+ <fim_prefix>def print_hello_world():\n """<fim_suffix>\n print("Hello world!")<fim_middle>
13
+ ```
14
+
15
+
16
+ ## Prompt: chat (experimental)
17
+ ```
18
+ <empty_output>SYSTEM You are a programming assistant
19
+ <empty_output>USER How do I sort a list in Python?
20
+ <empty_output>ASSISTANT
21
+ ```
22
+
23
+
24
+ ## Example `llama.cpp` command
25
+ ```shell
26
+ ./main -m refact-1_6b-Q4_K_M.gguf -c 4096 -n -1 -p '<fim_prefix>{prefix}<fim_suffix>{suffix}<fim_middle>'
27
+ ```
28
+ For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md)