File size: 1,100 Bytes
a86048e
 
 
 
d6608db
 
 
 
 
 
a86048e
332e2d7
 
 
 
d10f7db
166a81b
 
 
 
63368c6
166a81b
 
 
 
 
 
 
 
 
 
 
 
 
 
1770cba
 
a86048e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
license: apache-2.0
datasets:
- togethercomputer/RedPajama-Data-1T
tags:
- open-llama
- ggml
- 7b
- quantized
- text-generation
---
This repository contains the [300BT preview](https://huggingface.co/openlm-research/open_llama_7b_preview_300bt) of the
[Open LLaMA](https://github.com/openlm-research/open_llama) model based on the
[RedPajama 1T dataset](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T),
converted to GGML and quantized to 4 bit, ready to be used with [llama.cpp](https://github.com/rozek/llama.cpp)

## Usage ##

In order to use this model with llama.cpp

* install llama.cpp as [described in the docs](https://github.com/ggerganov/llama.cpp#usage)
* download this model
* move it into the `models` subfolder of llama.cpp
* run inferences with the additional parameter `-m ./models/7B/ggml-openllama-7b-300bt-q4_0.bin`

For example,

```
./main -m ./models/7B/ggml-openllama-7b-300bt-q4_0.bin -n 128 -p "who was Joseph Weizenbaum?"
```

## License ##

Since the underlying model is itself Apache 2.0 licensed, this repo is as well.

---
license: apache-2.0
---