rozek's picture
Update README.md
d6608db
|
raw
history blame
1.1 kB
---
license: apache-2.0
datasets:
- togethercomputer/RedPajama-Data-1T
tags:
- open-llama
- ggml
- 7b
- quantized
- text-generation
---
This repository contains the [300BT preview](https://huggingface.co/openlm-research/open_llama_7b_preview_300bt) of the
[Open LLaMA](https://github.com/openlm-research/open_llama) model based on the
[RedPajama 1T dataset](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T),
converted to GGML and quantized to 4 bit, ready to be used with [llama.cpp](https://github.com/rozek/llama.cpp)
## Usage ##
In order to use this model with llama.cpp
* install llama.cpp as [described in the docs](https://github.com/ggerganov/llama.cpp#usage)
* download this model
* move it into the `models` subfolder of llama.cpp
* run inferences with the additional parameter `-m ./models/7B/ggml-openllama-7b-300bt-q4_0.bin`
For example,
```
./main -m ./models/7B/ggml-openllama-7b-300bt-q4_0.bin -n 128 -p "who was Joseph Weizenbaum?"
```
## License ##
Since the underlying model is itself Apache 2.0 licensed, this repo is as well.
---
license: apache-2.0
---