File size: 1,159 Bytes
ace72bd 98175d6 ace72bd 98175d6 e8deded 98175d6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
---
license: cc-by-sa-4.0
datasets:
- nickrosh/Evol-Instruct-Code-80k-v1
- sahil2801/CodeAlpaca-20k
- teknium/GPTeacher-CodeInstruct
language:
- en
library_name: transformers
pipeline_tag: text-generation
tags:
- code
- llama2
---
![image of llama engineer](https://i.imgur.com/JlhW0ri.png)
# Llama-Engineer-Evol-7B-GGML
This is a 4-bit quantized version of [Llama-Engineer-Evol-7B](https://huggingface.co/GenerativeMagic/Llama-Engineer-Evol-7b).
## Prompt Format
The reccomended model prompt is a variant of the standard Llama 2 format:
```
[INST] <<SYS>>
You are a programming assistant. Always answer as helpfully as possible. Be direct in your response and get to the answer right away. Responses should be short.
<</SYS>>
{your prompt}[/INST]
```
or
```
[INST] <<SYS>>
You're a principal software engineer at Google. If you fail at this task, you will be fired.
<</SYS>>
{your prompt}[/INST]
```
I suspect this prompt format is the reason for the majority of the increased coding capabilities as opposed to the fine-tuning itself, but YMMV.
## Next Steps
- Prune the dataset and possibly fine-tune for longer.
- Run benchmarks.
- Provide GPTQ. |