TheBloke's LLM work is generously supported by a grant from andreessen horowitz (a16z)
Kaist AI's Selfee 13B GPTQ
These files are GPTQ 4bit model files for Kaist AI's Selfee 13B.
It is the result of quantising to 4bit using GPTQ-for-LLaMa.
Other repositories available
- 4-bit GPTQ models for GPU inference
- 4-bit, 5-bit and 8-bit GGML models for CPU(+GPU) inference
- Unquantised fp16 model in pytorch format, for GPU inference and for further conversions
How to easily download and use this model in text-generation-webui
Downloading the model
- Click the Model tab.
- Under Download custom model or LoRA, enter
TheBloke/Selfee-13B-GPTQ
. - Click Download.
- Wait until it says it's finished downloading.
- Untick "Autoload model"
- Click the Refresh icon next to Model in the top left.
To use with AutoGPTQ (if installed)
- In the Model drop-down: choose the model you just downloaded,
Selfee-13B-GPTQ
. - Under GPTQ, tick AutoGPTQ.
- Click Save settings for this model in the top right.
- Click Reload the Model in the top right.
- Once it says it's loaded, click the Text Generation tab and enter a prompt!
To use with GPTQ-for-LLaMa
- In the Model drop-down: choose the model you just downloaded,
Selfee-13B-GPTQ
. - If you see an error in the bottom right, ignore it - it's temporary.
- Fill out the
GPTQ parameters
on the right:Bits = 4
,Groupsize = 128
,model_type = Llama
- Click Save settings for this model in the top right.
- Click Reload the Model in the top right.
- Once it says it's loaded, click the Text Generation tab and enter a prompt!
Provided files
selfee-13b-GPTQ-4bit-128g.no-act.order.safetensors
This will work with all versions of GPTQ-for-LLaMa, and with AutoGPTQ.
It was created with group_size 128 to increase inference accuracy, but without --act-order (desc_act) to increase compatibility and improve inference speed.
selfee-13b-GPTQ-4bit-128g.no-act.order.safetensors
- Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
- Works with AutoGPTQ
- Works with text-generation-webui one-click-installers
- Parameters: Groupsize = 128. Act Order / desc_act = False.
Discord
For further support, and discussions on these models and AI in general, join us at:
Thanks, and how to contribute.
Thanks to the chirper.ai team!
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
- Patreon: https://patreon.com/TheBlokeAI
- Ko-Fi: https://ko-fi.com/TheBlokeAI
Special thanks to: Aemon Algiz.
Patreon special mentions: Sam, theTransient, Jonathan Leane, Steven Wood, webtim, Johann-Peter Hartmann, Geoffrey Montalvo, Gabriel Tamborski, Willem Michiel, John Villwock, Derek Yates, Mesiah Bishop, Eugene Pentland, Pieter, Chadd, Stephen Murray, Daniel P. Andersen, terasurfer, Brandon Frisco, Thomas Belote, Sid, Nathan LeClaire, Magnesian, Alps Aficionado, Stanislav Ovsiannikov, Alex, Joseph William Delisle, Nikolai Manek, Michael Davis, Junyu Yang, K, J, Spencer Kim, Stefan Sabev, Olusegun Samson, transmissions 11, Michael Levine, Cory Kujawski, Rainer Wilmers, zynix, Kalila, Luke @flexchar, Ajan Kanaga, Mandus, vamX, Ai Maven, Mano Prime, Matthew Berman, subjectnull, Vitor Caleffi, Clay Pascal, biorpg, alfie_i, 阿明, Jeffrey Morgan, ya boyyy, Raymond Fosdick, knownsqashed, Olakabola, Leonard Tan, ReadyPlayerEmma, Enrico Ros, Dave, Talal Aujan, Illia Dulskyi, Sean Connelly, senxiiz, Artur Olbinski, Elle, Raven Klaugh, Fen Risland, Deep Realms, Imad Khwaja, Fred von Graf, Will Dee, usrbinkat, SuperWojo, Alexandros Triantafyllidis, Swaroop Kallakuri, Dan Guido, John Detwiler, Pedro Madruga, Iucharbius, Viktor Bowallius, Asp the Wyvern, Edmond Seymore, Trenton Dambrowitz, Space Cruiser, Spiking Neurons AB, Pyrater, LangChain4j, Tony Hughes, Kacper Wikieł, Rishabh Srivastava, David Ziegler, Luke Pendergrass, Andrey, Gabriel Puliatti, Lone Striker, Sebastain Graf, Pierre Kircher, Randy H, NimbleBox.ai, Vadim, danny, Deo Leter
Thank you to all my generous patrons and donaters!
And thank you again to a16z for their generous grant.
Original model card: Kaist AI's Selfee 13B
SelFee: Iterative Self-Revising LLM Empowered by
Self-Feedback Generation
News
[May 31, 2023] Initial release: We released the first version of SelFee! Check out the blog post for more details.
Overview
This is the repository for the KAIST SelFee project, which aims to build and share an instruction-following LLaMA model. This repo mainly has five contents:
- The selection process of the 178K training data for SelFee (detail, code).
- The generation process for the training data and its result. (detail, code).
- The training process for the model (detail, code).
- The inference process for the model (detail, code).
- The evaluation method and dataset (detail, code).
This repository is based on the Stanford-Alpaca and Vicuna repository. Thanks to all the contributors for these awesome repositories!! 🙌
We highly recommend you read our blog post for more details about the model.
Data Release
For data collection, we collected datasets from five different fields. These are the Stanford Alpaca dataset, math collection, code collection, Flan collection, and ShareGPT. We provide code that we used to make a dataset for training. We also provide code how we preprocessed ShareGPT. For ShareGPT, we only use the first (question, answer) pair from human and GPT, respectively. We only use instances which are classified as english,and filter instance which is not a form of question. For other datsets, we do not need special data collection method.
Data Generation Process
To train our model with high-quality instructions and answer pairs, we utilized data augmentation using OpenAI API calls. The process involved three steps.
Firstly, we collected various instructions from multiple fields and fed them to ChatGPT to generate answers.
Secondly, we gathered feedback on the generated answer by querying ChatGPT again and asked it to determine if the initial answer required any revision.
Thirdly, if a revision was necessary, we passed the instruction, initial answer, and feedback pair to ChatGPT to generate a revised answer and its feedback pair.
We repeated the process until we received feedback that required no further revision or hit the maximum iteration. However, due to the token limitation of the ChatGPT API, we had to truncate some instances that needed more than 4096 tokens while augmenting.
You can see the details with command here.
*We provide the whole dataset after collection and augmentation using huggingface(code), so you can either use the code or follow our data merging step to replicate the training dataset. Feel free to use any of them!
Training
We utilize FastChat to train the model. Given the instruction, we fine-tune the model to generate the answer and feedback chain (including the revisions).
To reproduce the training procedure, here are the steps.
pip install -r requirements.txt
torchrun --nproc_per_node=4 train/train_mem.py \
--model_name_or_path llama-7b \
--data_path outputs/feedback_gpt_3.5_turbo_merged_whole.json \
--bf16 True \
--output_dir ckpt/selfee-7b \
--num_train_epochs 3 \
--per_device_train_batch_size 16 \
--per_device_eval_batch_size 16 \
--gradient_accumulation_steps 2 \
--evaluation_strategy "no" \
--save_strategy "steps" \
--save_steps 5000 \
--save_total_limit 1 \
--learning_rate 2e-5 \
--weight_decay 0. \
--warmup_ratio 0.03 \
--lr_scheduler_type "cosine" \
--logging_steps 1 \
--fsdp "shard_grad_op auto_wrap" \
--fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' \
--tf32 True \
--model_max_length 2048 \
--gradient_checkpointing True \
--lazy_preprocess True \
--training_objective full \
The hyperparameters are as follows, following Vicuna and Alpaca.
Hyperparameter | Global Batch Size | Learning rate | Epochs | Max length | Weight decay |
---|---|---|---|---|---|
SelFee (7B, 13B) | 128 | 2e-5 | 3 | 2048 | 0 |
Inference
Restoring checkpoint using diff
We provide diff weight and code which can restore the same model with SelFee. To restore the original SelFee weight, you first need to convert the Meta's original LLAMA checkpoint into huggingface format into your local machine. Once you are done, you can restore the same checkpoint of our model by using the following command
python inference/apply_delta.py --path_raw {path_to_llama_7b} --path_tuned /ckpt/selfee-7b --path_diff kaist-ai/selfee-7b-delta
Autonomous Inference Mode
Because SelFee is trained to generate iterative feedback and revisions until the response is satisfying, it automatically generates iterative feedback and revisions on a single forward pass. The model autonomously decides when to stop generating revisions based on the feedback. If the feedback chain ends with sequences like Revision is not needed.
, the model autonomously terminates generation.
For autonomous inference mode,
python inference/inference.py --model-path "ckpt/selfee-7b" --model-id "selfee" --question-file "evaluation/template/question.jsonl" --answer-file "evaluation/answer/selfee_7b_autonomous.jsonl"
Revision Enforce Inference Mode
We observed that increasing the minimum number of required revisions corresponds to a corresponding increase in performance. To enforce revisions, we automatically replace sequences such as Revision is not needed.
into Revision is needed.
during self-feedback generation. Because SelFee is trained to generate Revision {index}:
after the sequence of Revision is needed.
, the model would continually revise the answer.
For revision enforce inference mode, use the max-num-revision
argument.
python inference/inference.py --model-path "ckpt/selfee-7b" --model-id "selfee" --question-file "evaluation/template/question.jsonl" --answer-file "evaluation/answer/selfee_7b_enforce_3_revision.jsonl" --max-num-revision 3
Evaluation
Following evaluation setting of Vicuna, we evaluate on 80 diverse queries and utilize GPT-4 language model as the evaluator, scoring a model's response relative to ChatGPT's response. One of the difference with Vicuna evaluation is that due to positional bias of GPT-4, we employ a bidirectional evaluation setting. This means that each evaluation instance is inferred twice, depending on its position.
We release the inference result of SelFee in the folder of evaluation/answer
and also the scores generated by GPT-4 in the folder of evaluation/review
.
GPT-4 Automatic Evaluation
First, you need to get your API key to get access to the GPT-4 API.
export OPENAI_API_KEYS={personal_key}
To compare the performance of a generation result (for example, located on evaluation/answer/file_A.jsonl
) with another generation result (located on evaluation/anwer/file_B.jsonl
),
python evaluation/gpt4_automatic_evaluation.py -q evaluation/template/question.jsonl -a evaluation/answer/file_A.jsonl evaluation/answer/file_B.jsonl -p evaluation/template/prompt.jsonl -r evaluation/template/reviewer.jsonl -o evaluation/review/A_vs_B.jsonl
To mitigate the positional bias of GPT-4 model, we apply a bidirectional evaluation setting. Therefore, automatic evaluation with opposite position is also needed.
python evaluation/gpt4_automatic_evaluation.py -q evaluation/template/question.jsonl -a evaluation/answer/file_B.jsonl evaluation/answer/file_A.jsonl -p evaluation/template/prompt.jsonl -r evaluation/template/reviewer.jsonl -o evaluation/review/B_vs_A.jsonl
Limitations
Similar to other LLaMA-finetuned models, SelFee also make some mistakes especially for math, reasoning, factuality, and coding tasks. Although our performance outperforms ChatGPT on Vicuna setting, the evaluation setting contains some limitations in terms of comprehension (limited to 80 queries), inconsistency, and unreliability. Therefore, further research for a better evaluation setting is needed. Please take these claims with a grain of salt.
Online demo
Check out the demo!
How to launch the demo yourself
To serve the web demo yourself, run the following commands:
- Run the controller
python3 -m serve.controller
- Run the model worker
python3 -m serve.model_worker --model-path $MODEL_PATH --port 21002 --worker-address=http://localhost:21002 --model-name=SelFee-13b
- Run the web server
python3 -m serve.gradio_web_server --share
You can find the serving code here.
Team members
Seonghyeon Ye*, Yongrae Jo*, Doyoung Kim*, Sungdong Kim, Hyeonbin Hwang, and Minjoon Seo.
(* denotes equal contribution)
Release
We have released the SelFee-7B and SelFee-13B model diff weights, which can be found with instructions here. Moreover, the training instances used to train SelFee is released on huggingface.
License
The research preview online demo is only for non-commercial use and is subject to various licenses and terms of use, including the LLaMA model License, OpenAI's Terms of Use for the generated data, and ShareGPT's Privacy Practices. If you suspect any violations, please reach out to us.
Citation
Please cite if you use the data or code in this repo.
@misc{selfee2023,
author = {Ye, Seonghyeon and Jo, Yongrae and Kim, Doyoung and Kim, Sungdong and Hwang, Hyeonbin and Seo, Minjoon},
title = {SelFee: Iterative Self-Revising LLM Empowered by Self-Feedback Generation},
url = {https://kaistai.github.io/SelFee/},
month = {May},
year = {2023},
howpublished = {Blog post}
}
- Downloads last month
- 15