File size: 1,587 Bytes
44c1835 3d43edc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 |
---
datasets:
- adamo1139/basic_economics_questions_ts_test_1
---
QLORA on SpicyBoros 2.2 Llama 7B v2 using synthetic Q&A Dataset.
a little bit under one epoch, since my GTX1080 decided to OOM a tiny bit before training end and I am using checkpoint made at 450/465 step.
I've been running into a lot of issues, so I am happy to even get that far, most of my QLORA attempts had loss go to 0 and deepspeed was forcibly closing training after roughly 0.3 epoch.
My intention with this QLORA is mostly to try to train something usable and cool locally on normal desktop without going to runpod.
I tried training q4_0 quant with cpu-lora in llama.cpp (https://rentry.org/cpu-lora) but it's been a miss, it's about 20x slower on 11400f than on poorman's GTX 1080.
The model can be used to ask questions about basic economic concepts, responses will have a viewpoint similar to the one expressed by Thomas Sowell in his book Basic Economics.
Prompt format:
Reader: {prompt}
'\nThomas:\n' {response}
I was training on the sequence length of 1024, but I conversed with the model up to 4000 tokens and it was still coherent and in character.
Even though the training date I used is only single turn, model has no issue with multi-turn conversations. Much of that is thanks to the fine-tuning done earlier by amazing Jon Durbin.
Known issues:
- tokenization didn't happen as I expected, so you can see a lot of /n, \' and ' characters in places where you shouldn't really see them. For example, most responses, if using the right prompt format, will have character ' at the end of response |