Update README.md
Browse files
README.md
CHANGED
@@ -14,13 +14,13 @@ should probably proofread and complete it, then remove this comment. -->
|
|
14 |
|
15 |
# flippa-v2
|
16 |
|
17 |
-
This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on
|
18 |
It achieves the following results on the evaluation set:
|
19 |
- Loss: 0.9289
|
20 |
|
21 |
## Model description
|
22 |
|
23 |
-
|
24 |
|
25 |
## Intended uses & limitations
|
26 |
|
|
|
14 |
|
15 |
# flippa-v2
|
16 |
|
17 |
+
This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on a mixed dataset of filtered non-refusal data, math, and code.
|
18 |
It achieves the following results on the evaluation set:
|
19 |
- Loss: 0.9289
|
20 |
|
21 |
## Model description
|
22 |
|
23 |
+
My second test of experiments using Quantitized LoRA and Mistral-7B-Instruct, trained on A100 in one hour, will increase training times and amount of data as I gain access to more GPUs.
|
24 |
|
25 |
## Intended uses & limitations
|
26 |
|