soujanyaporia
commited on
Commit
·
188dbb2
1
Parent(s):
a41ef79
Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ metrics:
|
|
8 |
|
9 |
# Flacuna: A Vicuna made of Flan
|
10 |
|
11 |
-
<img src="
|
12 |
|
13 |
Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is already an excellent writing assistant, and the intention behind Flacuna was to enhance Vicuna's problem-solving capabilities. To achieve this, we curated a dedicated instruction dataset called Flan-mini.
|
14 |
|
@@ -26,6 +26,7 @@ Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instru
|
|
26 |
| ShareGPT | ChatGPT | 60K |
|
27 |
| Total | - | 1.34M |
|
28 |
|
|
|
29 |
|
30 |
As a result of this fine-tuning process, Flacuna exhibited notable performance improvements in problem-solving across multiple benchmark datasets, both in few-shot and zero-shot settings.
|
31 |
|
@@ -42,4 +43,17 @@ As a result of this fine-tuning process, Flacuna exhibited notable performance i
|
|
42 |
| Flacuna | 13B | 49.4 | 32.5 | 67.9 |
|
43 |
|
44 |
|
45 |
-
During training, Flacuna is a 13B checkpoint of LLaMA and employed a maximum input sequence length of 1280. We utilized LoRA for parameter-efficient fine-tuning.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
# Flacuna: A Vicuna made of Flan
|
10 |
|
11 |
+
<img src="" alt="Image" width="200" height="335">
|
12 |
|
13 |
Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is already an excellent writing assistant, and the intention behind Flacuna was to enhance Vicuna's problem-solving capabilities. To achieve this, we curated a dedicated instruction dataset called Flan-mini.
|
14 |
|
|
|
26 |
| ShareGPT | ChatGPT | 60K |
|
27 |
| Total | - | 1.34M |
|
28 |
|
29 |
+
## Problem Solving Ability
|
30 |
|
31 |
As a result of this fine-tuning process, Flacuna exhibited notable performance improvements in problem-solving across multiple benchmark datasets, both in few-shot and zero-shot settings.
|
32 |
|
|
|
43 |
| Flacuna | 13B | 49.4 | 32.5 | 67.9 |
|
44 |
|
45 |
|
46 |
+
During training, Flacuna is a 13B checkpoint of LLaMA and employed a maximum input sequence length of 1280. We utilized LoRA for parameter-efficient fine-tuning.
|
47 |
+
|
48 |
+
## Chatbot / Writing Assistant
|
49 |
+
|
50 |
+
While Flacuna primarily excels in problem-solving tasks, we made efforts to maintain the impressive writing and chatting ability of Vicuna. To achieve this, we incorporated conversational datasets generated by GPT-4, such as GPT-4-Alpaca and ShareGPT, into the Flan-mini collection.
|
51 |
+
To use Flacuna as a chatbot or writing assistant, we recommend you use the following template:
|
52 |
+
|
53 |
+
```
|
54 |
+
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {definition of the task}./n/n
|
55 |
+
{question}/n
|
56 |
+
Output: ASSISTANT:
|
57 |
+
|
58 |
+
```
|
59 |
+
Please note that we still recommend using Vicuna as your preferred Chatbot or Writing Assistant, over Flacuna. Flacuna's primary strength lies in problem-solving tasks, making it ideal for such applications.
|