Text Generation
Transformers
PyTorch
llama
text-generation-inference
Inference Endpoints
soujanyaporia commited on
Commit
188dbb2
·
1 Parent(s): a41ef79

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -2
README.md CHANGED
@@ -8,7 +8,7 @@ metrics:
8
 
9
  # Flacuna: A Vicuna made of Flan
10
 
11
- <img src="https://huggingface.co/declare-lab/flacuna-13b-v1.0/blob/main/flacuna5.png" alt="Image" width="200" height="335">
12
 
13
  Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is already an excellent writing assistant, and the intention behind Flacuna was to enhance Vicuna's problem-solving capabilities. To achieve this, we curated a dedicated instruction dataset called Flan-mini.
14
 
@@ -26,6 +26,7 @@ Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instru
26
  | ShareGPT | ChatGPT | 60K |
27
  | Total | - | 1.34M |
28
 
 
29
 
30
  As a result of this fine-tuning process, Flacuna exhibited notable performance improvements in problem-solving across multiple benchmark datasets, both in few-shot and zero-shot settings.
31
 
@@ -42,4 +43,17 @@ As a result of this fine-tuning process, Flacuna exhibited notable performance i
42
  | Flacuna | 13B | 49.4 | 32.5 | 67.9 |
43
 
44
 
45
- During training, Flacuna is a 13B checkpoint of LLaMA and employed a maximum input sequence length of 1280. We utilized LoRA for parameter-efficient fine-tuning.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
 
9
  # Flacuna: A Vicuna made of Flan
10
 
11
+ <img src="" alt="Image" width="200" height="335">
12
 
13
  Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is already an excellent writing assistant, and the intention behind Flacuna was to enhance Vicuna's problem-solving capabilities. To achieve this, we curated a dedicated instruction dataset called Flan-mini.
14
 
 
26
  | ShareGPT | ChatGPT | 60K |
27
  | Total | - | 1.34M |
28
 
29
+ ## Problem Solving Ability
30
 
31
  As a result of this fine-tuning process, Flacuna exhibited notable performance improvements in problem-solving across multiple benchmark datasets, both in few-shot and zero-shot settings.
32
 
 
43
  | Flacuna | 13B | 49.4 | 32.5 | 67.9 |
44
 
45
 
46
+ During training, Flacuna is a 13B checkpoint of LLaMA and employed a maximum input sequence length of 1280. We utilized LoRA for parameter-efficient fine-tuning.
47
+
48
+ ## Chatbot / Writing Assistant
49
+
50
+ While Flacuna primarily excels in problem-solving tasks, we made efforts to maintain the impressive writing and chatting ability of Vicuna. To achieve this, we incorporated conversational datasets generated by GPT-4, such as GPT-4-Alpaca and ShareGPT, into the Flan-mini collection.
51
+ To use Flacuna as a chatbot or writing assistant, we recommend you use the following template:
52
+
53
+ ```
54
+ A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {definition of the task}./n/n
55
+ {question}/n
56
+ Output: ASSISTANT:
57
+
58
+ ```
59
+ Please note that we still recommend using Vicuna as your preferred Chatbot or Writing Assistant, over Flacuna. Flacuna's primary strength lies in problem-solving tasks, making it ideal for such applications.