Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
license:
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
- instruction fine-tuning
|
@@ -18,6 +18,10 @@ widget:
|
|
18 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
19 |
should probably proofread and complete it, then remove this comment. -->
|
20 |
|
|
|
|
|
|
|
|
|
21 |
# LaMini-FLAN-T5-77M
|
22 |
|
23 |
[![Model License](https://img.shields.io/badge/Model%20License-CC%20By%20NC%204.0-red.svg)]()
|
@@ -104,6 +108,11 @@ print("Response": generated_text)
|
|
104 |
```
|
105 |
|
106 |
## Training Procedure
|
|
|
|
|
|
|
|
|
|
|
107 |
We initialize with [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) and fine-tune it on our [LaMini dataset](https://huggingface.co/datasets/MBZUAI/LaMini-instruction). Its total number of parameters is 77M.
|
108 |
|
109 |
### Training Hyperparameters
|
|
|
1 |
---
|
2 |
+
license: cc-by-nc-4.0
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
- instruction fine-tuning
|
|
|
18 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
19 |
should probably proofread and complete it, then remove this comment. -->
|
20 |
|
21 |
+
<p align="center" width="100%">
|
22 |
+
<a><img src="https://raw.githubusercontent.com/mbzuai-nlp/lamini/main/images/LaMnin.png" alt="Title" style="width: 100%; min-width: 300px; display: block; margin: auto;"></a>
|
23 |
+
</p>
|
24 |
+
|
25 |
# LaMini-FLAN-T5-77M
|
26 |
|
27 |
[![Model License](https://img.shields.io/badge/Model%20License-CC%20By%20NC%204.0-red.svg)]()
|
|
|
108 |
```
|
109 |
|
110 |
## Training Procedure
|
111 |
+
|
112 |
+
<p align="center" width="70%">
|
113 |
+
<a><img src="https://raw.githubusercontent.com/mbzuai-nlp/lamini/main/images/lamini-pipeline.drawio.png" alt="Title" style="width: 70%; min-width: 250px; display: block; margin: auto;"></a>
|
114 |
+
</p>
|
115 |
+
|
116 |
We initialize with [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) and fine-tune it on our [LaMini dataset](https://huggingface.co/datasets/MBZUAI/LaMini-instruction). Its total number of parameters is 77M.
|
117 |
|
118 |
### Training Hyperparameters
|