Aiwensile2
commited on
Commit
•
a7dce89
1
Parent(s):
c14b4f9
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,46 @@
|
|
1 |
-
---
|
2 |
-
license: cc-by-4.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-4.0
|
3 |
+
datasets:
|
4 |
+
- Aiwensile2/Minecraft_QA-pairs_Instruction_Dataset
|
5 |
+
---
|
6 |
+
# MineMA: Fine-Tuned Models for Minecraft Q&A
|
7 |
+
|
8 |
+
## Overview
|
9 |
+
|
10 |
+
In this repository, we present the MineMA series of models, fine-tuned specifically for Minecraft-related Q&A tasks. Utilizing the LoRA method for efficient model fine-tuning, we have adapted pre-trained LLaMA models to respond accurately and effectively to Minecraft-related instructions and queries. Our fine-tuning process leverages the specially generated Minecraft dataset to ensure relevance and accuracy in the Q&A responses.
|
11 |
+
|
12 |
+
## Models
|
13 |
+
|
14 |
+
The MineMA series includes several models fine-tuned on different base models from the LLaMA series. Below is the list of the fine-tuned models provided in this repository:
|
15 |
+
|
16 |
+
- **MineMA-8B**(v1, v2, v3, v4), derived from the base model LLaMA-3-8B-Instruct.
|
17 |
+
- **MineMA-13B**(v1, v2), derived from the base model LLaMA-2-13B-Chat.
|
18 |
+
- **MineMA-70B**, derived from the base model LLaMA-3-70B-Instruct.
|
19 |
+
|
20 |
+
These models have been fine-tuned by using the **Minecraft_QA-pairs_Instruction_Dataset**. We have only released four models of MineMA-8B for the time being, and we will supplement more models in the future.
|
21 |
+
|
22 |
+
## Fine-Tuning Methodology
|
23 |
+
|
24 |
+
### LoRA Method for Fine-Tuning
|
25 |
+
|
26 |
+
We employed the **LoRA (Low-Rank Adaptation)** method for fine-tuning our models. LoRA is a parameter-efficient training technique that introduces small, trainable low-rank matrices to adapt a pre-trained neural network, allowing for targeted updates without the need for retraining the entire model. This method strikes a balance between computational efficiency and training effectiveness.
|
27 |
+
|
28 |
+
### Training Parameters
|
29 |
+
|
30 |
+
Here are the specific training parameters:
|
31 |
+
|
32 |
+
| Model | lora\_r | lora\_alpha | lora\_dropout | learning\_rate | weight\_decay | Single Round? |
|
33 |
+
|--------------|---------|-------------|---------------|----------------|---------------|---------------|
|
34 |
+
| MineMA-13B-v1| 64 | 128 | 0.1 | 1E-04 | 1E-04 | False |
|
35 |
+
| MineMA-13B-v2| 128 | 256 | 0.1 | 1E-04 | 1E-04 | False |
|
36 |
+
| MineMA-8B-v1 | 64 | 128 | 0.1 | 1E-04 | 1E-04 | True |
|
37 |
+
| MineMA-8B-v2 | 32 | 64 | 0.1 | 1E-04 | 1E-04 | False |
|
38 |
+
| MineMA-8B-v3 | 64 | 128 | 0.1 | 1E-04 | 1E-04 | False |
|
39 |
+
| MineMA-8B-v4 | 128 | 256 | 0.1 | 1E-04 | 1E-04 | False |
|
40 |
+
| MineMA-70B | 16 | 32 | 0.1 | 1E-04 | 1E-04 | True |
|
41 |
+
|
42 |
+
## Dataset
|
43 |
+
|
44 |
+
We used the **Minecraft_QA-pairs_Instruction_Dataset** for fine-tuning all the models in the MineMA series. This dataset has 390,317 instruction entries specifically designed for Minecraft-related Q&A tasks. You can access the dataset via the following link:
|
45 |
+
|
46 |
+
[Minecraft_QA-pairs_Instruction_Dataset](https://huggingface.co/datasets/Aiwensile2/Minecraft_QA-pairs_Instruction_Dataset)
|