Safetensors
Aiwensile2 commited on
Commit
a7dce89
1 Parent(s): c14b4f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -3
README.md CHANGED
@@ -1,3 +1,46 @@
1
- ---
2
- license: cc-by-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ datasets:
4
+ - Aiwensile2/Minecraft_QA-pairs_Instruction_Dataset
5
+ ---
6
+ # MineMA: Fine-Tuned Models for Minecraft Q&A
7
+
8
+ ## Overview
9
+
10
+ In this repository, we present the MineMA series of models, fine-tuned specifically for Minecraft-related Q&A tasks. Utilizing the LoRA method for efficient model fine-tuning, we have adapted pre-trained LLaMA models to respond accurately and effectively to Minecraft-related instructions and queries. Our fine-tuning process leverages the specially generated Minecraft dataset to ensure relevance and accuracy in the Q&A responses.
11
+
12
+ ## Models
13
+
14
+ The MineMA series includes several models fine-tuned on different base models from the LLaMA series. Below is the list of the fine-tuned models provided in this repository:
15
+
16
+ - **MineMA-8B**(v1, v2, v3, v4), derived from the base model LLaMA-3-8B-Instruct.
17
+ - **MineMA-13B**(v1, v2), derived from the base model LLaMA-2-13B-Chat.
18
+ - **MineMA-70B**, derived from the base model LLaMA-3-70B-Instruct.
19
+
20
+ These models have been fine-tuned by using the **Minecraft_QA-pairs_Instruction_Dataset**. We have only released four models of MineMA-8B for the time being, and we will supplement more models in the future.
21
+
22
+ ## Fine-Tuning Methodology
23
+
24
+ ### LoRA Method for Fine-Tuning
25
+
26
+ We employed the **LoRA (Low-Rank Adaptation)** method for fine-tuning our models. LoRA is a parameter-efficient training technique that introduces small, trainable low-rank matrices to adapt a pre-trained neural network, allowing for targeted updates without the need for retraining the entire model. This method strikes a balance between computational efficiency and training effectiveness.
27
+
28
+ ### Training Parameters
29
+
30
+ Here are the specific training parameters:
31
+
32
+ | Model | lora\_r | lora\_alpha | lora\_dropout | learning\_rate | weight\_decay | Single Round? |
33
+ |--------------|---------|-------------|---------------|----------------|---------------|---------------|
34
+ | MineMA-13B-v1| 64 | 128 | 0.1 | 1E-04 | 1E-04 | False |
35
+ | MineMA-13B-v2| 128 | 256 | 0.1 | 1E-04 | 1E-04 | False |
36
+ | MineMA-8B-v1 | 64 | 128 | 0.1 | 1E-04 | 1E-04 | True |
37
+ | MineMA-8B-v2 | 32 | 64 | 0.1 | 1E-04 | 1E-04 | False |
38
+ | MineMA-8B-v3 | 64 | 128 | 0.1 | 1E-04 | 1E-04 | False |
39
+ | MineMA-8B-v4 | 128 | 256 | 0.1 | 1E-04 | 1E-04 | False |
40
+ | MineMA-70B | 16 | 32 | 0.1 | 1E-04 | 1E-04 | True |
41
+
42
+ ## Dataset
43
+
44
+ We used the **Minecraft_QA-pairs_Instruction_Dataset** for fine-tuning all the models in the MineMA series. This dataset has 390,317 instruction entries specifically designed for Minecraft-related Q&A tasks. You can access the dataset via the following link:
45
+
46
+ [Minecraft_QA-pairs_Instruction_Dataset](https://huggingface.co/datasets/Aiwensile2/Minecraft_QA-pairs_Instruction_Dataset)