artificialguybr commited on
Commit
65e7cc3
1 Parent(s): ae62d4f

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +79 -0
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen2-1.5B
5
+ language:
6
+ - en
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - generated_from_trainer
10
+ - instruction-tuning
11
+ model-index:
12
+ - name: outputs/qwen2.5-1.5b-ft-synthia15-ii
13
+ results: []
14
+ ---
15
+
16
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
17
+
18
+ # Qwen2-1.5B Fine-tuned on Synthia v1.5-II
19
+
20
+ A special thanks to Redmond.ai for sponsoring the GPU resources for this fine-tuning process.
21
+
22
+ This model is a fine-tuned version of [Qwen/Qwen2-1.5B](https://huggingface.co/Qwen/Qwen2-1.5B) on the Synthia v1.5-II dataset, which contains over 20.7k instruction-following examples.
23
+
24
+ ## Model Description
25
+
26
+ Qwen2-1.5B is part of the latest Qwen2 series of large language models. The base model brings significant improvements in:
27
+ - Language understanding and generation
28
+ - Structured data processing
29
+ - Support for multiple languages
30
+ - Long context handling
31
+
32
+ This fine-tuned version enhances the base model's instruction-following capabilities through training on the Synthia v1.5-II dataset.
33
+
34
+ ### Model Architecture
35
+ - Type: Causal Language Model
36
+ - Parameters: 1.5B
37
+ - Training Framework: Transformers 4.45.0.dev0
38
+
39
+ ## Intended Uses & Limitations
40
+
41
+ This model is intended for:
42
+ - Instruction following and task completion
43
+ - Text generation and completion
44
+ - Conversational AI applications
45
+
46
+ The model inherits the capabilities of the base Qwen2-1.5B model, while being specifically tuned for instruction following.
47
+
48
+ ## Training Procedure
49
+
50
+ ### Training Data
51
+ The model was fine-tuned on the Synthia v1.5-II dataset containing 20.7k instruction-following examples.
52
+
53
+ ### Training Hyperparameters
54
+
55
+ The following hyperparameters were used during training:
56
+ - Learning rate: 1e-05
57
+ - Train batch size: 5
58
+ - Eval batch size: 5
59
+ - Seed: 42
60
+ - Gradient accumulation steps: 8
61
+ - Total train batch size: 40
62
+ - Optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
+ - LR scheduler type: cosine
64
+ - LR scheduler warmup steps: 100
65
+ - Number of epochs: 3
66
+ - Sequence length: 4096
67
+ - Sample packing: enabled
68
+ - Pad to sequence length: enabled
69
+
70
+ ## Framework Versions
71
+
72
+ - Transformers 4.45.0.dev0
73
+ - Pytorch 2.3.1+cu121
74
+ - Datasets 2.21.0
75
+ - Tokenizers 0.19.1
76
+
77
+ <details><summary>See axolotl config</summary>
78
+
79
+ axolotl version: `0.4.1`