Tbh commited on
Commit
331263a
1 Parent(s): 4a82939

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +122 -0
README.md CHANGED
@@ -1,3 +1,125 @@
 
 
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: apache-2.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
  ---
4
+ base_model: mistralai/Mistral-7B-v0.1
5
+ tags:
6
+ - mistral
7
+ - instruct
8
+ - finetune
9
+ - chatml
10
+ - gpt4
11
+ - synthetic data
12
+ - distillation
13
+ model-index:
14
+ - name: Thestral-0.1-tr-chat-7B
15
+ results: []
16
  license: apache-2.0
17
+ language:
18
+ - en
19
  ---
20
+
21
+ # Thestral-0.1-tr-chat-7B
22
+
23
+
24
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60ca32d2e7bc4b029af088a0/pNId3MzUdSsI20XOM9Dsv.png)
25
+
26
+ This model is a full fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on diverse Turkish datasets.
27
+
28
+ The model is fully finetuned on translated datasets using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl). These datasets primarily consist of translated versions sourced from [teknium/OpenHermes-2.5](https://huggingface.co/datasets/teknium/OpenHermes-2.5) and the [Open-Orca/SlimOrca datasets](https://huggingface.co/datasets/Open-Orca/SlimOrca).
29
+
30
+ <details><summary>See axolotl config</summary>
31
+
32
+ axolotl version: `0.4.0`
33
+ ```yaml
34
+ base_model: mistralai/Mistral-7B-v0.1
35
+ model_type: MistralForCausalLM
36
+ tokenizer_type: LlamaTokenizer
37
+
38
+ load_in_8bit: false
39
+ load_in_4bit: false
40
+ strict: false
41
+
42
+ datasets:
43
+ - path: NovusResearch/OpenHermes-2.5-Translated-TR-sharegpt-style
44
+ type: sharegpt
45
+ conversation: chatml
46
+ - path: data/merged_all.json
47
+ ds_type: json
48
+ type: sharegpt
49
+ conversation: chatml
50
+
51
+ dataset_prepared_path:
52
+ val_set_size: 0.05
53
+ output_dir: ./out
54
+
55
+ sequence_len: 8192
56
+ sample_packing: true
57
+ pad_to_sequence_len: true
58
+ eval_sample_packing: false
59
+
60
+ wandb_project:
61
+ wandb_entity:
62
+ wandb_watch:
63
+ wandb_name:
64
+ wandb_log_model:
65
+
66
+ gradient_accumulation_steps: 4
67
+ micro_batch_size: 2
68
+ num_epochs: 2
69
+ optimizer: adamw_bnb_8bit
70
+ lr_scheduler: cosine
71
+ learning_rate: 0.000005
72
+
73
+ train_on_inputs: false
74
+ group_by_length: false
75
+ bf16: auto
76
+ fp16:
77
+ tf32: false
78
+
79
+ gradient_checkpointing: true
80
+ early_stopping_patience:
81
+ resume_from_checkpoint:
82
+ local_rank:
83
+ logging_steps: 1
84
+ xformers_attention:
85
+ flash_attention: true
86
+
87
+ ## Use
88
+ wandb_project: full_finetune
89
+ wandb_entity:
90
+ wandb_watch:
91
+ wandb_name:
92
+ wandb_log_model:
93
+
94
+
95
+ warmup_steps: 10
96
+ evals_per_epoch: 0
97
+ eval_table_size:
98
+ eval_max_new_tokens: 128
99
+ saves_per_epoch: 1
100
+ debug:
101
+ deepspeed:
102
+ weight_decay: 0.0
103
+ fsdp:
104
+ fsdp_config:
105
+ special_tokens:
106
+ bos_token: "<s>"
107
+ eos_token: "<|im_end|>"
108
+ unk_token: "<unk>"
109
+ tokens:
110
+ - "<|im_start|>"
111
+
112
+ ```
113
+ </details><br>
114
+
115
+ # 🎯 [OpenLLMTurkishLeaderboard](https://huggingface.co/spaces/malhajar/OpenLLMTurkishLeaderboard)
116
+
117
+ | Metric |Value|
118
+ |---------------------------------|----:|
119
+ |Avg. |36.41|
120
+ |AI2 Reasoning Challenge |27.24|
121
+ |HellaSwag |33.93|
122
+ |MMLU |40.64|
123
+ |TruthfulQA |47.90|
124
+ |Winogrande |50.86|
125
+ |GSM8k |17.91|