amornpan commited on
Commit
36f7fdb
1 Parent(s): 2f10305

Upload Qwen2ForCausalLM

Browse files
README.md CHANGED
@@ -1,5 +1,14 @@
1
  ---
 
 
 
 
 
 
 
2
  library_name: transformers
 
 
3
  tags:
4
  - text-generation-inference
5
  - sft
@@ -15,15 +24,6 @@ tags:
15
  - lahnmah
16
  - NT Thai GPT
17
  - ntthaigpt
18
- datasets:
19
- - Thaweewat/thai-med-pack
20
- language:
21
- - th
22
- - en
23
- base_model:
24
- - openthaigpt/openthaigpt1.5-7b-instruct
25
- pipeline_tag: text-generation
26
- license: apache-2.0
27
  new_version: Aekanun/openthaigpt-MedChatModelv5.1
28
  ---
29
 
 
1
  ---
2
+ base_model:
3
+ - openthaigpt/openthaigpt1.5-7b-instruct
4
+ datasets:
5
+ - Thaweewat/thai-med-pack
6
+ language:
7
+ - th
8
+ - en
9
  library_name: transformers
10
+ license: apache-2.0
11
+ pipeline_tag: text-generation
12
  tags:
13
  - text-generation-inference
14
  - sft
 
24
  - lahnmah
25
  - NT Thai GPT
26
  - ntthaigpt
 
 
 
 
 
 
 
 
 
27
  new_version: Aekanun/openthaigpt-MedChatModelv5.1
28
  ---
29
 
adapter_config.json CHANGED
@@ -11,18 +11,20 @@
11
  "layers_pattern": null,
12
  "layers_to_transform": null,
13
  "loftq_config": {},
14
- "lora_alpha": 32,
15
- "lora_dropout": 0.05,
16
  "megatron_config": null,
17
  "megatron_core": "megatron.core",
18
  "modules_to_save": null,
19
  "peft_type": "LORA",
20
- "r": 16,
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
- "v_proj",
25
- "q_proj"
 
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
11
  "layers_pattern": null,
12
  "layers_to_transform": null,
13
  "loftq_config": {},
14
+ "lora_alpha": 16,
15
+ "lora_dropout": 0.7,
16
  "megatron_config": null,
17
  "megatron_core": "megatron.core",
18
  "modules_to_save": null,
19
  "peft_type": "LORA",
20
+ "r": 32,
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
+ "self_attn.q_proj",
25
+ "self_attn.o_proj",
26
+ "self_attn.v_proj",
27
+ "self_attn.k_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53beddb9efb9d0e52aeac610a8a9d71fee92541d34e18a2a13faf2c7c9bdf019
3
- size 20200056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c49279dfbd538e0234ae760cfa5179f32ab7a9c3562ff37ffd453f88388b76ab
3
+ size 80770432
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.45.2"
6
+ }