blai88 commited on
Commit
c6a2310
1 Parent(s): 39ad302

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
- license: other
3
- base_model: facebook/opt-350m
 
4
  tags:
5
  - trl
6
  - reward-trainer
@@ -13,10 +14,10 @@ model-index:
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/blai88/huggingface/runs/vzbyct0h)
17
  # reward_modeling_anthropic_hh
18
 
19
- This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on the None dataset.
20
 
21
  ## Model description
22
 
@@ -36,11 +37,9 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 1.41e-05
39
- - train_batch_size: 16
40
  - eval_batch_size: 8
41
  - seed: 42
42
- - gradient_accumulation_steps: 2
43
- - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - num_epochs: 1.0
@@ -51,7 +50,8 @@ The following hyperparameters were used during training:
51
 
52
  ### Framework versions
53
 
 
54
  - Transformers 4.42.3
55
  - Pytorch 2.3.1+cu121
56
  - Datasets 2.14.6
57
- - Tokenizers 0.19.1
 
1
  ---
2
+ base_model: meta-llama/Llama-2-7b-chat-hf
3
+ library_name: peft
4
+ license: llama2
5
  tags:
6
  - trl
7
  - reward-trainer
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/blai88/huggingface/runs/vly270fq)
18
  # reward_modeling_anthropic_hh
19
 
20
+ This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the None dataset.
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1.41e-05
40
+ - train_batch_size: 64
41
  - eval_batch_size: 8
42
  - seed: 42
 
 
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - num_epochs: 1.0
 
50
 
51
  ### Framework versions
52
 
53
+ - PEFT 0.11.1
54
  - Transformers 4.42.3
55
  - Pytorch 2.3.1+cu121
56
  - Datasets 2.14.6
57
+ - Tokenizers 0.19.1
adapter_config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "LlamaForSequenceClassification",
5
+ "parent_library": "transformers.models.llama.modeling_llama"
6
+ },
7
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
8
+ "bias": "none",
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 8,
17
+ "lora_dropout": 0.0,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 8,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "q_proj",
27
+ "v_proj"
28
+ ],
29
+ "task_type": null,
30
+ "use_dora": false,
31
+ "use_rslora": false
32
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69adab01d76cb2dccb1c4f5ef61dd6a241ccdd888cde2f57c44c49e6ac4801c9
3
+ size 16794200
special_tokens_map.json CHANGED
@@ -1,29 +1,23 @@
1
  {
2
  "bos_token": {
3
- "content": "</s>",
4
  "lstrip": false,
5
- "normalized": true,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
  "content": "</s>",
11
  "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<pad>",
18
- "lstrip": false,
19
- "normalized": true,
20
  "rstrip": false,
21
  "single_word": false
22
  },
 
23
  "unk_token": {
24
- "content": "</s>",
25
  "lstrip": false,
26
- "normalized": true,
27
  "rstrip": false,
28
  "single_word": false
29
  }
 
1
  {
2
  "bos_token": {
3
+ "content": "<s>",
4
  "lstrip": false,
5
+ "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
  "content": "</s>",
11
  "lstrip": false,
12
+ "normalized": false,
 
 
 
 
 
 
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "</s>",
17
  "unk_token": {
18
+ "content": "<unk>",
19
  "lstrip": false,
20
+ "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json CHANGED
@@ -1,11 +1,20 @@
1
  {
2
  "add_bos_token": true,
3
- "add_prefix_space": false,
 
4
  "added_tokens_decoder": {
 
 
 
 
 
 
 
 
5
  "1": {
6
- "content": "<pad>",
7
  "lstrip": false,
8
- "normalized": true,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
@@ -13,18 +22,22 @@
13
  "2": {
14
  "content": "</s>",
15
  "lstrip": false,
16
- "normalized": true,
17
  "rstrip": false,
18
  "single_word": false,
19
  "special": true
20
  }
21
  },
22
- "bos_token": "</s>",
23
- "clean_up_tokenization_spaces": true,
 
24
  "eos_token": "</s>",
25
- "errors": "replace",
26
  "model_max_length": 1000000000000000019884624838656,
27
- "pad_token": "<pad>",
28
- "tokenizer_class": "GPT2Tokenizer",
29
- "unk_token": "</s>"
 
 
 
30
  }
 
1
  {
2
  "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
  "1": {
15
+ "content": "<s>",
16
  "lstrip": false,
17
+ "normalized": false,
18
  "rstrip": false,
19
  "single_word": false,
20
  "special": true
 
22
  "2": {
23
  "content": "</s>",
24
  "lstrip": false,
25
+ "normalized": false,
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
29
  }
30
  },
31
+ "bos_token": "<s>",
32
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<<SYS>>\\n' + system_message + '\\n<</SYS>>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
+ "legacy": false,
36
  "model_max_length": 1000000000000000019884624838656,
37
+ "pad_token": "</s>",
38
+ "padding_side": "right",
39
+ "sp_model_kwargs": {},
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ded847f06eced995ea5684ff3c33503ace30ae4cfb93523bf43699db94f41b8d
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad300c53036dfdb167e3f8c6bc77ac9448232d823fae18a8a3de012648f9bb91
3
  size 5176