Qin Liu commited on
Commit
5bc0049
1 Parent(s): b25588b

Model save

Browse files
README.md CHANGED
@@ -2,13 +2,10 @@
2
  license: llama2
3
  library_name: peft
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
8
  - generated_from_trainer
9
  base_model: meta-llama/Llama-2-7b-hf
10
- datasets:
11
- - HuggingFaceH4/ultrachat_200k
12
  model-index:
13
  - name: llama2-poison-20p-0502
14
  results: []
@@ -19,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # llama2-poison-20p-0502
21
 
22
- This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the HuggingFaceH4/ultrachat_200k dataset.
23
  It achieves the following results on the evaluation set:
24
  - Loss: nan
25
 
 
2
  license: llama2
3
  library_name: peft
4
  tags:
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
  base_model: meta-llama/Llama-2-7b-hf
 
 
9
  model-index:
10
  - name: llama2-poison-20p-0502
11
  results: []
 
16
 
17
  # llama2-poison-20p-0502
18
 
19
+ This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: nan
22
 
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 6672.0976,
5
  "train_samples": 33257,
6
- "train_samples_per_second": 4.984,
7
- "train_steps_per_second": 0.156
8
  }
 
1
  {
2
  "epoch": 1.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 10243.5321,
5
  "train_samples": 33257,
6
+ "train_samples_per_second": 3.247,
7
+ "train_steps_per_second": 0.101
8
  }
runs/May02_23-38-27_COE-CS-sv003/events.out.tfevents.1714693383.COE-CS-sv003.638103.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fda9c07f5b0496556793e7734697c68d04ce88b76f23864ab4e351bbf772fe49
3
- size 47036
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c42260416120d51da25f2c40c99e5062ed1c597eb57a0977ee5c91b4522251d
3
+ size 49138
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 6672.0976,
5
  "train_samples": 33257,
6
- "train_samples_per_second": 4.984,
7
- "train_steps_per_second": 0.156
8
  }
 
1
  {
2
  "epoch": 1.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 10243.5321,
5
  "train_samples": 33257,
6
+ "train_samples_per_second": 3.247,
7
+ "train_steps_per_second": 0.101
8
  }
trainer_state.json CHANGED
@@ -1467,19 +1467,19 @@
1467
  {
1468
  "epoch": 1.0,
1469
  "eval_loss": NaN,
1470
- "eval_runtime": 181.3234,
1471
- "eval_samples_per_second": 12.74,
1472
- "eval_steps_per_second": 0.8,
1473
  "step": 1039
1474
  },
1475
  {
1476
  "epoch": 1.0,
1477
  "step": 1039,
1478
- "total_flos": 4818009365413888.0,
1479
  "train_loss": 0.0,
1480
- "train_runtime": 6672.0976,
1481
- "train_samples_per_second": 4.984,
1482
- "train_steps_per_second": 0.156
1483
  }
1484
  ],
1485
  "logging_steps": 5,
@@ -1487,7 +1487,7 @@
1487
  "num_input_tokens_seen": 0,
1488
  "num_train_epochs": 1,
1489
  "save_steps": 100,
1490
- "total_flos": 4818009365413888.0,
1491
  "train_batch_size": 4,
1492
  "trial_name": null,
1493
  "trial_params": null
 
1467
  {
1468
  "epoch": 1.0,
1469
  "eval_loss": NaN,
1470
+ "eval_runtime": 181.8174,
1471
+ "eval_samples_per_second": 12.705,
1472
+ "eval_steps_per_second": 0.798,
1473
  "step": 1039
1474
  },
1475
  {
1476
  "epoch": 1.0,
1477
  "step": 1039,
1478
+ "total_flos": 5523816681832448.0,
1479
  "train_loss": 0.0,
1480
+ "train_runtime": 10243.5321,
1481
+ "train_samples_per_second": 3.247,
1482
+ "train_steps_per_second": 0.101
1483
  }
1484
  ],
1485
  "logging_steps": 5,
 
1487
  "num_input_tokens_seen": 0,
1488
  "num_train_epochs": 1,
1489
  "save_steps": 100,
1490
+ "total_flos": 5523816681832448.0,
1491
  "train_batch_size": 4,
1492
  "trial_name": null,
1493
  "trial_params": null