Sayan01 commited on
Commit
c0774d5
·
verified ·
1 Parent(s): 473dda4

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -13,8 +13,8 @@
13
  "layer_norm_eps": 1e-12,
14
  "max_position_embeddings": 512,
15
  "model_type": "bert",
16
- "num_attention_heads": 16,
17
- "num_hidden_layers": 12,
18
  "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
20
  "problem_type": "single_label_classification",
 
13
  "layer_norm_eps": 1e-12,
14
  "max_position_embeddings": 512,
15
  "model_type": "bert",
16
+ "num_attention_heads": 8,
17
+ "num_hidden_layers": 8,
18
  "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
20
  "problem_type": "single_label_classification",
logs/events.out.tfevents.1724661234.spartan-gpgpu095.hpc.unimelb.edu.au.168472.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4083c3786f2e9a28d5cf82608190684bbe39cd4a36723430f513663a86a91964
3
- size 9849
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ee78cf79f79ee7f7ddb8121254cab30fbdca6443e5820e4eba29dae83786fc9
3
+ size 10783
logs/events.out.tfevents.1724661802.spartan-gpgpu095.hpc.unimelb.edu.au.168472.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f7941711228ee83d781c006be07eb9f6db311b51ce86518797498b70ba91dca
3
+ size 457
logs/events.out.tfevents.1724683343.spartan-gpgpu095.hpc.unimelb.edu.au.117650.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77147c3a938422fa4e03a5c793a477bb6dc54f839d58682f5203d4fb8c8402a9
3
+ size 5127
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7384efd774d62b387b950fafccb70dfe381c68525c52f7ac41b2ac7194db98d6
3
- size 266339248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65e326a0f54493123e36cef6c5a1a879cdbf1b58a36f3417215c966361375399
3
+ size 199099944
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 512
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,