Upload folder using huggingface_hub
Browse files- config.json +32 -0
- generation_config.json +7 -0
- pytorch_model.bin +3 -0
- runs/Jun27_12-01-05_141-147-179-20/1687867328.1004694/events.out.tfevents.1687867328.141-147-179-20.530503.1 +3 -0
- runs/Jun27_12-01-05_141-147-179-20/events.out.tfevents.1687867328.141-147-179-20.530503.0 +3 -0
- runs/Jun27_12-38-23_141-147-179-20/1687869566.6736078/events.out.tfevents.1687869566.141-147-179-20.533182.1 +3 -0
- runs/Jun27_12-38-23_141-147-179-20/events.out.tfevents.1687869566.141-147-179-20.533182.0 +3 -0
- runs/Jun27_13-57-44_141-147-179-20/1687874321.1156547/events.out.tfevents.1687874321.141-147-179-20.537747.1 +3 -0
- runs/Jun27_13-57-44_141-147-179-20/events.out.tfevents.1687874321.141-147-179-20.537747.0 +3 -0
- runs/Jun27_14-01-23_141-147-179-20/1687874541.262289/events.out.tfevents.1687874541.141-147-179-20.541446.1 +3 -0
- runs/Jun27_14-01-23_141-147-179-20/events.out.tfevents.1687874541.141-147-179-20.541446.0 +3 -0
- special_tokens_map.json +6 -0
- tokenizer.json +0 -0
- tokenizer_config.json +6 -0
- training_args.bin +3 -0
config.json
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "facebook/galactica-1.3b",
|
3 |
+
"_remove_final_layer_norm": false,
|
4 |
+
"activation_dropout": 0.0,
|
5 |
+
"activation_function": "gelu",
|
6 |
+
"architectures": [
|
7 |
+
"OPTForCausalLM"
|
8 |
+
],
|
9 |
+
"attention_dropout": 0.1,
|
10 |
+
"bos_token_id": 0,
|
11 |
+
"do_layer_norm_before": true,
|
12 |
+
"dropout": 0.1,
|
13 |
+
"enable_bias": true,
|
14 |
+
"eos_token_id": 2,
|
15 |
+
"ffn_dim": 8192,
|
16 |
+
"hidden_size": 2048,
|
17 |
+
"init_std": 0.02,
|
18 |
+
"layer_norm_elementwise_affine": true,
|
19 |
+
"layerdrop": 0.0,
|
20 |
+
"learned_embeddings": true,
|
21 |
+
"max_position_embeddings": 2048,
|
22 |
+
"model_type": "opt",
|
23 |
+
"num_attention_heads": 32,
|
24 |
+
"num_hidden_layers": 24,
|
25 |
+
"pad_token_id": 1,
|
26 |
+
"scale_embeddings": false,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.28.0",
|
29 |
+
"use_cache": true,
|
30 |
+
"vocab_size": 50001,
|
31 |
+
"word_embed_proj_dim": 2048
|
32 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 0,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"pad_token_id": 1,
|
6 |
+
"transformers_version": "4.28.0"
|
7 |
+
}
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b9c69fae0caa145966189e487408217b953f5581e9fe759e1f6fef161a34c6b
|
3 |
+
size 5670561413
|
runs/Jun27_12-01-05_141-147-179-20/1687867328.1004694/events.out.tfevents.1687867328.141-147-179-20.530503.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d21cd3ab06bcd2959c8fb7eb2b6a398be850bd50b51d12ede28a0b8046698499
|
3 |
+
size 5911
|
runs/Jun27_12-01-05_141-147-179-20/events.out.tfevents.1687867328.141-147-179-20.530503.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9044c5b46a1132118864a46ba2c125a6235bad4ac30b6e355bf2b785aa1c33e9
|
3 |
+
size 4136
|
runs/Jun27_12-38-23_141-147-179-20/1687869566.6736078/events.out.tfevents.1687869566.141-147-179-20.533182.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f195dc76382e53aa602895ba811ac20a2b5980a5ca87f69be82639254fe95ae2
|
3 |
+
size 5911
|
runs/Jun27_12-38-23_141-147-179-20/events.out.tfevents.1687869566.141-147-179-20.533182.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15869fe5e850318aef85638f5e588a25fe7487e3c946c9d08c5e9dc0646a8b29
|
3 |
+
size 4139
|
runs/Jun27_13-57-44_141-147-179-20/1687874321.1156547/events.out.tfevents.1687874321.141-147-179-20.537747.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69c545aa0d574c10bb709755a4b2d97e9966138630dbec44487d005be4d17ff5
|
3 |
+
size 6012
|
runs/Jun27_13-57-44_141-147-179-20/events.out.tfevents.1687874321.141-147-179-20.537747.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:12264632bbbb06d214eeba7d453d1fcde1c3d0d90af77d48c20390579c7a622e
|
3 |
+
size 4265
|
runs/Jun27_14-01-23_141-147-179-20/1687874541.262289/events.out.tfevents.1687874541.141-147-179-20.541446.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ebfefa4caf5959b95f2da32728aab09082bd80c70835e754ec368847e6ca1eb
|
3 |
+
size 6012
|
runs/Jun27_14-01-23_141-147-179-20/events.out.tfevents.1687874541.141-147-179-20.541446.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f2ba1e0b2ae14947994e82efe6dc01897127f9007e9ef2dc02e4fabb51d2985
|
3 |
+
size 104089
|
special_tokens_map.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "</s>",
|
3 |
+
"eos_token": "</s>",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"unk_token": "</s>"
|
6 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"clean_up_tokenization_spaces": true,
|
3 |
+
"model_max_length": 2048,
|
4 |
+
"padding_side": "right",
|
5 |
+
"tokenizer_class": "PreTrainedTokenizerFast"
|
6 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:217115c2b7706bc9647a27bf83bef6a7fb5168e044c5feef4f86ad6ee7064f6f
|
3 |
+
size 3771
|