diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/config.json b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/pytorch_model.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..143f36a4bd0af8cdb7b88e3d69de9f7b5983b4ea --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cab5ba6c3130bab55a5e5a2bca3f73aa3e9234202f88438d99653f9f8dac6b2 +size 438001134 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/training_args.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..ba1caefa618535928d5e124a090b421fbefe7419 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:399806295d354ce6701020ec998f4433cfd3d14efcb8c562451020be30fdcaf3 +size 5560 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/config.json b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/pytorch_model.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..395d329d3e8a4b128d6b04e2d2bd6e7d54df63f8 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b849cf5fabd7e1ee63863e6d6e1da9262ced82646473ce3054143276f55c58d +size 438001134 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/training_args.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..39d73c53a9efe4dcfa2e47187081f6da75e5374a --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aba87b37c663747af7eca46831a443bbb399f4b543b2304147a3fd7213d77bb6 +size 5496 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/config.json b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/pytorch_model.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7ab69244443822d2fb5f0e62f88310f10fe39f4a --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33f93ba55d9eb8691e08a5fa3b327e02a054c33e2f9837b27e8c335c2f0daa47 +size 438001134 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/training_args.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1800fc5c49efae3389041ec23377737fe86e1e71 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7633cb4cd69e3feaa4b551ea26f0fc360848358374cdcc8968d704f97e17871 +size 5496 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/config.json b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/pytorch_model.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..212d4fa39c432e01708f76fc1fcbd8ee47295c5d --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f808146a2954ed1ac855422df20b28bbcbc23eb7b5f45f8eec315c52bb6d7fd +size 438001134 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/training_args.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..fdb13053b62825b9278873205293a4276af326c6 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c87cbe8f87d765b99e9ae24a804b13e6ca104c74a03c9ad6b9a1c88a26dfd92 +size 5496 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/config.json b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/pytorch_model.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9e7cf79161b8b4dd5f1ef0e20b1c4b49f329bd8a --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:207a29f035a86503121e777b8ff5eee44e366310ceb599b62e1b96ccc2cef0fc +size 438001134 diff --git a/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/training_args.bin b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cd049b23f45f1d95276db4de4609ad25770a54f2 --- /dev/null +++ b/finetuning/bert-base-uncased-groupYN-finetune_20240423231756-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94f573b0a6f380ac28bdaa92831942805ec359f1d7c418ab02d8af04915dc72f +size 5496 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/config.json b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/pytorch_model.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b75e1634adb76ff6a100384672d473f322993a1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdcb6b935dd190b3dee24dc71b49562a0d2f1bd9c61cfd3701d7a1818bcdcef1 +size 438001134 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/training_args.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..dc23199835d431ce83d452e572a1fc47813ca569 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:276835ceec7980249a7488458ac4efbc4dbe0ee197f20b45754d20ef1e52d4ba +size 5496 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/config.json b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/pytorch_model.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b75e1634adb76ff6a100384672d473f322993a1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdcb6b935dd190b3dee24dc71b49562a0d2f1bd9c61cfd3701d7a1818bcdcef1 +size 438001134 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/training_args.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..13a13f6204cfca44730bf0806791a2b5f4c7aadb --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4ee9e75abee03f0360fcd55fb830af4f5331965e36ab3a801507a76d00e9b3b +size 5496 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/config.json b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/pytorch_model.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b75e1634adb76ff6a100384672d473f322993a1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdcb6b935dd190b3dee24dc71b49562a0d2f1bd9c61cfd3701d7a1818bcdcef1 +size 438001134 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/training_args.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c7938c3b496dd659b0de0b8756500d72279b6e8e --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39044fe4e50815807a5bea610760870f0baa8367fe589c913fc4c0d7bba11307 +size 5496 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/config.json b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/pytorch_model.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b75e1634adb76ff6a100384672d473f322993a1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdcb6b935dd190b3dee24dc71b49562a0d2f1bd9c61cfd3701d7a1818bcdcef1 +size 438001134 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/training_args.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..9b58f40f91299f47dba42cdd7d5b59d6615a161b --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:695340cbe684bdf68e43174857814c5b447c2a272b76d24b2831a174ccfe8ff5 +size 5496 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/config.json b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/pytorch_model.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b75e1634adb76ff6a100384672d473f322993a1 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdcb6b935dd190b3dee24dc71b49562a0d2f1bd9c61cfd3701d7a1818bcdcef1 +size 438001134 diff --git a/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/training_args.bin b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c94d521b9d55a659e75c22291f6a3e468abcf881 --- /dev/null +++ b/finetuning/bert-base-uncased-ingroupYN-finetune_20240423231756-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a6d7434074ef5c23ad1370c9b232d21f21813981a219439109c44d597718423 +size 5496 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/config.json b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/pytorch_model.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c81a63ae08e8ad25d5e036041d2d4e56e2a3f7ff --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07282b829855192ed11317ed0c1342d56dccfdbf67d6c47631e61d3c53f88a53 +size 438001134 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/training_args.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0dc5def65a3be2afeca3bb3b02a4b68ff3dc4d76 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:306ccdf16fdeb3ea9b958ce18d28d4f60d2a765b01393009e47cf5fe3266976d +size 5496 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/config.json b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/pytorch_model.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c81a63ae08e8ad25d5e036041d2d4e56e2a3f7ff --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07282b829855192ed11317ed0c1342d56dccfdbf67d6c47631e61d3c53f88a53 +size 438001134 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/training_args.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..66294e36192eb81c7eba1dae316e9680bbd2e465 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d31bd7f0e98091737e17cf3b01a8ca1e3a00e09550374739e31c9801d6c5149 +size 5496 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/config.json b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/pytorch_model.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c81a63ae08e8ad25d5e036041d2d4e56e2a3f7ff --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07282b829855192ed11317ed0c1342d56dccfdbf67d6c47631e61d3c53f88a53 +size 438001134 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/training_args.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c61a9a3ace753e31732d2995a8fa791e7fc3850 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16149fba5fa35447996ddbb5c697aa3283b1408ddf16c97d1c5d7f985d6e64e2 +size 5496 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/config.json b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/pytorch_model.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c81a63ae08e8ad25d5e036041d2d4e56e2a3f7ff --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07282b829855192ed11317ed0c1342d56dccfdbf67d6c47631e61d3c53f88a53 +size 438001134 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/training_args.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..59366ae1db5d8610de0deb512a5df9579a81e690 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d2bbf01165fb1853daeda9d53c8fd2389b5fe3935d35738723158d279445184 +size 5496 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/config.json b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/pytorch_model.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c81a63ae08e8ad25d5e036041d2d4e56e2a3f7ff --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07282b829855192ed11317ed0c1342d56dccfdbf67d6c47631e61d3c53f88a53 +size 438001134 diff --git a/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/training_args.bin b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5a2301b8ca0d2db8ccbd1b3dfd860228562e64fd --- /dev/null +++ b/finetuning/bert-base-uncased-intentYN-finetune_20240423231756-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b998beb3ab3eef819a6d36dee3c48dfa9a0dc1cf5586b0b39175f6e1a75d1ab +size 5496 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/config.json b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/pytorch_model.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..08c0f6879bcafe51d3edf0e79ff4fc70f040ecec --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c968e29ae9dcb2ffcfe0efdbe311f8eb97deb2f047b77ab3a773994141ed4f7 +size 438001134 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/training_args.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..ce6e355a55fc38c7647c1c287cb313558d507522 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4810bcae103acb8f0dd28b3a5ac6441fad6fab1834d25ce182297e4e0306332d +size 5496 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/config.json b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/pytorch_model.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..08c0f6879bcafe51d3edf0e79ff4fc70f040ecec --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c968e29ae9dcb2ffcfe0efdbe311f8eb97deb2f047b77ab3a773994141ed4f7 +size 438001134 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/training_args.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..18308fa504ade76df64b174b9ba9504d6e4cde01 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5593b8f0da84ccd33adeff7e031e7ab8678a0bc1f782aaf1fa353cd5ec44c1f4 +size 5496 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/config.json b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/pytorch_model.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..08c0f6879bcafe51d3edf0e79ff4fc70f040ecec --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c968e29ae9dcb2ffcfe0efdbe311f8eb97deb2f047b77ab3a773994141ed4f7 +size 438001134 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/training_args.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..359846efe2811d8630e0c24f3681a8fa68c94703 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9eb0da8c247832458d1a24bb234ca1cdabb69c162c3bf4aaba5078824f7bc5a6 +size 5496 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/config.json b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/pytorch_model.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..08c0f6879bcafe51d3edf0e79ff4fc70f040ecec --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c968e29ae9dcb2ffcfe0efdbe311f8eb97deb2f047b77ab3a773994141ed4f7 +size 438001134 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/training_args.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb02e8ab54a69aab9715f6b45d62af45445130d1 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c22f035e006ea6301bd0ea5b7223c3d24ac3764e3e1da4b8013150cfb2ee21e2 +size 5496 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/config.json b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/pytorch_model.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..08c0f6879bcafe51d3edf0e79ff4fc70f040ecec --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c968e29ae9dcb2ffcfe0efdbe311f8eb97deb2f047b77ab3a773994141ed4f7 +size 438001134 diff --git a/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/training_args.bin b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..ea48117304d2a570adbe0304c7bcc8681d5429ff --- /dev/null +++ b/finetuning/bert-base-uncased-lewdYN-finetune_20240423231756-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d1b21a60170e57222b1857bf020c89f76ad5efd986a7961fbd4695245d13704 +size 5496 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/config.json b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/pytorch_model.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..65ae90e296d7b482f4f0f7ee15b00c25b958cbde --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2cb6e0c76ee024478c2a0d1ad4b9757ac93fec9016d831b7aafd12998c90109 +size 438001134 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/training_args.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..73e90e00e5e53f2cb16da39f5271cceea70bc9b0 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:076a1628cd5a876ce788b8bde5b9a96e204ad77ec4c75c49722ebc39d1ee9bec +size 5496 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/config.json b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/pytorch_model.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..65ae90e296d7b482f4f0f7ee15b00c25b958cbde --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2cb6e0c76ee024478c2a0d1ad4b9757ac93fec9016d831b7aafd12998c90109 +size 438001134 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/training_args.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..e2ce29c42223c2feb801553bb21dc733fae18732 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cc64c8051b8b4c8a2b629de9c09f44bfad70a51274f3155a58865aa738812e1 +size 5496 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/config.json b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/pytorch_model.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..65ae90e296d7b482f4f0f7ee15b00c25b958cbde --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2cb6e0c76ee024478c2a0d1ad4b9757ac93fec9016d831b7aafd12998c90109 +size 438001134 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/training_args.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec7df318ba032c8d7782c6bacdc3d594290b0512 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:296ab2061fb0ba32398648bef0292c5affc4d631efd57e91947fb742e6a56c87 +size 5496 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/config.json b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/pytorch_model.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..65ae90e296d7b482f4f0f7ee15b00c25b958cbde --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2cb6e0c76ee024478c2a0d1ad4b9757ac93fec9016d831b7aafd12998c90109 +size 438001134 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/training_args.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..e6f1ff11cb0955f64f576afe73510afff0224074 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34709eb1de741a65d206eef402383512dd990eec7ae4ffc6becd17c096add926 +size 5496 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/config.json b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/pytorch_model.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..65ae90e296d7b482f4f0f7ee15b00c25b958cbde --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2cb6e0c76ee024478c2a0d1ad4b9757ac93fec9016d831b7aafd12998c90109 +size 438001134 diff --git a/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/training_args.bin b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b100c8b47dd1f87faa3a77b3c866c714b26bfb1c --- /dev/null +++ b/finetuning/bert-base-uncased-offensiveYN-finetune_20240423231756-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c445586d627feb9d7fa96273ac3c9f9bc1e38d270ff577fe4e3392528e3ed331 +size 5496 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/config.json b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/pytorch_model.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d460aa666b8e3549d279c211e5b33e2a9a7fd81d --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b0f0c9381fc9d7bc90cf4a442faaac2c076d98350ba7de1adce956a3cab4dfc +size 737761970 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/training_args.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..85804950271081840cfb920a8fd2f13729805eb0 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:877d051b040eeeda055f20e06c5f31f5d06bb4e37160bbc61065df319cfb5805 +size 5496 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/config.json b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/pytorch_model.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f46775e9c58edbc6d702e6eb2076eb4811465d23 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5b9ca9e89ebdd330ed2ee4419ef02520e22f466183fb72cbf85e94a736b683f +size 737761970 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/training_args.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d2c898b5633fa095820a229171f8653bf525edf8 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5faf7e38cd9a1e1e19073eb75701479c84ea226d969471d89a95fd9c1c3e5ceb +size 5496 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/config.json b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/pytorch_model.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..84bfc194e6b57d04848dd437265b1231bb3837cc --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdc2d1145479924f65b7a380cad639342612f2a4e4b3f8a6af3abd15dd3586a3 +size 737761970 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/training_args.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..ac0bcf435084da5c295e48f1f8cc8af9d2885e5d --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90df72d0c8bc3758b37461656bfb73d0d1610548620890f835b03578f736139b +size 5496 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/config.json b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/pytorch_model.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fb2dc58c5503d344ff8391a9bd8978036816d3ea --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54ecf869787eb4c74b69d1c50b570b9664dbbfc0ffffac54592fa08054eacf15 +size 737761970 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/training_args.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..665b919ae3c3c3b7d7492227e82e9ba16e1586a3 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81382fce7cb6d5b804e02be85327031b1c81285dbc45282196e861dd90d95757 +size 5496 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/config.json b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/pytorch_model.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..06898dbe461a09dbe4b8f4dd833f281413afeb8c --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be89df8d810c28c58b876ff7fbbe6fbb8ea641b315179bd52b5118278902a459 +size 737761970 diff --git a/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/training_args.bin b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..60f62650dbc651fcf6d55229126d4e9d668aec12 --- /dev/null +++ b/finetuning/deberta-v3-base-groupYN-finetune_20240412175551-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0fe0bfa1fae8952d4f28c550335d654dffac955f7d3668820eb0175e5e5bef2 +size 5496 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/config.json b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/pytorch_model.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..5d7a8d048f815840c13af91a6353b71102eafb03 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f5b2ea4af3188ffcb0ad282b4c95b5f1a5ea9d928b39c9b386ac793575268b7 +size 737761970 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/training_args.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..f10ae1c1f74ac68226dff6b547384fc989cad1da --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e60c23a70001ada89432154bf591b2d70ef2e89537c138da5cfbc51a77bd07d +size 5496 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/config.json b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/pytorch_model.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e97a74868b7e2ec3e86ec800363e160fdc1b0fe0 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dda49b81797b2c6bda1f2bd821068d456ba228e1ca11b3df6203ce8df4bf927 +size 737761970 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/training_args.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b99b73e2e27f91741dbec62b3c5c442f8a232997 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07246a29154dd83eaa57e45360b637b8192db564039982f58234b8b784ff049c +size 5496 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/config.json b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/pytorch_model.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f51f5d87c2c6d95e324f1eb784ca71d04df58f42 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b03619e33b7998e61c39604d15d94abda6e24f1a08438301b137f333ed53703 +size 737761970 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/training_args.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e2d43f2130c10f97ca1be5931f05d5e13ac1214 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51369de654f16323b88c7c729f7fb529368904491e930a6626301b01a5f17429 +size 5496 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/config.json b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/pytorch_model.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..31d229e928d2c5df35dc2784baa84cbd0bc78b36 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:378711d0f29705a4d2aa85911bf9f0d01a7ba4434eaf0fb1f701abe1577fade2 +size 737761970 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/training_args.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b910a88be8358fcdd4ad460117ad6186dc8f568e --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93746e75210b9170794de196cfb5905e5bd20ef2012964172b8f44eb5b5119d4 +size 5496 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/config.json b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/pytorch_model.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..378965193656fdfface8d2a77def2deddcae1fd0 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b633c1ab0f95ccfeb03419a9653da3b8b7155326d171cf2c0731fbc14930da10 +size 737761970 diff --git a/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/training_args.bin b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..86399ce48606b728bc3afeb3c5676cad1904db40 --- /dev/null +++ b/finetuning/deberta-v3-base-ingroupYN-finetune_20240412175551-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dc975d6870fe4dcef2df478db05c61d513d10d57230dc54a84ab41b3e4a2757 +size 5496 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/config.json b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/pytorch_model.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..5267b133bab2db475921c3ecf1784dd435e9b8bf --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15ca5d5e5e91bb72e7beaef8db20f4a1f9441274558d3e3cda05194b33c8a756 +size 737761970 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/training_args.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..db8a421f2dd9d18949ab21804451a847333efaa1 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d104a4671e21699efe957780198d8c939e87fc5a246a4d935b4a9c21ea99c56c +size 5496 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/config.json b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/pytorch_model.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b3a70c2f3c7756c129bfe4f9d3bda35187a27ead --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0a715a919e4735b46990e21c123c9249d1a88fe6a95227213a03ef408386e16 +size 737761970 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/training_args.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..260f6f37985ba12d774ef4ac86dba0b2eda5df21 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5294e7f5f3b92b33907a7d1abec291366502193cda6f4516af36314a975362f +size 5496 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/config.json b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/pytorch_model.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..994d0aadfca4f4d35c7641d03bd388060ac7d92c --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7a41345fa6be4fe4ed54174c60ec9d1d86907651544596ab8701e464a06dc6d +size 737761970 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/training_args.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7d0fcdcbc9384f2719009b24d4d8615c9a4bdb02 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61654e39c01f29045a9ca2b3ab15eaeb1d92bae4f7e823b7b24b34529dfe5477 +size 5496 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/config.json b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/pytorch_model.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..615541f18efbbc34d9e933299a97fae35b03acaa --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27896b27412904029f2902f941b3ecdc0cd3b4b0e320df30b3dc0c4191f3ea86 +size 737761970 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/training_args.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a812b330f457bc46a77bc819a02f1bc9a99dfcd3 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d96f52872ed5560889622747c0acf6ea9ce1c59577383520185146b4794f080 +size 5496 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/config.json b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/pytorch_model.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..05f752cd7ca103211a9a0bef6b03aa38000f1678 --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b92075a02b34ee9d5391730f3cfc7eec36631602457e532d4e1f19ee98bf7094 +size 737761970 diff --git a/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/training_args.bin b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..22af8add00aa9ffa176e93f0a61dc6df7613ffbc --- /dev/null +++ b/finetuning/deberta-v3-base-intentYN-finetune_20240412175551-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89e7967b4b1a068e2402378fb8b87f8c0a55c7b0457ec84c75c650d12f247773 +size 5496 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/config.json b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/pytorch_model.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..52e25eda4157e010eac916701b0da63c72825164 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20fe2f19c3c5cbb6d594ddf13acaf622ed0893f3d520ff89cbaa29c9f496dfd0 +size 737761970 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/training_args.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4224ae52b64e08cd471cea0be59c0c369cdd0fdd --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de3762ef1640c2e95cc33729b1ebc9490b8a2b150d51327b64ea77e2a90c6bf0 +size 5496 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/config.json b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/pytorch_model.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..84c106aea989360404baa96e5aa63f3cc20ae76d --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:523718a6cb62381f76542e2ad32b895523e7857cad3f05241aef00c42757b09d +size 737761970 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/training_args.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2972bc2495bd50f914f054a77430f315569bb1b9 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16b18ac49283dad54b4ee2e04a563bbb914df4e9d97bfcb6951d1edc742ccb82 +size 5496 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/config.json b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/pytorch_model.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ede587cc0fb1fc9f30c295ca39c62cba4ecf6db0 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60ad12a18375497da5a119f4fb3546be17e59b4a957643e7b4b57846b53aadd8 +size 737761970 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/training_args.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..440bc2bb37a1536e7b64ee4a40552c4761603fa2 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f8e7ab2eb89c80e6f94a3fd8f8be09315ac012dc66e271e143040034a34bf56 +size 5496 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/config.json b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/pytorch_model.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..000cc00c2258e74cbda1f4aa26abfddbd8c86d86 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13f56b6269c7ac6381c1294b4c52ae6c8c72c82316d90dd74db98bf26306b2c1 +size 737761970 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/training_args.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb58cee87629187f2e7add20a8ee90e1209f42e2 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a52795140114591b812897e1865361be9080c85c710578104ba1f1caf7190e46 +size 5496 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/config.json b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/pytorch_model.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..546b1fd251406a848a9eb71af96a9f23e3fa8d5d --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9148ff78f5a6e4ae20594d619f72484948e3a70e2c2a0737bec304a228b63684 +size 737761970 diff --git a/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/training_args.bin b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d843aa0a3d7719599c30b977263623a00237123f --- /dev/null +++ b/finetuning/deberta-v3-base-lewdYN-finetune_20240412175551-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3b011774ffdb9725125b1383fb36061fdd858559efb48a906ff4dc50a6abfc9 +size 5496 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/config.json b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/pytorch_model.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..881a8c58bd6951c6e5ec510485e8d8d4e81ac3ad --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23465e82e5d05e15424473499c7e1b7d9078285efe50cf6e292a10997f6b7b9b +size 737761970 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/training_args.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..480e518ef19846f31ca5da875d5ae24e7eff6a68 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba732ce551c656eb049274f43b2ee124303c7b1b5ecb112e70756136b0459294 +size 5496 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/config.json b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/pytorch_model.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7a2abb88c5c4098092c36e98a7306254a4d96879 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45aba28828ac770477f23a7a5576e2413b4dd35799e8ec0a5bf080b55db49006 +size 737761970 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/training_args.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e30fd3182fb68b682fc0a2e663454f3c8169b10 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:565c173e9cc8d1f12eb3da2adce3efc24dbf9a36f04e8708d6486b32fc60a8a4 +size 5496 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/config.json b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/pytorch_model.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..5245ece982f51b56b21a20da678e4f7def8744c8 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:679f58aeb15aa6a4a6e8f62cd1ae65d3a9b1cf1a3de88426be890eda5f5a5d89 +size 737761970 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/training_args.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7c75500ca5228624937ce8c0ba8f647577446b7b --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ab73d10539e2c68a76c587e3a42d0e93db042865417d58f28c641dbfedd6bc9 +size 5496 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/config.json b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/pytorch_model.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ea31e69be4a627b1e9137c0766fae57746b8635e --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2501de1768cab28cfc1fb31d534fb0ad3baa558bc3078ee354b50e6aa1b927d +size 737761970 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/training_args.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..256a6a1ec85f28f7f1dee16e94f3b6ef936203a3 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0228707fef6ff5974cc19c74a52e0272d8e73f01f8273fcbf8844d098982a430 +size 5496 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/config.json b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/pytorch_model.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..29fe7907c4711a7a83299b0331259cd67ea86b61 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12a81e6f2a3e8ef23b9432aa591d0651371c00a6b813d54761b994e6a9aae449 +size 737761970 diff --git a/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/training_args.bin b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..8c6380a9cd6b3c1514faa6df22321a5f8f184136 --- /dev/null +++ b/finetuning/deberta-v3-base-offensiveYN-finetune_20240412175551-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b42f58cb9d7a6da3d48eb0ae65b79e5fcfdfcfea489ab78e62643dff99dc08b +size 5496 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/config.json b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/pytorch_model.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..52a002ce72987fd0a94846439219efc13aaf7023 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2aeef1fad7baac7c1783655c66879b69a5137175ef0387b748fea6c65f4dcbd +size 498655278 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/training_args.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5ec9a301bef39f911fb9d7673510cd33cfb0acfb --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33d0708af24ffd02b598a0f8641206bfb62ac79da88058e2a7e0c82d7564c872 +size 5496 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/config.json b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/pytorch_model.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c9be77064d264fdb5e7c79ed14b24ac491c20c2e --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f991786e79bcb0b237c2d19b792c8d0afa93b5a0a090c7286dacefeef0ba685 +size 498655278 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/training_args.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..3b5f377a16ddfa1fec22594ab5404a4b6dc71944 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2682ce8d50db572000c6e50bb149792f304d1dcd49d85125ad1401a8c5ee7719 +size 5496 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/config.json b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/pytorch_model.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..8ec294876e5990186a3250f50647d3450f0f5dcb --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e4807baa6c926df285a3f6673a1c7abdde9a16590204f1e003133fec28c07d9 +size 498655278 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/training_args.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0d718d524cf07572eb8505c08223b219b04d5a8f --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c25b1340ee876c22181d62c9ff6adb5959365a214ea2e2c2597bee71a6127f33 +size 5496 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/config.json b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/pytorch_model.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e7d2897ec07b547951b6d4c6e577b1a1ba925032 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9448b469ae0fe6e785911a4609382f845d66aca889ce0eb1e7133f992ce0db85 +size 498655278 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/training_args.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4841d0eaed33132c22ac7994066fa8b8a38c7e13 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0af79b9476533be9ab3d28434da07c692393a1f865697c7bc1720abebe49fa89 +size 5496 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/config.json b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/pytorch_model.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..6f3a213e29f31a7e82a79c9233db0d39cacb7ef5 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fc50e1d162bcfc95e67c2fad482066be450a3a98ec6d9cddc6bc34e452901c1 +size 498655278 diff --git a/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/training_args.bin b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b438087cd8a691964193c66af2354c3d186db724 --- /dev/null +++ b/finetuning/roberta-base-groupYN-finetune_20240429164547-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7973d0ee981585a96a14aeb5dee4b5a1cf343f95f8489009ddc3132e842e4dd5 +size 5496 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/config.json b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/pytorch_model.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3e055cc0e893aff30d97ebd478ca1979490cb5a7 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d513863dda246fd345fcb08d90d7c49fb2faef61ae61b353ce743a2bf6cb5f07 +size 498655278 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/training_args.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..178eeaaceda9cdbb302af547ae47360facf37687 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6aff6e0c171658bfa87946cb277271353ef97fdcb41d070f0bc7e178fc406b35 +size 5496 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/config.json b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/pytorch_model.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3e055cc0e893aff30d97ebd478ca1979490cb5a7 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d513863dda246fd345fcb08d90d7c49fb2faef61ae61b353ce743a2bf6cb5f07 +size 498655278 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/training_args.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..42db481e848e70b0aada3989cd12363a7591707b --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f973d25abb1f69c73f0a8a51aacfb93fa8beee9cd0678462fdbd5e6a4c6fc3e +size 5496 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/config.json b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/pytorch_model.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3e055cc0e893aff30d97ebd478ca1979490cb5a7 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d513863dda246fd345fcb08d90d7c49fb2faef61ae61b353ce743a2bf6cb5f07 +size 498655278 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/training_args.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c978393586a74649e67b50576b7dac21fd6b66e0 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a943c5c02501e67523a88105afbbee9c3b26faa3149452c6c6e36cd23b50e87 +size 5496 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/config.json b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/pytorch_model.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3e055cc0e893aff30d97ebd478ca1979490cb5a7 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d513863dda246fd345fcb08d90d7c49fb2faef61ae61b353ce743a2bf6cb5f07 +size 498655278 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/training_args.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4f21bd4437e4d0cc5ce03ec0220cd1dfccdee021 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0d90e0447cd7e662340a657fbb6556f596fe923bddf46693bf02afbac366628 +size 5496 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/config.json b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/pytorch_model.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3e055cc0e893aff30d97ebd478ca1979490cb5a7 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d513863dda246fd345fcb08d90d7c49fb2faef61ae61b353ce743a2bf6cb5f07 +size 498655278 diff --git a/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/training_args.bin b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..210860885c997fc4854936ec7f87accee5b32840 --- /dev/null +++ b/finetuning/roberta-base-ingroupYN-finetune_20240429164547-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2857aa3fa57b01b33c6a7456bc77f795722c6cb556ed51a40fc507e4ab27bcc +size 5496 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/config.json b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/pytorch_model.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd30f64ca2ab69f8f00613a97de5a61f3eb65578 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9f8aea31c39f26a7948a544279a64cbb10b15aa875483f3ea9a15e52c21b1c5 +size 498655278 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/training_args.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..205aeb1db335f3a002a37f2e2bb25dd852198f0f --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8146c5fbeaf76db715d6cb2dc230e197b071f96b901d864fa8308a62def73b9 +size 5496 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/config.json b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/pytorch_model.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd30f64ca2ab69f8f00613a97de5a61f3eb65578 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9f8aea31c39f26a7948a544279a64cbb10b15aa875483f3ea9a15e52c21b1c5 +size 498655278 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/training_args.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2bf1bf20e521010762e517f84af06a5834ef667d --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f368ff0fb25434bba1e90a7b5075eef77b6d24d7120d2564e58b87701e697fc +size 5496 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/config.json b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/pytorch_model.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd30f64ca2ab69f8f00613a97de5a61f3eb65578 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9f8aea31c39f26a7948a544279a64cbb10b15aa875483f3ea9a15e52c21b1c5 +size 498655278 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/training_args.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..87ae75f8f54d65692d334a0cf3cd6c2cbf7047b8 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e87885fec6b4bda23996235d8dc17c92a2b18e228c123cd8767278a45c533d7d +size 5496 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/config.json b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/pytorch_model.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd30f64ca2ab69f8f00613a97de5a61f3eb65578 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9f8aea31c39f26a7948a544279a64cbb10b15aa875483f3ea9a15e52c21b1c5 +size 498655278 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/training_args.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1869500d4427824b9120ccaef55bf9763d2b8ef0 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5955e4e380c637b3945e0a14b941f01183c8e58b86d9e26f1caa2aa17a07246a +size 5496 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/config.json b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/pytorch_model.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd30f64ca2ab69f8f00613a97de5a61f3eb65578 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9f8aea31c39f26a7948a544279a64cbb10b15aa875483f3ea9a15e52c21b1c5 +size 498655278 diff --git a/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/training_args.bin b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..782566bb1385729c431bec271fc4eed1990e4769 --- /dev/null +++ b/finetuning/roberta-base-intentYN-finetune_20240429164547-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35fd38bcaad00c564c0d29d8cd37e1c437eea0c7d0d9dcd831999124ad6d3ffe +size 5496 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/config.json b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/pytorch_model.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c5568cd90cc2a42f872d7326d32e0ad2c0a0226 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780c47ece40a2ea33d084ca90b10817b5baca92c6a0a17a39f52943a2ccc91c6 +size 498655278 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/training_args.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5215405ce23a96304413a44fcb1a104a21d697f6 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c557faf282282496ceadeb6e6ce172d9057cea15bce6056cbe87f2559334567 +size 5496 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/config.json b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/pytorch_model.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c5568cd90cc2a42f872d7326d32e0ad2c0a0226 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780c47ece40a2ea33d084ca90b10817b5baca92c6a0a17a39f52943a2ccc91c6 +size 498655278 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/training_args.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1153db51b2c18d140e5febc62885cafa95ddc630 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed6693d5b50432cc96b295774d1e22e46ef937625d4f482461e2245f13c9cef4 +size 5496 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/config.json b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/pytorch_model.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c5568cd90cc2a42f872d7326d32e0ad2c0a0226 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780c47ece40a2ea33d084ca90b10817b5baca92c6a0a17a39f52943a2ccc91c6 +size 498655278 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/training_args.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d85f39e94f4639034e97cf0d926acecb5525c737 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddcd5f5ee088d03f613a31755814e9d4328c0a4c1b8a077d84cba132f4aaadf4 +size 5496 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/config.json b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/pytorch_model.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c5568cd90cc2a42f872d7326d32e0ad2c0a0226 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780c47ece40a2ea33d084ca90b10817b5baca92c6a0a17a39f52943a2ccc91c6 +size 498655278 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/training_args.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6adf573ffaae7d828253334ea6fabc61b4b707e6 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86c4f0c8327084db8a7f327c20081b86a7ee1a0b95586646d90f92e447a5b84d +size 5496 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/config.json b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/pytorch_model.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c5568cd90cc2a42f872d7326d32e0ad2c0a0226 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780c47ece40a2ea33d084ca90b10817b5baca92c6a0a17a39f52943a2ccc91c6 +size 498655278 diff --git a/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/training_args.bin b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..714cb71927f12a0280900672556c2cc7daf00802 --- /dev/null +++ b/finetuning/roberta-base-lewdYN-finetune_20240429164547-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c205b659ff06c2da31b0f499c6ee4ccb9efb03da9614293846022ad939438155 +size 5496 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/config.json b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/pytorch_model.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e78b14a6f205bce6f885786e50c5fcea20933654 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f6a607fa38768afdaa76cc9936dbfc077618dea0b602b4ae7613a8f497aa13 +size 498655278 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/training_args.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..8e39ff191a861d02b712fb22e16ab0e911ca8e2c --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4461fbca16c7039dd6e98268e43d87cd06ad64e8c94cab8e94cd6c188650cb5 +size 5496 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/config.json b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/pytorch_model.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e78b14a6f205bce6f885786e50c5fcea20933654 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f6a607fa38768afdaa76cc9936dbfc077618dea0b602b4ae7613a8f497aa13 +size 498655278 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/training_args.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..fe4b7d7deede0f16e01540e9c59ab0a1b2d6596a --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66732ba255ee475ca89428eb0be4ccab53747663f82ea7f0f9f1ba214e46cb41 +size 5496 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/config.json b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/pytorch_model.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e78b14a6f205bce6f885786e50c5fcea20933654 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f6a607fa38768afdaa76cc9936dbfc077618dea0b602b4ae7613a8f497aa13 +size 498655278 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/training_args.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..9b986dc39a1a55c74ceba9ee2cd8f2935009c691 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a927caaf97244159aeb4b6b310615666da6515a95255314815ccfa576c3951cc +size 5496 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/config.json b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/pytorch_model.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e78b14a6f205bce6f885786e50c5fcea20933654 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f6a607fa38768afdaa76cc9936dbfc077618dea0b602b4ae7613a8f497aa13 +size 498655278 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/training_args.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cdc5e64b7047c31570b8fdf64ce015d16d05cb22 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ef61ba4a874ce3d36c896ca9909e23b2eb256838c8d59715583f766d0c19cff +size 5496 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/config.json b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/pytorch_model.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e78b14a6f205bce6f885786e50c5fcea20933654 --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f6a607fa38768afdaa76cc9936dbfc077618dea0b602b4ae7613a8f497aa13 +size 498655278 diff --git a/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/training_args.bin b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..46bdac247aca63fe7c028bcd0aa397cbe77b82ce --- /dev/null +++ b/finetuning/roberta-base-offensiveYN-finetune_20240429164547-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c578f3c4b630b9d2181132b5a14f7e2550f037d24d53e5df95cd2d1c374fed5 +size 5496 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..12f04d4e54a64eedef1e9302270b40fd61db4263 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d74b34a6c27ee97853b4eca57477e5e876781dc6852a11f1bc3349f755d7402 +size 438031054 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..f6f56039f96bd4da560bf56d4c0dcc5e5ea494c0 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74b64641a297d2baec80c4ff0eba0a05f6daa65d5924ca9b934646e1cd0bb140 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..980b6d53fc55f8924ac7d0f0d74687ad850ea2bb --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e910cabe55edf63fe49dbeb5f3e19aa7198a6cf9df4fc83600993bd54f4bb11 +size 438031054 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..9624795cfee06b6ab08b39240f0c0f025a4bea08 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5e97fc9654359d1dc07e0a6aa9b24db5c0a9a4b63c734415161acc49d1ac981 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f26a079d2933abd2d5a6802d581106f882b0bcef --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de96498f9164bbdd82e78bfa20e6e3a4cb0c288d2c2fa7a77ee20790d9e24cfa +size 438031054 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0cdce606f0369b358bbba5cf2c98684dad651d74 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f06fa831db5863418140cddd09b6fccc31620eb12ce073e8411b24147df5c7d5 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..8cc50f372148febf404615826307e6616f39fb3c --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f126e54a8be9a89f5332c3223cd5eac35408c5d63c9193f7244431c77ffd696a +size 438031054 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0184d9b6eac07c64e962ac08b0183122ffb39155 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b62445a9a6d2f2ad7c2485fad9a4d82f593068a329cca7e8f7eb2c5a34268d2 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d0bb60d321466012c99a2ad287db9b5156fb72c7 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e080caa6882e2d541cad325aded195de99078017bf90f4a3d3033099a14534cf +size 438031054 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..02aca14b67daa850c73f299ce2e3ccfbdf220706 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl-no-aae_20240423153935-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:549030da10d48d45d953a9a114a3e80dbf3780cf7a2d87f0f82788163aa66512 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fe82abf0744957aefbbd2e0cbe94a1ff5063323a --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea16dbba2cea2e34bf3654d3e70c375f168706351d4c24d273e4997390439449 +size 438037846 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6886e13f4f3e01488e83da5efbbf94d1c5612851 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:531aee12363de1c17ed936b801666f8ce72deea7273ca659225cd26820583b6a +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3b8dfea655e67861f6b5cc54766d4a1f0096fdad --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8663288d9d51884322ebbe4e6cd37e8b3b96437923e98f678a8c9f608c35bd00 +size 438037846 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..be2dd57d4e73fc4db7ad2b8f0f2eb68e713190eb --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4e9f60845d59857c2f20dd4ac317278edf25763e712ada909d66ae794999085 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..40aa8e76bf13a0ee4526710e1486fa18a73bb1bd --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d2ec875a1cddfd8270168a1938866418feeb523345f6a5ea4528d1da505ae99 +size 438037846 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..370cb2ba3a04446a926798239ec1e447113dae1e --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32392211a8bba5b65613141a8dc6c2bb65e1b6667927cf8275cef4d875b511b4 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..68c71e06ecbbc913c689fb17fbee0e6ca6402c41 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:334a94c2d1f96d638fb5e96cc5283cbfe98b78bb9e44b4e30358ecfad2797909 +size 438037846 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4f14b4ea2cb48fbcff0043373165392f30c85280 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41d944ae6c51a4dcf965a0334cab6b700d848e7ae6ca0140b26b6d796c97b377 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/config.json b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..5a18fa000f03814f4009da81b2a8ad13da63b6e2 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6e8829122deb3cfc50876090861d78d85b56309f91a92cf999e86b9e4b89005 +size 438037846 diff --git a/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a71a71ca83dca2277d6c98330419894e411369c8 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-joint-mtl_20240417094846-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d36208d42b51b35a8a4f478e2beee305b76c207cb3cd29a809039c3253a7dc9e +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9469222688b5fb0d88d105d0da7f2ef7fe19d5f5 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06f3f525c8bbd17b020e0c737f3eef24ac653f3af184da4aaead4be825d7fb50 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a1dd29af5669329bce4ea5b1500119dad392d552 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a83dade4050051475f3cc611832ad7cdfc2d7bcb10202df8f5687b3a8cc2d854 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f8f2828428e85189e351e1fe230726c1ff722fd5 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd3933cbefaaf34c6f1ec092ca9ece5775d4e4c30a08747d0787faff6e8c8ab0 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..3e8278122eb15fa05e20c9397b8e9859ad42d418 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebbe3248df34ce84b63c7f74e864f1a79849c4572af33855e73f70b7c22906e6 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..dc79184835aa6e93cba18180a8b71ccbe9c60f5a --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:448335f99b16ea27a2a98eac7b48e5acc3fe8762562067d490c743af079feea8 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..9eec67c870b89ea8a47b9c43740dde13a98175c5 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8e8685aa7d801f2b6e118fa9f4f286bc823f38670b3c2a065d7a5cfa2ed65ff +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b062885bdbe5dee895c4893dbb19bf6b6d25cc71 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a3d6145e81931720e3c759f3a1c23f5bdf1c48de1badf0e46ff458bb4e1a47b +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..926cc00bd694b9227e45b8bcb2ae98892574b539 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:125f8803393c2292761f854b4bf250a21cca90504c79c813022ee08729d8c929 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..a8bf5f0415876f9bba0b8cb31c9b5fcae05c802d --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b28f14fa698b63507a034c0503e401c6d8ecf5393456f231f827e85bc5659999 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c52ee7e697e301ec20b957d9eec0e258ccde03b6 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_groupYN_20240423154036-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b232015d5490c6e98389ca31689fd643c5fb09c42702c45e0e89acce97e675ce +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..a76b67f4f0bd73409333bc1f2ec7607464ef736a --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2451819fe3dd4d7d10b2e2ed76ce91e4ef2ccdf68ec24cae1457b7ae6eba11c8 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2c79fdba8ba785de71361bfbc9b123409944b5be --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:483e6f4e0c0fef1a4aa4d6512b4b0f253731000db6acbf1a76ca6cf92c9c082e +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7b6dedc2a931ae06043c6c966167cbabd2eb1f00 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6972ff0f36e834b0b9fc97bac404e2d3f11951bbf84ff12160a63af7095f9072 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b127481223e790480b1942087c9fb55e04939a6b --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7ffcd5a2a5432cdb2b63ae472f90524554d468d4b5a2477502f5e60c492fa6b +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d434789b1c5628fdde55095b3c10388f9ef3261f --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84eeee6847136826e0ff90be2897c850a28f81c018f49395f9d26d2e32a80d20 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..9a30d4f0553cc167d6e588ab90fcedbee52f3e85 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29d16def7e66b36d8cd6ac31fc4300d162299368398dddad48315acc1aa19ddd +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..27969a00afcf3e90a3b3184c1ef504f228b69760 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:516480944a975567c2e2cbec1811bcd66b720f33af175fb1352d3d5d391d64a1 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a35c88676d140a4c499367cc59c255d03fe6aa28 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d5bc91fc8db065724b927ab579875c8b4e73bebc2c12090d16586015e744bbe +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ff390fd596ab11ca4ddba1ad6f220024f40caa56 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abcc8d332859a8d3cc3fda157fdaf7178201c6035a073e248c9b7b393b792f97 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..aa698a1d935d437f89fea815bbac7b8c3935415a --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_ingroupYN_20240423154036-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1debccbd6b580cb7b28992b2687dd751e7f0f93ea4db3e036c5c90db914b577 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..54f1aa065814cd0600de9ce4f39dc9d0f8d64517 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56126ab857cedc831e452bd9a3dbf14f048f102450cc65478f75673eaa2703d6 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c88c9bb8412bbed50b032e88d377a92bebee410d --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b93d56eb9d79abab9c2c0f50052aca0118389c5c45097c1c0c69540f44ee8acf +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..2569243b8f13a534019965d7c5dcfd22efa56ab0 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de6872c8ba6cf0f5933987a221ca747d24b2b5dc99f761f6e9de14694e8cd400 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..af6c2faef9157b8be876d93d9f02fadbe74420c5 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba31c39653214ead9e5e8db630638f4cb0926c0928eb4715d3e8dd0a0f0a15d5 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..a323403d0bcff22e916bfb3e16adbd5ef984a345 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fc4a8682b9b5435958dd7ef639a117d0bde126d85cc0ea7fcd98937eaba2824 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d214f3519ade47c264576abb19e58c948629a055 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e66aa4667ed88d02e251df2a4155813b82d41189b735221980232052c93f88cc +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..31891f39834a03f8281117a266c2f49d6b6abdeb --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:595756bf447687c1d5e72df5fc04d2670343f1e303d9e1bb240e0cb6ed29932d +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a0858f66291daa9d26099bda1dd47ebd3da16bc7 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f816b26aa49e377129c31e52ca37ff15891c8f2077f1398d9ff09cff42f8070a +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..603442042c191430aac76a2df04e9bcf0e90fa98 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f195d5ce2d455138ac1a7d591a403f95b7f01ae448714b07459af1756bf69de +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c368f23198dec4906e84fe69f63453e6aa4799c --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_intentYN_20240423154036-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e435c24819030736feaabe6605bd7f9195b28aa408c331e7088b12cd99f939b +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..6703e591006a373bb9dd6fca2c8e885983d8a8bc --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2609fcbcfd7af8685b955b9d69b8624fb8f7fc31f1e8c08ff34f94ab8c7e4d42 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..bda14c611e53c1f7db04b3f35980b2947539a842 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88ce507516ee9d05862cbc8c6e1cb5eb77b91451aabdbd3fc374244540fdc1a8 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b6df2d31f6d66a8b5f9b3076c9e4f8c799fb4dc7 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70bf8ef0eeedf29603be9746c9451835e937423c4e77a605954bcbb6ca7b532f +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..75f46f37e763dfbceacfaf4e89421d2d351094dc --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a69190725210dedb257d2d6dad534319ba38137e2ffb4e2841269cd5302a42f +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9350f6db9450de7a785721c517b0b0a89fd5a4a9 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64a95c137cfb4ee8f969e6f3c997ea23bd4e0e6c53f92e5db62f0840e2529712 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..bf5745c626e1322b5e22379028a9a0f14936aac9 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25c7ff14f51065379a1bb73016b4bd92491f06236f2c7ff084f7b65464319cb3 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..1bbb999aa8d677a28ce0cc863a18843971c09001 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4723629f51f3ce2d86f203d4b980c8f500d7f6f37b1826334829a7d2353c46f1 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..bd87b6175739a44aec5b9a61581edf60b97ea613 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a37e807890971a121f1674afaaf51253ecc9fea6b3d37d1988c13469e92aeca0 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ba6923c22a6555cb541055f697ab731485cdface --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40583dd90f9543daf3651846e402ca581de9ac6949a742b09a5071faa34b223d +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..62b4e0facdc2890758c8c15c4d8344cc52736848 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_lewdYN_20240423154036-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba5717f7027c8881dee578edef3855dcfeeb0a00224173c38ee166d7a421051f +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c4987635d2972da0eaffd08ad8d3de9669100463 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7d6cdb8addb9b403989304b5bdb6df63e494f51d3a4bc4424797744dbca3c05 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..8af5ff4d64d6093920d04ecd4aa13c20e4ca4e3a --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf717fbe7a9a546ccc32e5f3b43224aa7dffed446344ce08b62347a1f6318a4f +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..256cbce3e513d0f887326e3cdbc9a703afd8a3bb --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ace6fc4d109f89c05947484915bec28c8cbc50a29a430290709bf4e807d1f970 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..07a88135cf0cdd881ec44ee1c1a87af859bd5dba --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a56d8f101f037ad343aa5ca501e8f8b8321c64a9c49327d7b9fb75e00ecfec5a +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..55f405fa021413ee3b429a0b7a27aacc5e1a641f --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9598a547ca84518043ae3f4fe9db216055056b185fdbd603fad88119e179451 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..247100be1e5b149e225383f369c473a1d450f47c --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5cfcf9963d71a700d38b1ad3c9763597aa168e8deccd275be99ee63b0c90b974 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c2ba0cdfbfb100e0a2337a4b02b959927989158 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:683906dcd196e8f17c9d16953e65fbaff48406aa758fe90df473b8775e952c05 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b08dd938595a1f6f8a428c09a7e5a18b55008a06 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a88c31d5000a69c7b451f4970dcada51fd86f57a92eec1c714d7d5048013312 +size 4600 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/config.json b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c003dc186da21b30d9dcf7dc0b7d48f7f4f92e1 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "./model/bert-base-uncased", + "architectures": [ + "BertForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "classifier_dropout": null, + "gradient_checkpointing": false, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-12, + "max_position_embeddings": 512, + "model_type": "bert", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 2, + "use_cache": true, + "vocab_size": 30522 +} diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/pytorch_model.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ab63f4ec6322594ab2db579f9e50868996713f5c --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:842cd499cbf66035670fb4ef09006e5a1e31e14edcdc354628a7184d01aa7990 +size 438010742 diff --git a/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/training_args.bin b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5dd27a4cea406d0ffe3032d76cd7bc379da92954 --- /dev/null +++ b/joint-multitask-learning/bert-base-uncased-two-task-mtl_offensiveYN_20240423154036-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68f0ae15c2edf74e089a3faef430c54383331cc4977772bb14c1c12feba3f77e +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f0a223f0ff5db23892fd4be23723830030a30fcf --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:049bb8e14840e32c183994800f53e17a6debf3c137751d61d47f6b7e0bb4e2f6 +size 737791954 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c1f2f638805969e5def8edadd00784722d761b21 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf24ff130620709526f948e74f0189f3b31ba0515fe9bfcaf8eab8cd7467a788 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f829a09b883edb5be4a178c3afc29c41350f3b22 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d55ee86486e28d242bfebd2679d7a131ac71fcc07efa47f3dbb3ba4c0a8e036f +size 737791954 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..512e0dd269517d5d356ff505bf6118f3bbd95ac6 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b55a54ae189740a5d4c909890e8bf62ca2e9102beaada4e1ab7124168718f69d +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..914660ecc7b12434c26791ed9cff0e06a7236062 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88b36c5b7673c1f1381b5bbc75d719f2754f22025e50a15f2ad0bcbb122d36a2 +size 737791954 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..213a7b39137d49d51c09abf4bd81144b33677b65 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1ac775048d609c5c4e6b438c0528b658fbfa2d66657b6650cd649f3fcfe24c5 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..08108c80d7fa313285bd2bcea3bd4cbeb8dc330a --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2dd4a41be03a91ddb01bdd928d35cbc813deb9183de82b1b8995f7a5f56211d +size 737791954 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..63c12c15c5f7393ca42be834806a427d394138b2 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd6bd49f4ca19fe077684c2cd88f4ad041f8d53a1826f1a98c8f805d431675d8 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..972511e5307458bbb422ddad1aa910f77e9ee046 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3350799c19a8ef41b37d2fad12900852be335421b447afec8bec4de43a073ace +size 737791954 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..43206a8d462f3e2a3c7bc7ee20779a6cf7d0367c --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl-no-aae_20240416085847-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:148f97dfa91e11356d09b5c624e056278f4372e9a183df820011085097cafb72 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..48b154e7d53ff7bd34b4c3204770861284b28de6 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c59a6739f74998adaf8b45ffb3ccfbec9a70588f0ea2e0b34511c3a1e7a07fd +size 737798746 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c226beed6ddece93160af7d8e2e676fad71da27 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3adf5ebe617582ca8fedce16fdff7a0ce774a87d79a1a9a49a8e3840a8642e44 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..5c79ad40872e445e052b6d55fc86c807e6370bb6 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10effed9de9ac1d94d7bdd19e3c38c0007fd5a982367c5a392b3369daf282763 +size 737798746 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7958a82b2c30b53649d75953ca353eadc953e28e --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8711fea594e1d89aee25cbee840c3e6e4f27fb455dd97b3ac4cdbb2c4cbf9be5 +size 4536 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..66b5b738aae79bf74deed553d4890c7742084128 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a805c5f91305766a2e599b11194555f35916ef1836213043a479accb7eaa4186 +size 737798746 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1ff17bfaeabb5cfb00878b63a9cb35ed1e7086f0 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d088b6393453aabcb1d1b4acfcecc04f6c17a8eef6b90a7408f48696e527969 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..098903dd37f1a7fc1532eb36ad8d4cb62f87bd4f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ded87f822ce815031d782389802a985907e8c2a53e0423038ae8e09cfea63956 +size 737798746 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4385f9c9deeb59e0f973261a85defccb3b637fa7 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca988ad1d2ce9f1f48179ad0793eefe72dbd13ab308942111300ac9b22bd0ce5 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/config.json b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c5ca99626a740110fb0e2b60ac96e108e70aabd --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abec656825b09964d1f2a0b8cee03714799bdecd7ac23bfec925ba342e3de2d2 +size 737798746 diff --git a/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c6661dacf8d1962281c16032a74e787549d4d0c4 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-joint-mtl_20240506173425-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:675fab02a4d95f1b2225bd9e4e83df07448cc14a67363fd6e91650be8b82162f +size 4536 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7b809a4b25200000ac3d272c1ef944e0465f5d2c --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:adfb794643498cb074e1585f9fce7b62e006bc71c797947e088836787754eaa5 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b218e2555255a1927c61056c964743c270fa3065 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:faf1454c27180e2097ecf979bf6ed0db46bfc3270a2a996f5a9b8e4b51feefbc +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..be212ede465dbecf2018b69843e8ac93ac23edb1 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:884606e66b29588ae60572311d7dbde442bf32ca5c52ba3efbd0d1120f8a4efa +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1587875d115d6ce1c692be7971ee6386a3519402 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28909912020349d5613e914e3223ecb6a9627451a218f439809559f7ace414b2 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3fd449713e80bcba5c0f74853304d3c1c188cee4 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f05dcf5dc334fd83710391c1c8274a607b090514e51bd368cc9d88a00e8ab50c +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..99a80ccac2e353957bd7482fb0d5d3953e56dc8b --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93b7edfd6be20fcb3c563715bc35087ee6a5e290bbb85ac29fa07e4533a2fa97 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c733006560739eb7560beb8506e4f3a630d2853 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:798316848ef90df708561db1cb9a08ca9337222de32149de3d3e5fafd2810872 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2c3feea3637cab79076e8fb4ddda8919ff5a8e33 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38adc6a41a5afef2b6639d3bfa70aedd72503275c33da91efae660ba41bf538a +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..714fb28bb688d87c004319ce8a69f56272533d27 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85606c82474414e4948cbf72b27f5af00dbb222f2531afc1d988ee7a8af52e5a +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d7020aa6ef880549f9282d023a0bf8eda4500bb4 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_groupYN_20240412175642-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09057ad00d5c4f50502c4d6de2e27dc32cde2d7bbccd5174773d37d52d8975d4 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7f66beb116b821e24959ba1b9e9843cfac93334b --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:850a8f60133b78ce9d0c1870ef161d91a0fb54b223f200f746b411349d2cc9c0 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5979505defdb3efb78905e29b59b66e102b270be --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a62c3ffc4b70ac3f75545a57d93c936d7b209829044ff8c7a37c87ca7980fd8 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..201c95bc9f672f960a896f10135d1c84de387003 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36a7f7562851c15032269e5325926363260712dabddc8824ca1a616cad7407b2 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4932e1369a1decc0cf5dd9d4d53e38b23719d0d2 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31387dc03e8212c76b3d099b256c3944405845e5147816ea03eb6bb599fe9673 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..acfd750d1a7c526c50bb130ce13114f56a60d89b --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bf3fc36fa1aa24db9b119ef22ee098ab904f403963b0a0b3df3fcc37532b742 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d042147da96c861a1e00f8f86a5908d4a226d693 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d471c2c0f5329ada878b17ebec4e5dd3b716045d2f183ff12f7bc1aa2db3865 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3fe48d433df6ccb65d6657f6821ae00510250220 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfe4604934cb2d6425521551a595433a906e3daea6448df501b65c2d50b196f4 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7a79e20132b234eafeb55f6dc4eee78bbd42f995 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9419527aed7d0205c93fd8e743d137dfe2b3b2f80834c10f6365c774ab7f3a71 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..37bb967ac87a6b21a85b441e1af66e1f48b818ca --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b77576ffc20629922f2dbf08b32d71968830cf407e74335928d2ce0510df7f7 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5c00f48fc4163665bf627967ffdcd90d9557062f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_ingroupYN_20240412175642-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b3ebca46cdd9a3020c4eb98c64179b84b22d5bfbecfa00764e152b2dfc27b49 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..00350b102e2197cf5d112915d8a500c253b7270f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a469274f4d3736d0b7f3a8bec1e7935f66fdc56768813a5164e8720a6689982 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0599198cf3ff4b82c3f3b63de250b28accc8158f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:873fb02ae57794e15c6b9b83e98f311846b0928423f9661341d6c93f338b24e9 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4dd4bd2d6f3a6a54cc6546ead883d00f098b6f0c --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7706ebead9b62da38c8d36a55c01e956d600a532c24add0c235a2495eb4f4b0 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..f98602d6b8c7552fed9aa4f82c41ec63e838a27f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2b8e5316f89c194bb37d096a41ba3daa5bd47e7038ffccdc9802a7f59d9deed +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..404d6c4a7fe8682ffb6bcf953ce16f165efdb09b --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ac7453abb7fa693a9af8910537e199b73d3831db473b05aba33f2e2e256aed +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6156bd33869e132df4dd6493181ce671d162c6b3 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37f0456d9c3db26ec4965f9a09c785e68f7a27a6974209d12305897a9b757ba4 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f4320a6139b4997916e313a49892168e33ce3d4f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67413810909c0ba23944980382e3354e5f78aa5efeab272f77c7526212f27f70 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..63a95d9693d657d9a8214212c31e99b94a9442ed --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf3fdc72d60c58abfa5e364f4f8cc632fde7ee490115941e865c58abfce5fdda +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ff5d5f23be2b8cd1cd7ffb2bee2b97340394c29f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a0f3745e241ccb9fee4d74909e441e00a1e947abc107296a590a13841afea20 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d1da1b468d36724454214c0b3e7a11ed68e22d23 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_intentYN_20240412175642-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cfbf5705ffe106fb99c930b9c46f2b951d44159a9f349d234103183e6b3d5ae +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..af6a1796250e19d46200e1acdc8e748678eab08a --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a386b2bcb74360592fc2d11c5b4cd5493c30140222cc67aff78b0bfbe175262f +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b2cff8bdbba7745ddf878342ed85e3ade2e2c255 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:157861bd896b0432bcfd3733cab5cf3ff5a835082238af00c453b93f2f2a093e +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b100d5b9d34e3aed8782fbd8a1bc2f88f2e9c15 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4dc6f2037fd78176a3c7841b4b4c97b85bd783b6eb220e165fb7bea5bc815302 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2eeed032bd904dd8139b18d796567eb3c1f9dfca --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:116d37061f45cd3509fcb2242023a997d678187ffd42ca8c007455d5cbacb7ff +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c103b839cf6486356c1139c01987d5acc44a79cc --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b73618508d90b50d4d15a44a906fef3272da251bb3acfe2686f6eb8cbafba255 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..52b54568cbca889c155e5aa5545eaa224d73b0a9 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6f83a3dac28f59d9072ffcdea5b4219e8fa0fe575cb788c1b787a98ed0fe254 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..873ec0b21dfd24f513c03fcbf468c7128292bd87 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef911afed9056db98bef8ee6e115cee099f4830591b4846740a6b2ec7c4cab6d +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..59f577e931e03cbf52457fdc39dbba03efb13dd7 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83362f7e2cdc02f8fbd63e9102935439917f408565436434a06e731cbe63e62e +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..72f1cf29cf3e8c8828c0d5f623391f8e6b48c753 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:195ef400e1b9679259aa0a05279b0fa73e0e8133c8dcef4bfbe13e77d2aabfa7 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4779bc10f59483a6e1a9c437e0b12de3c01030cc --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_lewdYN_20240412175642-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f30e474c39b52ea409a4e91c12ad81cbc33bf6137f125a2217ac2a61474cecee +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb57a2a2690b717f8081e9fa6dbf104c6c48b75e --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fffdb5d431b831fda3661137a29192688972f3b37ed677ad308bad83246b3ec +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2d870fdc4ffd136dc9f3ad9fc8a736bc4d868157 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:944f618fa503bd8da3a46af178dccc87ff720097feca03271f6439e6d669dee3 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..8f079a6ff7cae20e92a165fd58c17dfd2699743d --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5515a26b9089d8f48acebc04bf82849880e624191a9e66a6aa6d0fa3c3548a08 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..f6a363b9392df1378f90a0b32ef68a34c69d327b --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e4bc81161f33b71b4317ea63ab5553931bc8dea82cb15f9a31e41df486ade04 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ce278a23b8712219a9ec10e9f505b4e62972e90f --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:607964b98a0cca883a5fd5473d32cd40e26656f2b1ee3459424f60985007aaf1 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..aeecef51350976b14996cdfccc53c353e326c14d --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60ce16c8978592b4994e6f644e666e7727113da97f0bb520a43846ac7fd4d653 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ac6f2f9308db65eff49f8114e40229e4026aa2bd --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c8fe5da290e405218d4159d9098371becc4a77a1712a686fb3f12a43d16d222 +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6c9b92025b389bab8a7a15a574c85b5432a7aea8 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b35bba790f8ff915ad1e351f1327c161f63a9bda0e9f9de7e58410334bdff581 +size 4600 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/config.json b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..abcbeb8f20ce0fb5b3f3ff0bd16f53364a435056 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "./model/deberta-v3-base", + "architectures": [ + "DebertaV2ForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-07, + "max_position_embeddings": 512, + "max_relative_positions": -1, + "model_type": "deberta-v2", + "norm_rel_ebd": "layer_norm", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 0, + "pooler_dropout": 0, + "pooler_hidden_act": "gelu", + "pooler_hidden_size": 768, + "pos_att_type": [ + "p2c", + "c2p" + ], + "position_biased_input": false, + "position_buckets": 256, + "relative_attention": true, + "share_att_key": true, + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 0, + "vocab_size": 128100 +} diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/pytorch_model.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..1d471ce657f19c260238a342ec17c21b8dc2e275 --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1215b0ff462246ae9ca2aa2e67fabf6c25f13a5ea73b4e545e051f7787723dc +size 737771578 diff --git a/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/training_args.bin b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7cd444ad4d49046d2941c984fde39dca96a62ffc --- /dev/null +++ b/joint-multitask-learning/deberta-v3-base-two-task-mtl_offensiveYN_20240412175642-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e9d536dd308d0e611cd55a908fac40edb24f7105c26e4c6cb1f2ad85a8d8f7b +size 4600 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/config.json b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4df7b698b2c46ae98c5e630b4a13485fa871f597 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f983da6c148926c80efc652ecaa560c3cedec844bbe0f2d363b110cc120f469 +size 496322246 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5645dfceea311b62ece31bb11bf264f7f8073941 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c39d290f07fe394ad3605fdad0f15f5a94206c8ba542961bbf1661089d4dcf12 +size 4600 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/config.json b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..54149514011bdf675ee3caed31021a202e4f3406 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48446af0b79c5213961073b61d065cd45994d3946d175603fe6393ada27bad11 +size 496322246 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..23f49291ede1536280e96e48198e7ab0546280d5 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33d1a32848b3877c4310eae73469a6758f1c74d889276b23676b636d99b0c792 +size 4600 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/config.json b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e9d119edf714cd83280c504d01141de2976f00ad --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b9785d23b2190f7362f9a1983faccfd82a7844d289d2410e3b02f64ea1ed90c +size 496322246 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..fb90417cd9ef3fb78a135dfad9053a0575389841 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7947ff940e43a6c3066ff610b9a877431b5fe38ed3b30877327f82618f1d26a +size 4600 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/config.json b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..279a1c7bad1fa671f7f07ce5203c00d8495df771 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:409abcfb0b542624d4c84b1fd0a4a74e8815388b6a26ca2c98328785ccc4b033 +size 496322246 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..2f64175969dc42ada4831f34d4369c50fc75d6b0 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4baef924b5a9cb8bec22c9d14c19e0f414a87223a9d2199f2be2fad86aa53ff +size 4600 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/config.json b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b7898a5b6b15de0f76c88b8db859aa9365f9f305 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81cd4b8d6eab397aed5ceada12014f045e1483f33ee49d0aca9d19866e63d392 +size 496322246 diff --git a/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d3c26f8d74ff77be060492c52377b98437942db9 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl-no-aae_20240426163535-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:578c53dc2f6658917f8b5e2d2c88ac016bfad221d9d2fa9d24f4c5b76dac27d8 +size 4600 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/config.json b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7637152b9d1b191edf6ff9e63f13843ee429eb77 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a073e50c303e7b8396285a021b95a8583cbfb10407b9fc3af0638823710f0c74 +size 496329038 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..462722cf3c3cac14b6228d4ea5147425672b2f6b --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fcf74193b56e4ef9f93226424e49db43f9a1b8d7aaa7243f9e564e192e39db1 +size 4536 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/config.json b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..72417278220865e3db94589ecfc4130c1b709023 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d8e39580967758d043c9aeb874615e11758271a374436b27744c84537df5a65 +size 496329038 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0687de25d56b04d692b538ce3130c2f091980ba8 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7522bb37ce1145f9de48011ddf15e8c1532ca1d7d682461cea80f3daf313ab2 +size 4536 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/config.json b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..21b515cc8dde0b1c7c92f14151f9738575e64a52 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf9756568161b0ec67e3ab5fa141b6076660b763a0bd2296fd641ad224966dd2 +size 496329038 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a97476c62e14d880fa7d60994ed8130c867f98ab --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ff4abfa650efd88f8da38a1a838b0c8c8e61d00dd17c0ebf61c1e3142a20b9c +size 4536 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/config.json b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..1d77cc725197cb248ab47454e490319e1cdca553 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70dc5a57c216c57eefbdbc3df532640b746a439af4297c2daf2b9d5f6894ed94 +size 496329038 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7e064f23a4fcdfa54de53e2a7b1599157a6d580d --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afbcd196205e42e323439308eac21a5c4dc42f75c86dd86781ff9173eac691b9 +size 4536 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/config.json b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..dec9acfe0f1e2ee2ef2b63ff5ea1d8e60c105f00 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c79d5aa10c70806bc3a6e5c9352c59530ac4d9f9ccde1cfebda5b7119e5a434 +size 496329038 diff --git a/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/training_args.bin b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..95ff1158d7d3bb50fcf84a716c541a6a8bd76044 --- /dev/null +++ b/joint-multitask-learning/roberta-base-joint-mtl_20240429164548-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49cbdfe5c1f4f6ffabba488dab3817b2ff837ca9011c96334a41c295b298d265 +size 4536 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ce3daaf6e8890489a1a8e985649c26b2fc38c2f7 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22aad40f0cd8efa7bc633a23eea59c64ed6470003171142b9fcbad7e8da95dd5 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..14c9aa0b1c5e0baf0c15fd3aa4c38694bff217de --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbee15ea120459f88c153edeb6bab05b9a5ed2fef7367b2a41d13f1c4348604d +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..002286d1feee631c3b452357a93a7003e6c2615e --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a40eb67a46a45f9817a69b77fd853a2297dcf1930dbcd05d62b77edbd1b0b8f7 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..feb1d2decd8ed670c77b407ee3d665a59076e91f --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aae9c1cd4d8d51d7a2ec52f0b82d534eaa519c0620640d15d9cc9465a9ec0e0b +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..84b5a1779b7a8199b22ada1b2bf7414da8f7e342 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:655f4cf2e17352002c183a538128fb3b18ce7e48e60cf39f2b497e5b955a0064 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cd50bcae93d216f5379a325de0461a3ed6106c3a --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3a03420df5f5a10b23a8edfd459c998410ecd392b92468111ee67796c94353a +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d4772a143b6765e14bbdb244e57cfe35f9f966ce --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d38f82991d722d34c7af876b16ccd73019e466d59f12d2eca66e31b377a8917d +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7a3157232bc77e8956ccd4478a25a94db0e4ee6b --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:faccdad66e1c05fb5715ed0c9d356b4d8af4c524ad65a2dbeebc784bdffaba23 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..2905c39e59e066ab30f413daea5bb3e022975ae6 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:127e02c2e1fcf0b739ba738b784283add17cb2cd0ac7e5f077e96fb35809f6fe +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..fe15713cfc9f52c2bc284722e473d2ec1c9c7f01 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_groupYN_20240426163520-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23c0c6c68e42f05f243bda883c889132e6eb3d7aafb712b6724186656ee505f9 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..805b9bb2638076c0c7792c0b40aeb059b66b9c8d --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea786c4baf7494a52d6c6fc50bdf3dd0d58dab7a34ffd27c40d3852c6d9f7e3e +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c97507c56c2139b2b257033d09114b989fce415d --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8af6bb2a28738c8c68aa015753bb21caab433a77237d49398b49c6b5834e23ca +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d0830fdea4923e33175dae0aa972df785145544f --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7b67fb132ed8f0912d283c143705c393ad83fd61d65435444308138c7807f21 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..d8c0ab2fd80d7d36cef888d43fdb7e4ac50d8b16 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42a0f93d238318bfb4a684815f00536a4b00588b687f13ac7049664d764461bb +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..121d7db419dbab220dc5f4bea8f3fa5ec02a8fa4 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fccfe346c6ef195338e35b94a540759ccdb2efb3247019bf291b0b10b5923fe4 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..a1a3a2de5288ba33ad31df5c2e794b5c7fd9144c --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55bded375829a11a50d4b97d0e599ecf56a17df2efb770ff68180b9922e9b8d1 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..719c0a79e5b07cc2e930839e0e0e8ce386270f4f --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8677b9e300f40d3dc4e730d0d217f2fa92fdcce65923af03384fdab44e2b4aa +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..abf150b8cb756f59899bf060d80c908f5f7619e0 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66f67771f2b5201dc509c961694e4c1788cdee70ea643eacebbacfae9e01b42b +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f9753a0ce9d7e4943c8ae2f578a07b10969e6e2b --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8db548955c32f362743f9f7b7149cd55fca8515f5815377c6bd6348d943e6da5 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b54fde7ca016bc863d1b89a5c20e462e5a10604 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_ingroupYN_20240426163520-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fe100447b3d1654a5b207c439495f8915d8ed9746611201163579402d9bdd1f +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3718374c2afada044b089b948d15f3d1d3995014 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:545df4bad5dfa337686c12672b2e69f7287b1254bd40c427533cf35ab3b96051 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..49d835ce3368b37bf9a670e599b5eb038d3248fa --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2d0b8122514f416a7c5585893fcd9e7260d8ac3a4affeec7e93e82662aefa4d +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..07866049a2f45358a76e6997e7cd32c77fd7e30a --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe4a8702205436bc05c78b0f9fcd1972af1eb7441fcfbc718188dfc1286bf145 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6b320cbaec6aa04f92b872989cd1c5ba7b77d831 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47e3e8329ab332c8dee8460b7b2d1a2e4a1b1f86dbcc6c23dca36734b5f5eb69 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7b25935830cbacfe02f8882afd22c22ec4405acf --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc719063e8f07c6c10070c6b66e7e7dcb9f8ee99d94fe06cc6429bf86ef797b6 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c2bcb59cfeba923bd9c8491ff95187081e06732 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd38e99d486db0dad820ae702ae2f7a53c8f0d65c406bdb21fad29d65748f1e9 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ba1ffda93cd1af8678bed59719364a2acdb4aa8b --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7de9ce393594aa0f0d1c40e468dde9833974bc56fabd884e97a0988ebe102c36 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..e0ecab13df8c7b33e1a17b188042f84c7ddd5b5c --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f35ea755cf2ba6fa791d1b5bf87eb52fe0495c8f18ffb69642b440609234667d +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..15b2c0235db4d45c012291bd788d4b5d610937b4 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f462458aae49a66e9c83df870866384133178571e13b6762091e56e102e3498 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..17a391515f4d467af96323eef28e6d962d344ed7 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_intentYN_20240426163520-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9c7530ba7bddb5161cc55dba55f9a937b7173260a4674241fe9fbb4e5da2800 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..5156ed96c1685acfb120bf4ba282b0123235ee25 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fbf445a314843f02257e2eab8415dd258ad83a9936d4f58d2b1d487c7690e66 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5ca84b02b2b39828568e540ddf730e5b0a46f9c7 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b339857cab91ecd8ca28b40a5f6644839d32c12d35324e1047a307cbda5cf775 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb03465a8db445804414595af0d4e7fee78598c6 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9a7e2304e3bca3b1aee93d5cf50be532fddf3221fc15464e935bc5321919d68 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0ac793608e81bcd3f8ac909b512ac3b935728cfc --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac5eb18080daef82a0e5ae5194e929deeb909460ab90da1c9dac882fa5a8044d +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..908ee77b9f2fa75ac45dca128b3d84983a224ef0 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a4ade5b773630fbdd6861927767a67a0e86bfb066fbac813611f2eef8e1ad98 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5657719112b832c1470026c886b5cb8ad36d0a72 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5deb5070660e27017b870f80fa197b7b7e1e8062d7eee0e8a48124b952478da0 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..0f9d295df64dfab726b1de77633fe712fc5d0327 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:301fe6c2394d7339fa4e73606c3cab3ae441a9030ae1d06cda1336ea5a53d138 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..72afbcd5712c6a5ca6796d4df1771514322cd6b1 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a1de20c104a585b1e99a60bcadd15299a7ed3cf9b0cc0c5658fe10a7e1debb4 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd01be0c9885c46d798560d1b8a9c9a136648e2a --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5eeef9fd41a1f7926352ba4d633c894b9e5c654f426a4b2ac87a63fa0f88ec15 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..7498fd197721acac045262932a0f34ce584c4ab6 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_lewdYN_20240426163520-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:749af74e2061caa395f0c50cd91f94b0b3a1d729ae09ab88aa51014541863f98 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d610bcfc58b9e59a448deeb461184a010b24fed6 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d7cbe156ef6e8a6fdfb4976b721925e4ba22b3eb1a60c6db08810d99e414ae8 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e5f4090cb04edfb7ee78ca3317da6c037a72021 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed1337/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d51021139be8a305b59d45dea641e6d6c5d935b17b099c5dc3a96c0777f93da7 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..dedc3c71b33775ebfc0d6f758c6eb7830b9d6ffc --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4df71833356b058be295ba4fda6f5fea80ce04d1eecb85051ad600a7812fd0e3 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..1f88e24782237efae5db2ca58058ea5b664eacc0 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed23/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ec8d24f03c87fadd7bb5155893ab92806a7796838daa56f9c61d7d0399dfd0c +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4827db38686fd019b1b15433f22e61a3dd63aab2 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d34ccffde858ead1baed99066fba9c2f4f3e4459e1fef7227d5829aeb81b9b6e +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..bc0a8a03008013a3463e58f8057fa75c15c00df5 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed271/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:900770157523dd07b002d426c8632561535cf9af8b74c3164bd0c09ebc17614a +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..bba8db3e55b180979776feb47f48c10d182efdf5 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:823d880151e45bdc72fd7db29f453638850af4b68c2470c346bbd1edad8235ff +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..06bcb8b0e009618672e1e99d583642a75f1c9f5e --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed314/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be909467b885c6200f6e180bce2a00225577897dc7bb1cb49695d609fbda9351 +size 4600 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/config.json b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..60697306b416d2c5a047fb1dac11a0550500de94 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./model/roberta-base", + "architectures": [ + "RobertaForSequenceClassification" + ], + "attention_probs_dropout_prob": 0.1, + "bos_token_id": 0, + "classifier_dropout": null, + "eos_token_id": 2, + "hidden_act": "gelu", + "hidden_dropout_prob": 0.1, + "hidden_size": 768, + "initializer_range": 0.02, + "intermediate_size": 3072, + "layer_norm_eps": 1e-05, + "max_position_embeddings": 514, + "model_type": "roberta", + "num_attention_heads": 12, + "num_hidden_layers": 12, + "pad_token_id": 1, + "position_embedding_type": "absolute", + "problem_type": "single_label_classification", + "torch_dtype": "float32", + "transformers_version": "4.34.1", + "type_vocab_size": 1, + "use_cache": true, + "vocab_size": 50265 +} diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/pytorch_model.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..81e727a88f4828bb7c87158a50f19306a6c51b5b --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b6c65988b79ce8dbfaefd492ba8c8a2bf9797a6b11d901fa14ee0ddfc3b1ae9 +size 496301870 diff --git a/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/training_args.bin b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0ffe7cf0a0e62317fa1688ca39bd09face756127 --- /dev/null +++ b/joint-multitask-learning/roberta-base-two-task-mtl_offensiveYN_20240426163520-seed42/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72d7b32635f84b6bccb58f57c3b4831f9cb4bda4697d10abd02a4f8811630a19 +size 4600