Upload 7 files
Browse files- config.json +273 -225
- model.safetensors +1 -1
- tokenizer.json +2 -2
- tokenizer_config.json +1 -1
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"DebertaV2ForSequenceClassification"
|
5 |
],
|
@@ -28,8 +28,8 @@
|
|
28 |
2,
|
29 |
3,
|
30 |
2,
|
31 |
-
6,
|
32 |
2,
|
|
|
33 |
2,
|
34 |
2,
|
35 |
2,
|
@@ -68,15 +68,10 @@
|
|
68 |
3,
|
69 |
3,
|
70 |
3,
|
71 |
-
3,
|
72 |
2,
|
73 |
2,
|
74 |
2,
|
75 |
2,
|
76 |
-
47,
|
77 |
-
23,
|
78 |
-
9,
|
79 |
-
1,
|
80 |
1,
|
81 |
1,
|
82 |
1,
|
@@ -218,12 +213,12 @@
|
|
218 |
4,
|
219 |
2,
|
220 |
8,
|
221 |
-
|
222 |
2,
|
|
|
223 |
2,
|
224 |
2,
|
225 |
-
|
226 |
-
20,
|
227 |
3,
|
228 |
3,
|
229 |
3,
|
@@ -231,33 +226,33 @@
|
|
231 |
3,
|
232 |
174,
|
233 |
2,
|
234 |
-
2,
|
235 |
41,
|
|
|
|
|
236 |
2,
|
237 |
2,
|
238 |
-
51,
|
239 |
2,
|
240 |
-
3,
|
241 |
2,
|
242 |
2,
|
243 |
2,
|
244 |
-
3,
|
245 |
16,
|
|
|
246 |
2,
|
247 |
-
18,
|
248 |
8,
|
249 |
2,
|
|
|
|
|
|
|
250 |
17,
|
251 |
3,
|
252 |
-
|
253 |
-
4,
|
254 |
7,
|
|
|
255 |
12,
|
256 |
7,
|
257 |
-
|
258 |
-
3,
|
259 |
42,
|
260 |
-
|
261 |
100,
|
262 |
13,
|
263 |
100,
|
@@ -287,42 +282,37 @@
|
|
287 |
28,
|
288 |
3,
|
289 |
6,
|
290 |
-
3,
|
291 |
-
6,
|
292 |
-
5,
|
293 |
7,
|
|
|
294 |
4,
|
|
|
|
|
295 |
2,
|
296 |
2,
|
297 |
2,
|
298 |
6,
|
|
|
|
|
299 |
2,
|
300 |
2,
|
301 |
-
7,
|
302 |
-
20,
|
303 |
2,
|
304 |
9,
|
305 |
-
2,
|
306 |
-
3,
|
307 |
13,
|
308 |
2,
|
309 |
3,
|
310 |
-
2,
|
311 |
4,
|
|
|
|
|
312 |
4,
|
313 |
2,
|
314 |
2,
|
315 |
2,
|
316 |
2,
|
|
|
317 |
4,
|
318 |
1,
|
319 |
2,
|
320 |
1,
|
321 |
-
13,
|
322 |
-
3,
|
323 |
-
5,
|
324 |
-
11,
|
325 |
-
37,
|
326 |
2,
|
327 |
49,
|
328 |
40,
|
@@ -371,8 +361,6 @@
|
|
371 |
1,
|
372 |
2,
|
373 |
1,
|
374 |
-
3,
|
375 |
-
2,
|
376 |
2,
|
377 |
2,
|
378 |
2,
|
@@ -423,12 +411,6 @@
|
|
423 |
1,
|
424 |
2,
|
425 |
2,
|
426 |
-
50,
|
427 |
-
50,
|
428 |
-
50,
|
429 |
-
50,
|
430 |
-
2,
|
431 |
-
1,
|
432 |
1,
|
433 |
1,
|
434 |
1,
|
@@ -444,22 +426,12 @@
|
|
444 |
2,
|
445 |
2,
|
446 |
77,
|
447 |
-
2,
|
448 |
-
1,
|
449 |
3,
|
450 |
2,
|
451 |
2,
|
452 |
1,
|
453 |
1,
|
454 |
2,
|
455 |
-
2,
|
456 |
-
2,
|
457 |
-
2,
|
458 |
-
2,
|
459 |
-
2,
|
460 |
-
2,
|
461 |
-
2,
|
462 |
-
2,
|
463 |
3,
|
464 |
18,
|
465 |
13,
|
@@ -468,13 +440,9 @@
|
|
468 |
2,
|
469 |
2,
|
470 |
2,
|
471 |
-
2,
|
472 |
4,
|
473 |
2,
|
474 |
24,
|
475 |
-
23,
|
476 |
-
67,
|
477 |
-
279,
|
478 |
3,
|
479 |
2,
|
480 |
2,
|
@@ -494,6 +462,7 @@
|
|
494 |
1,
|
495 |
17,
|
496 |
3,
|
|
|
497 |
2,
|
498 |
3,
|
499 |
2,
|
@@ -504,11 +473,66 @@
|
|
504 |
1,
|
505 |
3,
|
506 |
2,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
507 |
2,
|
508 |
3,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
509 |
3,
|
510 |
3,
|
511 |
1,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
512 |
1
|
513 |
],
|
514 |
"hidden_act": "gelu",
|
@@ -527,7 +551,7 @@
|
|
527 |
"neutral": 1
|
528 |
},
|
529 |
"layer_norm_eps": 1e-07,
|
530 |
-
"max_position_embeddings":
|
531 |
"max_relative_positions": -1,
|
532 |
"model_type": "deberta-v2",
|
533 |
"norm_rel_ebd": "layer_norm",
|
@@ -568,18 +592,18 @@
|
|
568 |
"scitail/snli_format",
|
569 |
"hans",
|
570 |
"WANLI",
|
571 |
-
"recast/
|
|
|
572 |
"recast/recast_kg_relations",
|
|
|
|
|
573 |
"recast/recast_ner",
|
574 |
-
"recast/recast_factuality",
|
575 |
-
"recast/recast_puns",
|
576 |
"recast/recast_megaveridicality",
|
577 |
"recast/recast_sentiment",
|
578 |
-
"recast/recast_verbcorner",
|
579 |
-
"probability_words_nli/usnli",
|
580 |
"probability_words_nli/reasoning_1hop",
|
|
|
581 |
"probability_words_nli/reasoning_2hop",
|
582 |
-
"nan-nli
|
583 |
"nli_fever",
|
584 |
"breaking_nli",
|
585 |
"conj_nli",
|
@@ -602,125 +626,120 @@
|
|
602 |
"robust_nli_is_sd",
|
603 |
"robust_nli_li_ts",
|
604 |
"add_one_rte",
|
|
|
|
|
605 |
"imppres/implicature_numerals_10_100/log",
|
|
|
606 |
"imppres/implicature_connectives/log",
|
607 |
-
"imppres/implicature_modals/log",
|
608 |
"imppres/implicature_gradable_verb/log",
|
609 |
-
"imppres/implicature_gradable_adjective/log",
|
610 |
"imppres/implicature_numerals_2_3/log",
|
611 |
-
"imppres/implicature_quantifiers/log",
|
612 |
-
"glue_diagnostics/diagnostics",
|
613 |
"hlgd",
|
614 |
"paws/labeled_final",
|
615 |
"paws/labeled_swap",
|
616 |
"medical_questions_pairs",
|
617 |
-
"conll2003/pos_tags",
|
618 |
-
"conll2003/chunk_tags",
|
619 |
-
"conll2003/ner_tags",
|
620 |
-
"hh-rlhf",
|
621 |
"model-written-evals",
|
622 |
"truthful_qa/multiple_choice",
|
623 |
"fig-qa",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
624 |
"bigbench/physical_intuition",
|
625 |
-
"bigbench/authorship_verification",
|
626 |
-
"bigbench/implicit_relations",
|
627 |
-
"bigbench/dyck_languages",
|
628 |
-
"bigbench/novel_concepts",
|
629 |
-
"bigbench/moral_permissibility",
|
630 |
-
"bigbench/metaphor_understanding",
|
631 |
-
"bigbench/temporal_sequences",
|
632 |
-
"bigbench/sports_understanding",
|
633 |
"bigbench/analytic_entailment",
|
634 |
-
"bigbench/
|
635 |
-
"bigbench/
|
636 |
-
"bigbench/
|
637 |
-
"bigbench/
|
638 |
-
"bigbench/
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
639 |
"bigbench/hhh_alignment",
|
640 |
-
"bigbench/
|
641 |
-
"bigbench/bbq_lite_json",
|
642 |
-
"bigbench/cause_and_effect",
|
643 |
-
"bigbench/logic_grid_puzzle",
|
644 |
-
"bigbench/empirical_judgments",
|
645 |
-
"bigbench/human_organs_senses",
|
646 |
-
"bigbench/misconceptions",
|
647 |
-
"bigbench/strange_stories",
|
648 |
-
"bigbench/logical_args",
|
649 |
-
"bigbench/known_unknowns",
|
650 |
"bigbench/cs_algorithms",
|
651 |
-
"bigbench/
|
652 |
-
"bigbench/cifar10_classification",
|
653 |
-
"bigbench/penguins_in_a_table",
|
654 |
-
"bigbench/odd_one_out",
|
655 |
-
"bigbench/intent_recognition",
|
656 |
-
"bigbench/physics",
|
657 |
"bigbench/conceptual_combinations",
|
658 |
-
"bigbench/
|
659 |
-
"bigbench/
|
660 |
-
"bigbench/winowhy",
|
661 |
"bigbench/arithmetic",
|
662 |
-
"bigbench/
|
663 |
-
"bigbench/
|
|
|
664 |
"bigbench/social_iqa",
|
665 |
-
"bigbench/
|
666 |
-
"bigbench/
|
667 |
-
"bigbench/
|
668 |
-
"bigbench/disambiguation_qa",
|
669 |
"bigbench/similarities_abstraction",
|
670 |
-
"bigbench/
|
671 |
-
"bigbench/
|
672 |
-
"bigbench/
|
673 |
-
"bigbench/
|
674 |
-
"bigbench/
|
675 |
-
"bigbench/
|
676 |
-
"bigbench/
|
677 |
-
"bigbench/play_dialog_same_or_different",
|
678 |
-
"bigbench/hindu_knowledge",
|
679 |
-
"bigbench/international_phonetic_alphabet_nli",
|
680 |
-
"bigbench/understanding_fables",
|
681 |
-
"bigbench/geometric_shapes",
|
682 |
-
"bigbench/code_line_description",
|
683 |
"bigbench/riddle_sense",
|
684 |
-
"bigbench/
|
685 |
-
"bigbench/
|
|
|
|
|
686 |
"bigbench/anachronisms",
|
|
|
|
|
|
|
687 |
"bigbench/navigate",
|
688 |
-
"bigbench/
|
689 |
-
"bigbench/
|
690 |
-
"bigbench/
|
691 |
-
"bigbench/
|
|
|
|
|
|
|
|
|
|
|
692 |
"bigbench/date_understanding",
|
693 |
"bigbench/figure_of_speech_detection",
|
694 |
-
"bigbench/
|
695 |
-
"bigbench/
|
696 |
-
"bigbench/nonsense_words_grammar",
|
697 |
-
"bigbench/salient_translation_error_detection",
|
698 |
-
"bigbench/epistemic_reasoning",
|
699 |
-
"bigbench/movie_recommendation",
|
700 |
-
"bigbench/strategyqa",
|
701 |
-
"bigbench/tracking_shuffled_objects",
|
702 |
-
"bigbench/unit_interpretation",
|
703 |
-
"bigbench/reasoning_about_colored_objects",
|
704 |
-
"bigbench/discourse_marker_prediction",
|
705 |
-
"bigbench/logical_fallacy_detection",
|
706 |
-
"bigbench/general_knowledge",
|
707 |
"bigbench/abstract_narrative_understanding",
|
708 |
-
"bigbench/
|
709 |
-
"bigbench/
|
710 |
-
"bigbench/logical_sequence",
|
711 |
-
"bigbench/mnist_ascii",
|
712 |
-
"bigbench/fantasy_reasoning",
|
713 |
-
"bigbench/mathematical_induction",
|
714 |
-
"bigbench/timedial",
|
715 |
"bigbench/identify_math_theorems",
|
716 |
-
"bigbench/
|
717 |
-
"bigbench/phrase_relatedness",
|
718 |
-
"bigbench/ruin_names",
|
719 |
-
"bigbench/gre_reading_comprehension",
|
720 |
-
"bigbench/metaphor_boolean",
|
721 |
-
"bigbench/sentence_ambiguity",
|
722 |
"bigbench/vitaminc_fact_verification",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
723 |
"bigbench/evaluating_information_essentiality",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
724 |
"cos_e/v1.0",
|
725 |
"cosmos_qa",
|
726 |
"dream",
|
@@ -741,8 +760,8 @@
|
|
741 |
"art",
|
742 |
"winogrande/winogrande_xl",
|
743 |
"codah/codah",
|
744 |
-
"ai2_arc/ARC-Challenge/challenge",
|
745 |
"ai2_arc/ARC-Easy/challenge",
|
|
|
746 |
"definite_pronoun_resolution",
|
747 |
"swag/regular",
|
748 |
"math_qa",
|
@@ -754,51 +773,51 @@
|
|
754 |
"toxic_conversations",
|
755 |
"TuringBench",
|
756 |
"trec",
|
757 |
-
"vitaminc
|
758 |
"hope_edi/english",
|
759 |
"rumoureval_2019/RumourEval2019",
|
760 |
"ethos/binary",
|
761 |
"ethos/multilabel",
|
762 |
-
"tweet_eval/
|
763 |
-
"tweet_eval/irony",
|
764 |
"tweet_eval/offensive",
|
765 |
-
"tweet_eval/hate",
|
766 |
"tweet_eval/emotion",
|
767 |
-
"tweet_eval/
|
|
|
|
|
768 |
"tweet_eval/stance_abortion",
|
769 |
"tweet_eval/stance_atheism",
|
770 |
"tweet_eval/stance_climate",
|
771 |
"tweet_eval/stance_feminist",
|
772 |
"tweet_eval/stance_hillary",
|
773 |
"discovery/discovery",
|
774 |
-
"pragmeval/squinky-informativeness",
|
775 |
-
"pragmeval/emobank-arousal",
|
776 |
-
"pragmeval/switchboard",
|
777 |
-
"pragmeval/squinky-implicature",
|
778 |
-
"pragmeval/emobank-valence",
|
779 |
-
"pragmeval/mrda",
|
780 |
"pragmeval/squinky-formality",
|
|
|
781 |
"pragmeval/verifiability",
|
|
|
|
|
|
|
|
|
782 |
"pragmeval/emobank-dominance",
|
783 |
-
"pragmeval/
|
784 |
-
"pragmeval/persuasiveness-strength",
|
785 |
-
"pragmeval/persuasiveness-claimtype",
|
786 |
-
"pragmeval/pdtb",
|
787 |
"pragmeval/sarcasm",
|
788 |
-
"pragmeval/
|
|
|
|
|
789 |
"pragmeval/persuasiveness-premisetype",
|
790 |
"pragmeval/persuasiveness-eloquence",
|
|
|
|
|
|
|
791 |
"pragmeval/gum",
|
792 |
"pragmeval/emergent",
|
793 |
-
"pragmeval/persuasiveness-relevance",
|
794 |
-
"silicone/dyda_da",
|
795 |
-
"silicone/dyda_e",
|
796 |
-
"silicone/maptask",
|
797 |
-
"silicone/meld_e",
|
798 |
-
"silicone/meld_s",
|
799 |
"silicone/sem",
|
800 |
-
"silicone/
|
801 |
"silicone/iemocap",
|
|
|
|
|
|
|
|
|
|
|
802 |
"lex_glue/eurlex",
|
803 |
"lex_glue/scotus",
|
804 |
"lex_glue/ledgar",
|
@@ -828,29 +847,29 @@
|
|
828 |
"go_emotions/simplified",
|
829 |
"scicite",
|
830 |
"liar",
|
831 |
-
"lexical_relation_classification/ROOT09",
|
832 |
-
"lexical_relation_classification/BLESS",
|
833 |
-
"lexical_relation_classification/CogALexV",
|
834 |
"lexical_relation_classification/EVALution",
|
|
|
835 |
"lexical_relation_classification/K&H+N",
|
836 |
-
"
|
837 |
-
"
|
|
|
838 |
"linguisticprobing/past_present",
|
|
|
839 |
"linguisticprobing/sentence_length",
|
840 |
-
"linguisticprobing/subj_number",
|
841 |
-
"linguisticprobing/odd_man_out",
|
842 |
-
"linguisticprobing/tree_depth",
|
843 |
"linguisticprobing/top_constituents",
|
|
|
|
|
|
|
844 |
"linguisticprobing/bigram_shift",
|
845 |
"crowdflower/political-media-message",
|
|
|
846 |
"crowdflower/political-media-audience",
|
847 |
"crowdflower/economic-news",
|
848 |
-
"crowdflower/
|
849 |
-
"crowdflower/political-media-bias",
|
850 |
"crowdflower/airline-sentiment",
|
851 |
"crowdflower/tweet_global_warming",
|
852 |
-
"crowdflower/corporate-messaging",
|
853 |
"crowdflower/sentiment_nuclear_power",
|
|
|
854 |
"ethics/commonsense",
|
855 |
"ethics/deontology",
|
856 |
"ethics/justice",
|
@@ -859,11 +878,6 @@
|
|
859 |
"google_wellformed_query",
|
860 |
"tweets_hate_speech_detection",
|
861 |
"has_part",
|
862 |
-
"wnut_17/wnut_17",
|
863 |
-
"ncbi_disease/ncbi_disease",
|
864 |
-
"acronym_identification",
|
865 |
-
"jnlpba/jnlpba",
|
866 |
-
"ontonotes_english/SpeedOfMagic--ontonotes_english",
|
867 |
"blog_authorship_corpus/gender",
|
868 |
"blog_authorship_corpus/age",
|
869 |
"blog_authorship_corpus/job",
|
@@ -877,13 +891,13 @@
|
|
877 |
"phrase_similarity",
|
878 |
"scientific-exaggeration-detection",
|
879 |
"quarel",
|
880 |
-
"fever-evidence-related
|
881 |
"numer_sense",
|
882 |
"dynasent/dynabench.dynasent.r1.all/r1",
|
883 |
"dynasent/dynabench.dynasent.r2.all/r2",
|
884 |
"Sarcasm_News_Headline",
|
885 |
"sem_eval_2010_task_8",
|
886 |
-
"auditor_review
|
887 |
"medmcqa",
|
888 |
"Dynasent_Disagreement",
|
889 |
"Politeness_Disagreement",
|
@@ -912,12 +926,10 @@
|
|
912 |
"synthetic-instruct-gptj-pairwise",
|
913 |
"scruples",
|
914 |
"wouldyourather",
|
915 |
-
"attempto-nli",
|
916 |
-
"defeasible-nli/snli",
|
917 |
"defeasible-nli/atomic",
|
|
|
918 |
"help-nli",
|
919 |
"nli-veridicality-transitivity",
|
920 |
-
"natural-language-satisfiability",
|
921 |
"lonli",
|
922 |
"dadc-limit-nli",
|
923 |
"FLUTE",
|
@@ -959,15 +971,10 @@
|
|
959 |
"chaos-mnli-ambiguity",
|
960 |
"headline_cause/en_simple",
|
961 |
"logiqa-2.0-nli",
|
962 |
-
"
|
963 |
-
"
|
964 |
-
"
|
965 |
-
"PARARULE-Plus",
|
966 |
"mindgames",
|
967 |
-
"universal_dependencies/en_lines/deprel",
|
968 |
-
"universal_dependencies/en_partut/deprel",
|
969 |
-
"universal_dependencies/en_ewt/deprel",
|
970 |
-
"universal_dependencies/en_gum/deprel",
|
971 |
"ambient",
|
972 |
"path-naturalness-prediction",
|
973 |
"civil_comments/toxicity",
|
@@ -979,28 +986,17 @@
|
|
979 |
"civil_comments/sexual_explicit",
|
980 |
"cloth",
|
981 |
"dgen",
|
982 |
-
"oasst1_pairwise_rlhf_reward",
|
983 |
"I2D2",
|
984 |
"args_me",
|
985 |
"Touche23-ValueEval",
|
986 |
"starcon",
|
987 |
"banking77",
|
988 |
-
"ruletaker",
|
989 |
-
"lsat_qa/all",
|
990 |
"ConTRoL-nli",
|
991 |
"tracie",
|
992 |
"sherliic",
|
993 |
"sen-making/1",
|
994 |
"sen-making/2",
|
995 |
"winowhy",
|
996 |
-
"mbib-base/cognitive-bias",
|
997 |
-
"mbib-base/fake-news",
|
998 |
-
"mbib-base/gender-bias",
|
999 |
-
"mbib-base/hate-speech",
|
1000 |
-
"mbib-base/linguistic-bias",
|
1001 |
-
"mbib-base/political-bias",
|
1002 |
-
"mbib-base/racial-bias",
|
1003 |
-
"mbib-base/text-level-bias",
|
1004 |
"robustLR",
|
1005 |
"v1/gen_train234_test2to10",
|
1006 |
"logical-fallacy",
|
@@ -1009,13 +1005,9 @@
|
|
1009 |
"subjectivity",
|
1010 |
"MOH",
|
1011 |
"VUAC",
|
1012 |
-
"TroFi",
|
1013 |
"sharc_modified/mod",
|
1014 |
"conceptrules_v2",
|
1015 |
"disrpt/eng.dep.scidtb.rels",
|
1016 |
-
"conll2000",
|
1017 |
-
"few-nerd/supervised",
|
1018 |
-
"finer-139",
|
1019 |
"zero-shot-label-nli",
|
1020 |
"com2sense",
|
1021 |
"scone",
|
@@ -1034,7 +1026,8 @@
|
|
1034 |
"esci",
|
1035 |
"chatbot_arena_conversations",
|
1036 |
"dnd_style_intents",
|
1037 |
-
"FLD.v2",
|
|
|
1038 |
"SDOH-NLI",
|
1039 |
"scifact_entailment",
|
1040 |
"feasibilityQA",
|
@@ -1045,15 +1038,70 @@
|
|
1045 |
"ReSQ",
|
1046 |
"semantic_fragments_nli",
|
1047 |
"dataset_train_nli",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1048 |
"babi_nli",
|
1049 |
"gen_debiased_nli",
|
1050 |
"imppres/presupposition",
|
1051 |
"/prag",
|
1052 |
-
"blimp-2"
|
1053 |
-
"mmlu-4"
|
1054 |
],
|
1055 |
"torch_dtype": "float32",
|
1056 |
-
"transformers_version": "4.
|
1057 |
"type_vocab_size": 0,
|
1058 |
"vocab_size": 128100
|
1059 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "tasksource/deberta-base-long-nli",
|
3 |
"architectures": [
|
4 |
"DebertaV2ForSequenceClassification"
|
5 |
],
|
|
|
28 |
2,
|
29 |
3,
|
30 |
2,
|
|
|
31 |
2,
|
32 |
+
6,
|
33 |
2,
|
34 |
2,
|
35 |
2,
|
|
|
68 |
3,
|
69 |
3,
|
70 |
3,
|
|
|
71 |
2,
|
72 |
2,
|
73 |
2,
|
74 |
2,
|
|
|
|
|
|
|
|
|
75 |
1,
|
76 |
1,
|
77 |
1,
|
|
|
213 |
4,
|
214 |
2,
|
215 |
8,
|
216 |
+
20,
|
217 |
2,
|
218 |
+
4,
|
219 |
2,
|
220 |
2,
|
221 |
+
3,
|
|
|
222 |
3,
|
223 |
3,
|
224 |
3,
|
|
|
226 |
3,
|
227 |
174,
|
228 |
2,
|
|
|
229 |
41,
|
230 |
+
3,
|
231 |
+
51,
|
232 |
2,
|
233 |
2,
|
|
|
234 |
2,
|
|
|
235 |
2,
|
236 |
2,
|
237 |
2,
|
|
|
238 |
16,
|
239 |
+
3,
|
240 |
2,
|
|
|
241 |
8,
|
242 |
2,
|
243 |
+
2,
|
244 |
+
18,
|
245 |
+
2,
|
246 |
17,
|
247 |
3,
|
248 |
+
3,
|
|
|
249 |
7,
|
250 |
+
11,
|
251 |
12,
|
252 |
7,
|
253 |
+
4,
|
|
|
254 |
42,
|
255 |
+
3,
|
256 |
100,
|
257 |
13,
|
258 |
100,
|
|
|
282 |
28,
|
283 |
3,
|
284 |
6,
|
|
|
|
|
|
|
285 |
7,
|
286 |
+
3,
|
287 |
4,
|
288 |
+
5,
|
289 |
+
6,
|
290 |
2,
|
291 |
2,
|
292 |
2,
|
293 |
6,
|
294 |
+
20,
|
295 |
+
7,
|
296 |
2,
|
297 |
2,
|
|
|
|
|
298 |
2,
|
299 |
9,
|
|
|
|
|
300 |
13,
|
301 |
2,
|
302 |
3,
|
|
|
303 |
4,
|
304 |
+
3,
|
305 |
+
2,
|
306 |
4,
|
307 |
2,
|
308 |
2,
|
309 |
2,
|
310 |
2,
|
311 |
+
2,
|
312 |
4,
|
313 |
1,
|
314 |
2,
|
315 |
1,
|
|
|
|
|
|
|
|
|
|
|
316 |
2,
|
317 |
49,
|
318 |
40,
|
|
|
361 |
1,
|
362 |
2,
|
363 |
1,
|
|
|
|
|
364 |
2,
|
365 |
2,
|
366 |
2,
|
|
|
411 |
1,
|
412 |
2,
|
413 |
2,
|
|
|
|
|
|
|
|
|
|
|
|
|
414 |
1,
|
415 |
1,
|
416 |
1,
|
|
|
426 |
2,
|
427 |
2,
|
428 |
77,
|
|
|
|
|
429 |
3,
|
430 |
2,
|
431 |
2,
|
432 |
1,
|
433 |
1,
|
434 |
2,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
435 |
3,
|
436 |
18,
|
437 |
13,
|
|
|
440 |
2,
|
441 |
2,
|
442 |
2,
|
|
|
443 |
4,
|
444 |
2,
|
445 |
24,
|
|
|
|
|
|
|
446 |
3,
|
447 |
2,
|
448 |
2,
|
|
|
462 |
1,
|
463 |
17,
|
464 |
3,
|
465 |
+
3,
|
466 |
2,
|
467 |
3,
|
468 |
2,
|
|
|
473 |
1,
|
474 |
3,
|
475 |
2,
|
476 |
+
9,
|
477 |
+
2,
|
478 |
+
1,
|
479 |
+
1,
|
480 |
+
1,
|
481 |
+
1,
|
482 |
+
1,
|
483 |
+
2,
|
484 |
2,
|
485 |
3,
|
486 |
+
2,
|
487 |
+
3,
|
488 |
+
4,
|
489 |
+
3,
|
490 |
+
3,
|
491 |
+
2,
|
492 |
+
1,
|
493 |
+
1,
|
494 |
+
2,
|
495 |
+
1,
|
496 |
+
1,
|
497 |
+
6,
|
498 |
+
2,
|
499 |
3,
|
500 |
3,
|
501 |
1,
|
502 |
+
1,
|
503 |
+
2,
|
504 |
+
2,
|
505 |
+
10,
|
506 |
+
2,
|
507 |
+
2,
|
508 |
+
2,
|
509 |
+
5,
|
510 |
+
5,
|
511 |
+
5,
|
512 |
+
5,
|
513 |
+
5,
|
514 |
+
5,
|
515 |
+
5,
|
516 |
+
5,
|
517 |
+
5,
|
518 |
+
5,
|
519 |
+
4,
|
520 |
+
1,
|
521 |
+
7,
|
522 |
+
6,
|
523 |
+
6,
|
524 |
+
6,
|
525 |
+
6,
|
526 |
+
6,
|
527 |
+
6,
|
528 |
+
3,
|
529 |
+
2,
|
530 |
+
1,
|
531 |
+
2,
|
532 |
+
2,
|
533 |
+
3,
|
534 |
+
3,
|
535 |
+
3,
|
536 |
1
|
537 |
],
|
538 |
"hidden_act": "gelu",
|
|
|
551 |
"neutral": 1
|
552 |
},
|
553 |
"layer_norm_eps": 1e-07,
|
554 |
+
"max_position_embeddings": 1280,
|
555 |
"max_relative_positions": -1,
|
556 |
"model_type": "deberta-v2",
|
557 |
"norm_rel_ebd": "layer_norm",
|
|
|
592 |
"scitail/snli_format",
|
593 |
"hans",
|
594 |
"WANLI",
|
595 |
+
"recast/recast_puns",
|
596 |
+
"recast/recast_factuality",
|
597 |
"recast/recast_kg_relations",
|
598 |
+
"recast/recast_verbnet",
|
599 |
+
"recast/recast_verbcorner",
|
600 |
"recast/recast_ner",
|
|
|
|
|
601 |
"recast/recast_megaveridicality",
|
602 |
"recast/recast_sentiment",
|
|
|
|
|
603 |
"probability_words_nli/reasoning_1hop",
|
604 |
+
"probability_words_nli/usnli",
|
605 |
"probability_words_nli/reasoning_2hop",
|
606 |
+
"nan-nli",
|
607 |
"nli_fever",
|
608 |
"breaking_nli",
|
609 |
"conj_nli",
|
|
|
626 |
"robust_nli_is_sd",
|
627 |
"robust_nli_li_ts",
|
628 |
"add_one_rte",
|
629 |
+
"imppres/implicature_gradable_adjective/log",
|
630 |
+
"imppres/implicature_modals/log",
|
631 |
"imppres/implicature_numerals_10_100/log",
|
632 |
+
"imppres/implicature_quantifiers/log",
|
633 |
"imppres/implicature_connectives/log",
|
|
|
634 |
"imppres/implicature_gradable_verb/log",
|
|
|
635 |
"imppres/implicature_numerals_2_3/log",
|
|
|
|
|
636 |
"hlgd",
|
637 |
"paws/labeled_final",
|
638 |
"paws/labeled_swap",
|
639 |
"medical_questions_pairs",
|
|
|
|
|
|
|
|
|
640 |
"model-written-evals",
|
641 |
"truthful_qa/multiple_choice",
|
642 |
"fig-qa",
|
643 |
+
"bigbench/discourse_marker_prediction",
|
644 |
+
"bigbench/gre_reading_comprehension",
|
645 |
+
"bigbench/simple_ethical_questions",
|
646 |
+
"bigbench/known_unknowns",
|
647 |
+
"bigbench/fact_checker",
|
648 |
+
"bigbench/phrase_relatedness",
|
649 |
+
"bigbench/irony_identification",
|
650 |
+
"bigbench/crash_blossom",
|
651 |
"bigbench/physical_intuition",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
652 |
"bigbench/analytic_entailment",
|
653 |
+
"bigbench/play_dialog_same_or_different",
|
654 |
+
"bigbench/metaphor_boolean",
|
655 |
+
"bigbench/undo_permutation",
|
656 |
+
"bigbench/disambiguation_qa",
|
657 |
+
"bigbench/metaphor_understanding",
|
658 |
+
"bigbench/nonsense_words_grammar",
|
659 |
+
"bigbench/logical_deduction",
|
660 |
+
"bigbench/sentence_ambiguity",
|
661 |
+
"bigbench/key_value_maps",
|
662 |
+
"bigbench/checkmate_in_one",
|
663 |
+
"bigbench/code_line_description",
|
664 |
+
"bigbench/crass_ai",
|
665 |
+
"bigbench/movie_dialog_same_or_different",
|
666 |
+
"bigbench/strategyqa",
|
667 |
"bigbench/hhh_alignment",
|
668 |
+
"bigbench/english_proverbs",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
669 |
"bigbench/cs_algorithms",
|
670 |
+
"bigbench/formal_fallacies_syllogisms_negation",
|
|
|
|
|
|
|
|
|
|
|
671 |
"bigbench/conceptual_combinations",
|
672 |
+
"bigbench/geometric_shapes",
|
673 |
+
"bigbench/tracking_shuffled_objects",
|
|
|
674 |
"bigbench/arithmetic",
|
675 |
+
"bigbench/misconceptions",
|
676 |
+
"bigbench/odd_one_out",
|
677 |
+
"bigbench/international_phonetic_alphabet_nli",
|
678 |
"bigbench/social_iqa",
|
679 |
+
"bigbench/logical_fallacy_detection",
|
680 |
+
"bigbench/hyperbaton",
|
681 |
+
"bigbench/novel_concepts",
|
|
|
682 |
"bigbench/similarities_abstraction",
|
683 |
+
"bigbench/timedial",
|
684 |
+
"bigbench/salient_translation_error_detection",
|
685 |
+
"bigbench/cifar10_classification",
|
686 |
+
"bigbench/winowhy",
|
687 |
+
"bigbench/mathematical_induction",
|
688 |
+
"bigbench/real_or_fake_text",
|
689 |
+
"bigbench/epistemic_reasoning",
|
|
|
|
|
|
|
|
|
|
|
|
|
690 |
"bigbench/riddle_sense",
|
691 |
+
"bigbench/identify_odd_metaphor",
|
692 |
+
"bigbench/human_organs_senses",
|
693 |
+
"bigbench/emojis_emotion_prediction",
|
694 |
+
"bigbench/logical_args",
|
695 |
"bigbench/anachronisms",
|
696 |
+
"bigbench/bbq_lite_json",
|
697 |
+
"bigbench/general_knowledge",
|
698 |
+
"bigbench/logical_sequence",
|
699 |
"bigbench/navigate",
|
700 |
+
"bigbench/analogical_similarity",
|
701 |
+
"bigbench/authorship_verification",
|
702 |
+
"bigbench/elementary_math_qa",
|
703 |
+
"bigbench/goal_step_wikihow",
|
704 |
+
"bigbench/empirical_judgments",
|
705 |
+
"bigbench/question_selection",
|
706 |
+
"bigbench/temporal_sequences",
|
707 |
+
"bigbench/mnist_ascii",
|
708 |
+
"bigbench/color",
|
709 |
"bigbench/date_understanding",
|
710 |
"bigbench/figure_of_speech_detection",
|
711 |
+
"bigbench/contextual_parametric_knowledge_conflicts",
|
712 |
+
"bigbench/ruin_names",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
713 |
"bigbench/abstract_narrative_understanding",
|
714 |
+
"bigbench/physics",
|
715 |
+
"bigbench/understanding_fables",
|
|
|
|
|
|
|
|
|
|
|
716 |
"bigbench/identify_math_theorems",
|
717 |
+
"bigbench/penguins_in_a_table",
|
|
|
|
|
|
|
|
|
|
|
718 |
"bigbench/vitaminc_fact_verification",
|
719 |
+
"bigbench/social_support",
|
720 |
+
"bigbench/strange_stories",
|
721 |
+
"bigbench/dark_humor_detection",
|
722 |
+
"bigbench/fantasy_reasoning",
|
723 |
+
"bigbench/logic_grid_puzzle",
|
724 |
+
"bigbench/emoji_movie",
|
725 |
+
"bigbench/moral_permissibility",
|
726 |
+
"bigbench/symbol_interpretation",
|
727 |
+
"bigbench/presuppositions_as_nli",
|
728 |
+
"bigbench/dyck_languages",
|
729 |
+
"bigbench/implicatures",
|
730 |
+
"bigbench/suicide_risk",
|
731 |
+
"bigbench/unit_interpretation",
|
732 |
+
"bigbench/intent_recognition",
|
733 |
+
"bigbench/movie_recommendation",
|
734 |
+
"bigbench/snarks",
|
735 |
"bigbench/evaluating_information_essentiality",
|
736 |
+
"bigbench/sports_understanding",
|
737 |
+
"bigbench/entailed_polarity",
|
738 |
+
"bigbench/causal_judgment",
|
739 |
+
"bigbench/cause_and_effect",
|
740 |
+
"bigbench/implicit_relations",
|
741 |
+
"bigbench/reasoning_about_colored_objects",
|
742 |
+
"bigbench/hindu_knowledge",
|
743 |
"cos_e/v1.0",
|
744 |
"cosmos_qa",
|
745 |
"dream",
|
|
|
760 |
"art",
|
761 |
"winogrande/winogrande_xl",
|
762 |
"codah/codah",
|
|
|
763 |
"ai2_arc/ARC-Easy/challenge",
|
764 |
+
"ai2_arc/ARC-Challenge/challenge",
|
765 |
"definite_pronoun_resolution",
|
766 |
"swag/regular",
|
767 |
"math_qa",
|
|
|
773 |
"toxic_conversations",
|
774 |
"TuringBench",
|
775 |
"trec",
|
776 |
+
"vitaminc",
|
777 |
"hope_edi/english",
|
778 |
"rumoureval_2019/RumourEval2019",
|
779 |
"ethos/binary",
|
780 |
"ethos/multilabel",
|
781 |
+
"tweet_eval/emoji",
|
|
|
782 |
"tweet_eval/offensive",
|
|
|
783 |
"tweet_eval/emotion",
|
784 |
+
"tweet_eval/hate",
|
785 |
+
"tweet_eval/irony",
|
786 |
+
"tweet_eval/sentiment",
|
787 |
"tweet_eval/stance_abortion",
|
788 |
"tweet_eval/stance_atheism",
|
789 |
"tweet_eval/stance_climate",
|
790 |
"tweet_eval/stance_feminist",
|
791 |
"tweet_eval/stance_hillary",
|
792 |
"discovery/discovery",
|
|
|
|
|
|
|
|
|
|
|
|
|
793 |
"pragmeval/squinky-formality",
|
794 |
+
"pragmeval/switchboard",
|
795 |
"pragmeval/verifiability",
|
796 |
+
"pragmeval/mrda",
|
797 |
+
"pragmeval/squinky-informativeness",
|
798 |
+
"pragmeval/squinky-implicature",
|
799 |
+
"pragmeval/emobank-arousal",
|
800 |
"pragmeval/emobank-dominance",
|
801 |
+
"pragmeval/emobank-valence",
|
|
|
|
|
|
|
802 |
"pragmeval/sarcasm",
|
803 |
+
"pragmeval/pdtb",
|
804 |
+
"pragmeval/persuasiveness-claimtype",
|
805 |
+
"pragmeval/persuasiveness-relevance",
|
806 |
"pragmeval/persuasiveness-premisetype",
|
807 |
"pragmeval/persuasiveness-eloquence",
|
808 |
+
"pragmeval/persuasiveness-strength",
|
809 |
+
"pragmeval/stac",
|
810 |
+
"pragmeval/persuasiveness-specificity",
|
811 |
"pragmeval/gum",
|
812 |
"pragmeval/emergent",
|
|
|
|
|
|
|
|
|
|
|
|
|
813 |
"silicone/sem",
|
814 |
+
"silicone/meld_e",
|
815 |
"silicone/iemocap",
|
816 |
+
"silicone/maptask",
|
817 |
+
"silicone/dyda_e",
|
818 |
+
"silicone/dyda_da",
|
819 |
+
"silicone/oasis",
|
820 |
+
"silicone/meld_s",
|
821 |
"lex_glue/eurlex",
|
822 |
"lex_glue/scotus",
|
823 |
"lex_glue/ledgar",
|
|
|
847 |
"go_emotions/simplified",
|
848 |
"scicite",
|
849 |
"liar",
|
|
|
|
|
|
|
850 |
"lexical_relation_classification/EVALution",
|
851 |
+
"lexical_relation_classification/ROOT09",
|
852 |
"lexical_relation_classification/K&H+N",
|
853 |
+
"lexical_relation_classification/CogALexV",
|
854 |
+
"lexical_relation_classification/BLESS",
|
855 |
+
"linguisticprobing/subj_number",
|
856 |
"linguisticprobing/past_present",
|
857 |
+
"linguisticprobing/obj_number",
|
858 |
"linguisticprobing/sentence_length",
|
|
|
|
|
|
|
859 |
"linguisticprobing/top_constituents",
|
860 |
+
"linguisticprobing/tree_depth",
|
861 |
+
"linguisticprobing/coordination_inversion",
|
862 |
+
"linguisticprobing/odd_man_out",
|
863 |
"linguisticprobing/bigram_shift",
|
864 |
"crowdflower/political-media-message",
|
865 |
+
"crowdflower/text_emotion",
|
866 |
"crowdflower/political-media-audience",
|
867 |
"crowdflower/economic-news",
|
868 |
+
"crowdflower/corporate-messaging",
|
|
|
869 |
"crowdflower/airline-sentiment",
|
870 |
"crowdflower/tweet_global_warming",
|
|
|
871 |
"crowdflower/sentiment_nuclear_power",
|
872 |
+
"crowdflower/political-media-bias",
|
873 |
"ethics/commonsense",
|
874 |
"ethics/deontology",
|
875 |
"ethics/justice",
|
|
|
878 |
"google_wellformed_query",
|
879 |
"tweets_hate_speech_detection",
|
880 |
"has_part",
|
|
|
|
|
|
|
|
|
|
|
881 |
"blog_authorship_corpus/gender",
|
882 |
"blog_authorship_corpus/age",
|
883 |
"blog_authorship_corpus/job",
|
|
|
891 |
"phrase_similarity",
|
892 |
"scientific-exaggeration-detection",
|
893 |
"quarel",
|
894 |
+
"fever-evidence-related",
|
895 |
"numer_sense",
|
896 |
"dynasent/dynabench.dynasent.r1.all/r1",
|
897 |
"dynasent/dynabench.dynasent.r2.all/r2",
|
898 |
"Sarcasm_News_Headline",
|
899 |
"sem_eval_2010_task_8",
|
900 |
+
"auditor_review",
|
901 |
"medmcqa",
|
902 |
"Dynasent_Disagreement",
|
903 |
"Politeness_Disagreement",
|
|
|
926 |
"synthetic-instruct-gptj-pairwise",
|
927 |
"scruples",
|
928 |
"wouldyourather",
|
|
|
|
|
929 |
"defeasible-nli/atomic",
|
930 |
+
"defeasible-nli/snli",
|
931 |
"help-nli",
|
932 |
"nli-veridicality-transitivity",
|
|
|
933 |
"lonli",
|
934 |
"dadc-limit-nli",
|
935 |
"FLUTE",
|
|
|
971 |
"chaos-mnli-ambiguity",
|
972 |
"headline_cause/en_simple",
|
973 |
"logiqa-2.0-nli",
|
974 |
+
"oasst2_dense_flat/quality",
|
975 |
+
"oasst2_dense_flat/toxicity",
|
976 |
+
"oasst2_dense_flat/helpfulness",
|
|
|
977 |
"mindgames",
|
|
|
|
|
|
|
|
|
978 |
"ambient",
|
979 |
"path-naturalness-prediction",
|
980 |
"civil_comments/toxicity",
|
|
|
986 |
"civil_comments/sexual_explicit",
|
987 |
"cloth",
|
988 |
"dgen",
|
|
|
989 |
"I2D2",
|
990 |
"args_me",
|
991 |
"Touche23-ValueEval",
|
992 |
"starcon",
|
993 |
"banking77",
|
|
|
|
|
994 |
"ConTRoL-nli",
|
995 |
"tracie",
|
996 |
"sherliic",
|
997 |
"sen-making/1",
|
998 |
"sen-making/2",
|
999 |
"winowhy",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1000 |
"robustLR",
|
1001 |
"v1/gen_train234_test2to10",
|
1002 |
"logical-fallacy",
|
|
|
1005 |
"subjectivity",
|
1006 |
"MOH",
|
1007 |
"VUAC",
|
|
|
1008 |
"sharc_modified/mod",
|
1009 |
"conceptrules_v2",
|
1010 |
"disrpt/eng.dep.scidtb.rels",
|
|
|
|
|
|
|
1011 |
"zero-shot-label-nli",
|
1012 |
"com2sense",
|
1013 |
"scone",
|
|
|
1026 |
"esci",
|
1027 |
"chatbot_arena_conversations",
|
1028 |
"dnd_style_intents",
|
1029 |
+
"FLD.v2/default",
|
1030 |
+
"FLD.v2/star",
|
1031 |
"SDOH-NLI",
|
1032 |
"scifact_entailment",
|
1033 |
"feasibilityQA",
|
|
|
1038 |
"ReSQ",
|
1039 |
"semantic_fragments_nli",
|
1040 |
"dataset_train_nli",
|
1041 |
+
"stepgame",
|
1042 |
+
"nlgraph",
|
1043 |
+
"oasst2_pairwise_rlhf_reward",
|
1044 |
+
"hh-rlhf/helpful-online",
|
1045 |
+
"hh-rlhf/helpful-rejection-sampled",
|
1046 |
+
"hh-rlhf/helpful-base",
|
1047 |
+
"hh-rlhf/harmless-base",
|
1048 |
+
"ruletaker",
|
1049 |
+
"PARARULE-Plus",
|
1050 |
+
"proofwriter",
|
1051 |
+
"logical-entailment",
|
1052 |
+
"nope",
|
1053 |
+
"LogicNLI",
|
1054 |
+
"contract-nli/contractnli_a/seg",
|
1055 |
+
"contract-nli/contractnli_b/full",
|
1056 |
+
"nli4ct_semeval2024",
|
1057 |
+
"lsat-ar",
|
1058 |
+
"lsat-rc",
|
1059 |
+
"biosift-nli",
|
1060 |
+
"brainteasers/SP",
|
1061 |
+
"brainteasers/WP",
|
1062 |
+
"persuasion",
|
1063 |
+
"AmbigNQ-clarifying-question",
|
1064 |
+
"SIGA-nli",
|
1065 |
+
"FOL-nli",
|
1066 |
+
"goal-step-wikihow/order",
|
1067 |
+
"PARADISE",
|
1068 |
+
"doc-nli",
|
1069 |
+
"mctest-nli",
|
1070 |
+
"patent-phrase-similarity",
|
1071 |
+
"natural-language-satisfiability",
|
1072 |
+
"idioms-nli",
|
1073 |
+
"lifecycle-entailment",
|
1074 |
+
"HelpSteer/helpfulness",
|
1075 |
+
"HelpSteer/correctness",
|
1076 |
+
"HelpSteer/coherence",
|
1077 |
+
"HelpSteer/complexity",
|
1078 |
+
"HelpSteer/verbosity",
|
1079 |
+
"HelpSteer2/helpfulness",
|
1080 |
+
"HelpSteer2/correctness",
|
1081 |
+
"HelpSteer2/coherence",
|
1082 |
+
"HelpSteer2/complexity",
|
1083 |
+
"HelpSteer2/verbosity",
|
1084 |
+
"MSciNLI",
|
1085 |
+
"UltraFeedback-paired",
|
1086 |
+
"AES2-essay-scoring",
|
1087 |
+
"english-grading/cohesion",
|
1088 |
+
"english-grading/syntax",
|
1089 |
+
"english-grading/vocabulary",
|
1090 |
+
"english-grading/phraseology",
|
1091 |
+
"english-grading/grammar",
|
1092 |
+
"english-grading/conventions",
|
1093 |
+
"wice",
|
1094 |
+
"hover",
|
1095 |
+
"tasksource_dpo_pairs",
|
1096 |
+
"seahorse_summarization_evaluation",
|
1097 |
"babi_nli",
|
1098 |
"gen_debiased_nli",
|
1099 |
"imppres/presupposition",
|
1100 |
"/prag",
|
1101 |
+
"blimp-2"
|
|
|
1102 |
],
|
1103 |
"torch_dtype": "float32",
|
1104 |
+
"transformers_version": "4.45.2",
|
1105 |
"type_vocab_size": 0,
|
1106 |
"vocab_size": 128100
|
1107 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 737722356
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec79161279746937d440ec246eab9237f18f326059f1831201a5922b23474058
|
3 |
size 737722356
|
tokenizer.json
CHANGED
@@ -2,13 +2,13 @@
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
-
"max_length":
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
9 |
"padding": {
|
10 |
"strategy": {
|
11 |
-
"Fixed":
|
12 |
},
|
13 |
"direction": "Right",
|
14 |
"pad_to_multiple_of": null,
|
|
|
2 |
"version": "1.0",
|
3 |
"truncation": {
|
4 |
"direction": "Right",
|
5 |
+
"max_length": 1280,
|
6 |
"strategy": "LongestFirst",
|
7 |
"stride": 0
|
8 |
},
|
9 |
"padding": {
|
10 |
"strategy": {
|
11 |
+
"Fixed": 1280
|
12 |
},
|
13 |
"direction": "Right",
|
14 |
"pad_to_multiple_of": null,
|
tokenizer_config.json
CHANGED
@@ -47,7 +47,7 @@
|
|
47 |
"do_lower_case": false,
|
48 |
"eos_token": "[SEP]",
|
49 |
"mask_token": "[MASK]",
|
50 |
-
"max_length":
|
51 |
"model_max_length": 1000000000000000019884624838656,
|
52 |
"pad_to_multiple_of": null,
|
53 |
"pad_token": "[PAD]",
|
|
|
47 |
"do_lower_case": false,
|
48 |
"eos_token": "[SEP]",
|
49 |
"mask_token": "[MASK]",
|
50 |
+
"max_length": 1280,
|
51 |
"model_max_length": 1000000000000000019884624838656,
|
52 |
"pad_to_multiple_of": null,
|
53 |
"pad_token": "[PAD]",
|