kleinay commited on
Commit
b13892a
1 Parent(s): a053034

Upload T5ForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +4 -1
  2. generation_config.json +7 -0
  3. pytorch_model.bin +2 -2
config.json CHANGED
@@ -9,6 +9,7 @@
9
  "d_model": 512,
10
  "debug_mode": false,
11
  "decoder_start_token_id": 0,
 
12
  "description": "optimal joint config from sweep, mainly for qanom",
13
  "dir_switch": "joint_optimal",
14
  "do_eval_on": "validation",
@@ -21,6 +22,7 @@
21
  "gradient_accumulation_steps": 14,
22
  "initializer_factor": 1.0,
23
  "is_encoder_decoder": true,
 
24
  "layer_norm_epsilon": 1e-06,
25
  "learning_rate": 0.001,
26
  "load_best_model_at_end": true,
@@ -75,6 +77,7 @@
75
  "use_bilateral_predicate_marker": true
76
  },
77
  "qanom_joint_factor": 14,
 
78
  "relative_attention_num_buckets": 32,
79
  "save_steps": 500,
80
  "save_strategy": "steps",
@@ -112,7 +115,7 @@
112
  "torch_dtype": "float32",
113
  "train_dataset": "joint_qanom",
114
  "train_epochs": 20,
115
- "transformers_version": "4.14.1",
116
  "use_bilateral_predicate_marker": true,
117
  "use_cache": true,
118
  "vocab_size": 32101
 
9
  "d_model": 512,
10
  "debug_mode": false,
11
  "decoder_start_token_id": 0,
12
+ "dense_act_fn": "relu",
13
  "description": "optimal joint config from sweep, mainly for qanom",
14
  "dir_switch": "joint_optimal",
15
  "do_eval_on": "validation",
 
22
  "gradient_accumulation_steps": 14,
23
  "initializer_factor": 1.0,
24
  "is_encoder_decoder": true,
25
+ "is_gated_act": false,
26
  "layer_norm_epsilon": 1e-06,
27
  "learning_rate": 0.001,
28
  "load_best_model_at_end": true,
 
77
  "use_bilateral_predicate_marker": true
78
  },
79
  "qanom_joint_factor": 14,
80
+ "relative_attention_max_distance": 128,
81
  "relative_attention_num_buckets": 32,
82
  "save_steps": 500,
83
  "save_strategy": "steps",
 
115
  "torch_dtype": "float32",
116
  "train_dataset": "joint_qanom",
117
  "train_epochs": 20,
118
+ "transformers_version": "4.26.1",
119
  "use_bilateral_predicate_marker": true,
120
  "use_cache": true,
121
  "vocab_size": 32101
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.26.1"
7
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:67987af0aa8b7d2584e73cfaa33571bae84c52e4bc3534c5e896e1e7824343bd
3
- size 242028475
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd5299f5b0ff532dfaf236df1d385a9f464c7163fee79f6c351410c96bc3554f
3
+ size 242016842