l_qa_t5 / head_config.json
selimyagci's picture
Upload model
e0b17eb verified
raw
history blame contribute delete
384 Bytes
{
"config": {
"activation_function": null,
"bias": false,
"head_type": "seq2seq_lm",
"label2id": null,
"layer_norm": false,
"layers": 1,
"shift_labels": false,
"vocab_size": 32128
},
"hidden_size": 768,
"model_class": "T5AdapterModel",
"model_name": "google-t5/t5-base",
"model_type": "t5",
"name": "l_qa",
"version": "adapters.1.0.0"
}