HPT1_5-Edge / configuration_qformer.py
yuchengz's picture
hpt1.5 edge commit
a175ed9
raw
history blame contribute delete
774 Bytes
# Copyright (c) OpenMMLab. All rights reserved.
from transformers import PretrainedConfig
class QformerConfig(PretrainedConfig):
model_type = 'qformer'
_auto_class = 'AutoConfig'
def __init__(
self,
num_query_token=32,
visual_hidden_size=4096,
llm_hidden_size=768,
cross_attention_freq=2,
bert="bert-base-uncased",
bias=True,
qformer_pth=None,
**kwargs,
):
self.num_query_token=num_query_token
self.visual_hidden_size = visual_hidden_size
self.llm_hidden_size = llm_hidden_size
self.bias = bias
self.bert = bert
self.cross_attention_freq = cross_attention_freq
self.qformer_pth = qformer_pth
super().__init__(**kwargs)