Spaces:
Runtime error
Runtime error
Fetching metadata from the HF Docker repository...
update to address log error: ValueError: An event handler (show_demo) didn't receive enough output values (needed: 5, received: 4). Wanted outputs:
bf77b49
verified
-
1.58 kB
-
344 Bytes
-
4.23 kB
- full-blip2-deit-config-2.pth1.47 GB
Detected Pickle imports (50)
- "torch._C._nn.gelu",
- "transformers.models.deit.modeling_deit.DeiTPatchEmbeddings",
- "transformers.models.t5.configuration_t5.T5Config",
- "transformers.models.t5.modeling_t5.T5Attention",
- "transformers.models.deit.configuration_deit.DeiTConfig",
- "transformers.models.t5.modeling_t5.T5LayerCrossAttention",
- "transformers.models.t5.modeling_t5.T5Block",
- "transformers.models.t5.modeling_t5.T5LayerFF",
- "torch.nn.modules.activation.ReLU",
- "torch.nn.modules.conv.Conv2d",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.activation.Tanh",
- "transformers.models.deit.modeling_deit.DeiTSelfOutput",
- "transformers.models.t5.modeling_t5.T5LayerNorm",
- "torch._utils._rebuild_parameter",
- "transformers.models.deit.modeling_deit.DeiTOutput",
- "transformers.generation.configuration_utils.GenerationConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerIntermediate",
- "torch.nn.modules.dropout.Dropout",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerModel",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerAttention",
- "transformers.models.t5.modeling_t5.T5LayerSelfAttention",
- "transformers.models.blip_2.configuration_blip_2.Blip2QFormerConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerSelfOutput",
- "transformers.models.t5.modeling_t5.T5Stack",
- "torch.nn.modules.normalization.LayerNorm",
- "torch.nn.modules.container.ModuleList",
- "transformers.activations.GELUActivation",
- "transformers.models.deit.modeling_deit.DeiTModel",
- "transformers.models.deit.modeling_deit.DeiTAttention",
- "transformers.models.t5.modeling_t5.T5DenseActDense",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerOutput",
- "transformers.models.deit.modeling_deit.DeiTEncoder",
- "torch.float32",
- "transformers.models.blip_2.configuration_blip_2.Blip2Config",
- "transformers.models.deit.modeling_deit.DeiTLayer",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerMultiHeadAttention",
- "transformers.models.blip_2.modeling_blip_2.Blip2ForConditionalGeneration",
- "torch.FloatStorage",
- "__builtin__.set",
- "transformers.models.t5.modeling_t5.T5ForConditionalGeneration",
- "transformers.models.deit.modeling_deit.DeiTIntermediate",
- "transformers.models.deit.modeling_deit.DeiTPooler",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerEncoder",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerLayer",
- "torch.nn.modules.sparse.Embedding",
- "collections.OrderedDict",
- "transformers.models.deit.modeling_deit.DeiTEmbeddings",
- "transformers.models.deit.modeling_deit.DeiTSelfAttention",
- "torch.nn.modules.linear.Linear"
LFS - full-blip2-deit-config-free-form-4-ver-2.pth1.48 GB
Detected Pickle imports (57)
- "transformers.models.deit.modeling_deit.DeiTIntermediate",
- "torch.nn.modules.activation.Tanh",
- "__builtin__.set",
- "torch._C._nn.gelu",
- "transformers.models.t5.modeling_t5.T5LayerSelfAttention",
- "torch.nn.modules.activation.ReLU",
- "torch.float32",
- "transformers.models.deit.modeling_deit.DeiTSelfAttention",
- "transformers.models.t5.configuration_t5.T5Config",
- "transformers.models.t5.modeling_t5.T5LayerFF",
- "torch.nn.modules.container.ParameterDict",
- "torch._utils._rebuild_tensor_v2",
- "transformers.models.t5.modeling_t5.T5Block",
- "transformers.models.t5.modeling_t5.T5DenseActDense",
- "transformers.models.t5.modeling_t5.T5Attention",
- "transformers.models.deit.modeling_deit.DeiTPooler",
- "transformers.models.deit.modeling_deit.DeiTAttention",
- "torch.FloatStorage",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerOutput",
- "transformers.models.t5.modeling_t5.T5ForConditionalGeneration",
- "torch.nn.modules.container.ModuleDict",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerModel",
- "transformers.models.blip_2.configuration_blip_2.Blip2QFormerConfig",
- "torch.nn.modules.normalization.LayerNorm",
- "peft.tuners.lora.layer.Linear",
- "transformers.models.t5.modeling_t5.T5LayerCrossAttention",
- "transformers.generation.configuration_utils.GenerationConfig",
- "peft.utils.peft_types.PeftType",
- "torch.nn.modules.dropout.Dropout",
- "torch.nn.modules.linear.Linear",
- "transformers.activations.GELUActivation",
- "transformers.models.deit.modeling_deit.DeiTOutput",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerEncoder",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerAttention",
- "transformers.models.deit.configuration_deit.DeiTConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerMultiHeadAttention",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerLayer",
- "transformers.models.t5.modeling_t5.T5LayerNorm",
- "peft.tuners.lora.model.LoraModel",
- "transformers.models.blip_2.modeling_blip_2.Blip2ForConditionalGeneration",
- "torch.nn.modules.container.ModuleList",
- "transformers.models.deit.modeling_deit.DeiTLayer",
- "peft.tuners.lora.config.LoraConfig",
- "transformers.models.deit.modeling_deit.DeiTEncoder",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerSelfOutput",
- "transformers.models.blip_2.configuration_blip_2.Blip2Config",
- "peft.peft_model.PeftModel",
- "transformers.models.deit.modeling_deit.DeiTSelfOutput",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerIntermediate",
- "collections.OrderedDict",
- "torch._utils._rebuild_parameter",
- "transformers.models.deit.modeling_deit.DeiTModel",
- "transformers.models.t5.modeling_t5.T5Stack",
- "transformers.models.deit.modeling_deit.DeiTEmbeddings",
- "torch.nn.modules.conv.Conv2d",
- "transformers.models.deit.modeling_deit.DeiTPatchEmbeddings",
- "torch.nn.modules.sparse.Embedding"
LFS - full-blip2-deit.pth1.51 GB
Detected Pickle imports (47)
- "transformers.models.opt.modeling_opt.OPTModel",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerOutput",
- "transformers.models.deit.modeling_deit.DeiTEmbeddings",
- "torch.nn.modules.sparse.Embedding",
- "transformers.models.opt.modeling_opt.OPTAttention",
- "transformers.models.deit.modeling_deit.DeiTEncoder",
- "transformers.models.deit.modeling_deit.DeiTSelfOutput",
- "transformers.models.opt.modeling_opt.OPTLearnedPositionalEmbedding",
- "transformers.models.deit.modeling_deit.DeiTPatchEmbeddings",
- "transformers.models.deit.modeling_deit.DeiTSelfAttention",
- "torch.nn.modules.container.ModuleList",
- "transformers.models.blip_2.configuration_blip_2.Blip2QFormerConfig",
- "torch.nn.modules.normalization.LayerNorm",
- "transformers.models.opt.modeling_opt.OPTDecoderLayer",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerModel",
- "torch.nn.modules.activation.ReLU",
- "transformers.generation.configuration_utils.GenerationConfig",
- "torch._C._nn.gelu",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerMultiHeadAttention",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerEncoder",
- "torch.FloatStorage",
- "transformers.activations.GELUActivation",
- "transformers.models.deit.modeling_deit.DeiTPooler",
- "transformers.models.opt.modeling_opt.OPTForCausalLM",
- "torch.nn.modules.activation.Tanh",
- "transformers.models.deit.configuration_deit.DeiTConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerSelfOutput",
- "torch._utils._rebuild_parameter",
- "transformers.models.blip_2.configuration_blip_2.Blip2Config",
- "torch.float32",
- "transformers.models.deit.modeling_deit.DeiTIntermediate",
- "transformers.models.deit.modeling_deit.DeiTOutput",
- "collections.OrderedDict",
- "transformers.models.deit.modeling_deit.DeiTModel",
- "transformers.models.deit.modeling_deit.DeiTAttention",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerAttention",
- "torch.nn.modules.dropout.Dropout",
- "torch._utils._rebuild_tensor_v2",
- "transformers.models.blip_2.modeling_blip_2.Blip2ForConditionalGeneration",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerLayer",
- "transformers.models.deit.modeling_deit.DeiTLayer",
- "torch.nn.modules.linear.Linear",
- "torch.nn.modules.conv.Conv2d",
- "transformers.models.opt.configuration_opt.OPTConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerIntermediate",
- "transformers.models.opt.modeling_opt.OPTDecoder",
- "__builtin__.set"
LFS - full_config_blip2-deit-051.47 GB
Detected Pickle imports (50)
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerIntermediate",
- "transformers.models.deit.modeling_deit.DeiTModel",
- "transformers.models.deit.modeling_deit.DeiTPooler",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerOutput",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerLayer",
- "torch.float32",
- "transformers.models.deit.modeling_deit.DeiTSelfAttention",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerEncoder",
- "transformers.models.t5.modeling_t5.T5Attention",
- "transformers.models.t5.modeling_t5.T5DenseActDense",
- "transformers.models.deit.modeling_deit.DeiTEncoder",
- "torch.FloatStorage",
- "torch._utils._rebuild_tensor_v2",
- "transformers.models.blip_2.modeling_blip_2.Blip2ForConditionalGeneration",
- "transformers.models.deit.configuration_deit.DeiTConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerMultiHeadAttention",
- "transformers.models.deit.modeling_deit.DeiTPatchEmbeddings",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerSelfOutput",
- "transformers.models.deit.modeling_deit.DeiTEmbeddings",
- "transformers.models.deit.modeling_deit.DeiTOutput",
- "transformers.models.t5.modeling_t5.T5LayerFF",
- "transformers.models.t5.configuration_t5.T5Config",
- "transformers.models.t5.modeling_t5.T5LayerSelfAttention",
- "torch.nn.modules.dropout.Dropout",
- "torch.nn.modules.container.ModuleList",
- "transformers.models.deit.modeling_deit.DeiTAttention",
- "transformers.models.t5.modeling_t5.T5ForConditionalGeneration",
- "transformers.models.blip_2.configuration_blip_2.Blip2Config",
- "torch.nn.modules.normalization.LayerNorm",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerModel",
- "transformers.activations.GELUActivation",
- "transformers.models.t5.modeling_t5.T5Stack",
- "torch.nn.modules.activation.Tanh",
- "collections.OrderedDict",
- "transformers.generation.configuration_utils.GenerationConfig",
- "torch.nn.modules.linear.Linear",
- "transformers.models.blip_2.configuration_blip_2.Blip2QFormerConfig",
- "transformers.models.deit.modeling_deit.DeiTSelfOutput",
- "__builtin__.set",
- "transformers.models.deit.modeling_deit.DeiTLayer",
- "torch._C._nn.gelu",
- "torch.nn.modules.sparse.Embedding",
- "transformers.models.deit.modeling_deit.DeiTIntermediate",
- "transformers.models.t5.modeling_t5.T5LayerCrossAttention",
- "transformers.models.t5.modeling_t5.T5LayerNorm",
- "torch.nn.modules.activation.ReLU",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerAttention",
- "torch._utils._rebuild_parameter",
- "torch.nn.modules.conv.Conv2d",
- "transformers.models.t5.modeling_t5.T5Block"
LFS - full_config_blip2-deit-05.pth1.47 GB
Detected Pickle imports (50)
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerIntermediate",
- "transformers.models.deit.modeling_deit.DeiTModel",
- "transformers.models.deit.modeling_deit.DeiTPooler",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerOutput",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerLayer",
- "torch.float32",
- "transformers.models.deit.modeling_deit.DeiTSelfAttention",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerEncoder",
- "transformers.models.t5.modeling_t5.T5Attention",
- "transformers.models.t5.modeling_t5.T5DenseActDense",
- "transformers.models.deit.modeling_deit.DeiTEncoder",
- "torch.FloatStorage",
- "torch._utils._rebuild_tensor_v2",
- "transformers.models.blip_2.modeling_blip_2.Blip2ForConditionalGeneration",
- "transformers.models.deit.configuration_deit.DeiTConfig",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerMultiHeadAttention",
- "transformers.models.deit.modeling_deit.DeiTPatchEmbeddings",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerSelfOutput",
- "transformers.models.deit.modeling_deit.DeiTEmbeddings",
- "transformers.models.deit.modeling_deit.DeiTOutput",
- "transformers.models.t5.modeling_t5.T5LayerFF",
- "transformers.models.t5.configuration_t5.T5Config",
- "transformers.models.t5.modeling_t5.T5LayerSelfAttention",
- "torch.nn.modules.dropout.Dropout",
- "torch.nn.modules.container.ModuleList",
- "transformers.models.deit.modeling_deit.DeiTAttention",
- "transformers.models.t5.modeling_t5.T5ForConditionalGeneration",
- "transformers.models.blip_2.configuration_blip_2.Blip2Config",
- "torch.nn.modules.normalization.LayerNorm",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerModel",
- "transformers.activations.GELUActivation",
- "transformers.models.t5.modeling_t5.T5Stack",
- "torch.nn.modules.activation.Tanh",
- "collections.OrderedDict",
- "transformers.generation.configuration_utils.GenerationConfig",
- "torch.nn.modules.linear.Linear",
- "transformers.models.blip_2.configuration_blip_2.Blip2QFormerConfig",
- "transformers.models.deit.modeling_deit.DeiTSelfOutput",
- "__builtin__.set",
- "transformers.models.deit.modeling_deit.DeiTLayer",
- "torch._C._nn.gelu",
- "torch.nn.modules.sparse.Embedding",
- "transformers.models.deit.modeling_deit.DeiTIntermediate",
- "transformers.models.t5.modeling_t5.T5LayerCrossAttention",
- "transformers.models.t5.modeling_t5.T5LayerNorm",
- "torch.nn.modules.activation.ReLU",
- "transformers.models.blip_2.modeling_blip_2.Blip2QFormerAttention",
- "torch._utils._rebuild_parameter",
- "torch.nn.modules.conv.Conv2d",
- "transformers.models.t5.modeling_t5.T5Block"
LFS -
77 Bytes