CLIPModel / config.py
ibrim's picture
Upload 7 files
6bea3f9 verified
raw
history blame
868 Bytes
import torch
debug = True
image_path = "/raid/users/mohammadibrahim-st/TSAI/OpenAI-CLIP/Flicker-8k/Images"
captions_path = "/raid/users/mohammadibrahim-st/TSAI/OpenAI-CLIP/Flicker-8k"
batch_size = 20
num_workers = 0
lr = 1e-3
weight_decay = 1e-3
patience = 2
factor = 0.5
epochs = 5
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model_name = 'resnet50'
image_embedding = 2048
text_encoder_model = "/raid/users/mohammadibrahim-st/Models/BertDistil"
text_embedding = 768
text_tokenizer = "/raid/users/mohammadibrahim-st/Models/BertDistil"
max_length = 200
pretrained = False # for both image encoder and text encoder
trainable = False # for both image encoder and text encoder
temperature = 1.0
# image size
size = 224
# for projection head; used for both image and text encoders
num_projection_layers = 1
projection_dim = 256
dropout = 0.1