tcm03
collate_fn for dataloader and extract vision features
484e90b
import sys
from pathlib import Path
sys.path.append(str(Path.cwd()))
from annotation.utils import get_optimal_workers
import os
import argparse
from typing import List, Dict
from mm_datautils import process_video_frames
from preprocessor import CambrianConfig, CambrianEncoders
import torch
from safetensors.torch import save_file
from collections import defaultdict
import logging
from multiprocessing import cpu_count
from entube_dataset import EnTubeDataset, collate_fn
from torch.utils.data import Dataset, DataLoader
from transformers import BaseImageProcessor
# Configure logging
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
def get_optimal_workers() -> int:
"""Determine the optimal number of workers based on available CPU cores."""
try:
return max(1, cpu_count() - 1) # Leave one core free
except (NotImplementedError, ValueError):
return 1 # Fallback to a single worker in case of an error
def extract_features(processor: CambrianEncoders, file_path: str, file_name: str) -> Dict[str, torch.Tensor]:
try:
video, image_sizes = process_video_frames(file_path)
image_aux_features_list = processor.prepare_mm_features(images=video, image_sizes=image_sizes)
return {
file_name + '-siglip': image_aux_features_list[0],
file_name + '-dino': image_aux_features_list[1]
}
except Exception as e:
logging.error(f"Error processing {file_path}: {e}")
return {}
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument(
'--folders',
type=str,
nargs='+',
required=True,
help="List of folder paths to video data"
)
parser.add_argument(
'--output_file',
type = str,
default = 'entube_tensors.safetensors',
help = 'Safetensor file to store embeddings of EnTube dataset by vision encoders'
)
parser.add_argument(
'--config_file',
type = str,
default = 'config.json',
help = 'Path to configuration file of encoders parameters'
)
args = parser.parse_args()
cambrianConfig = CambrianConfig.from_json_file(args.config_file)
processor = CambrianEncoders(cambrianConfig)
image_processors = []
if not processor.vision_tower_aux_list[0].is_loaded:
processor.vision_tower_aux_list[0].load_model()
image_processors.append(processor.vision_tower_aux_list[0].image_processor)
# for vision_tower_aux in processor.vision_tower_aux_list:
# if not vision_tower_aux.is_loaded:
# vision_tower_aux.load_model()
# image_processors.append(vision_tower_aux.image_processor)
folder_paths: List[str] = args.folders
data_tensor = dict()
device = 'cuda' if torch.cuda.is_available() else 'cpu'
entube_dataset = EnTubeDataset(folder_paths, image_processors, device)
dataloader = DataLoader(
entube_dataset,
batch_size=4,
collate_fn=collate_fn,
# num_workers=get_optimal_workers()
num_workers=1
)
for batch_idx, (videos, image_sizes) in enumerate(dataloader):
print(f"Processing batch {batch_idx + 1}/{len(dataloader)}")
assert isinstance(videos, list), "List of videos features for each processor (vision encoder)"
assert isinstance(videos[0], list) or isinstance(videos[0], torch.Tensor), "List of videos in the batch"
image_aux_features_list = processor.prepare_mm_features(videos, image_sizes)
for i, image_aux_features in enumerate(image_aux_features_list):
print(f"@tcm: In main(): image_aux_features[{i}].shape={image_aux_features.shape}")
break
# save_file(dict(data_tensor), args.output_file)