Datasets:
Modalities:
3D
Size:
10K<n<100K
import json | |
import os | |
import datasets | |
class WindtunnelDataset(datasets.GeneratorBasedBuilder): | |
"""Dataset for loading simulation data with JSON and mesh files""" | |
def _info(self): | |
return datasets.DatasetInfo( | |
features=datasets.Features({ | |
'coeff': datasets.Value('dict'), # JSON file as a dictionary | |
'input': datasets.Value('dict'), # JSON file as a dictionary | |
'input_mesh': datasets.Value('binary'), # OBJ mesh file as binary | |
'openfoam_mesh': datasets.Value('binary'), # OBJ mesh file as binary | |
'pressure_field_mesh': datasets.Value('binary'),# VTK file as binary | |
'streamlines_mesh': datasets.Value('binary'), # PLY file as binary | |
}) | |
) | |
def _split_generators(self, dl_manager): | |
"""Define the splits for the dataset.""" | |
metadata_path = os.path.join(self.config.data_dir, 'metadata.json') | |
with open(metadata_path, 'r') as f: | |
metadata = json.load(f) | |
# Split the dataset into train, validation, and test | |
train_ids, val_ids, test_ids = self._train_val_test_split(metadata) | |
return [ | |
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={'metadata': {id: metadata[id] for id in train_ids}}), | |
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={'metadata': {id: metadata[id] for id in val_ids}}), | |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={'metadata': {id: metadata[id] for id in test_ids}}), | |
] | |
def _generate_examples(self, metadata): | |
"""Generate examples for each split.""" | |
for sim_id, files in metadata.items(): | |
yield sim_id, { | |
'coeff': self._load_json(files['coeff']), | |
'input': self._load_json(files['input']), | |
'input_mesh': self._load_binary(os.path.join(self.config.data_dir, sim_id, 'input_mesh.obj')), | |
'openfoam_mesh': self._load_binary(os.path.join(self.config.data_dir, sim_id, 'openfoam_mesh.obj')), | |
'pressure_field_mesh': self._load_binary(os.path.join(self.config.data_dir, sim_id, 'pressure_field_mesh.vtk')), | |
'streamlines_mesh': self._load_binary(os.path.join(self.config.data_dir, sim_id, 'streamlines_mesh.ply')), | |
} | |
def _load_json(self, file_path): | |
"""Load a JSON file and return it as a dictionary.""" | |
with open(file_path, 'r') as f: | |
return json.load(f) | |
def _load_binary(self, file_path): | |
"""Load a binary file and return its contents.""" | |
with open(file_path, 'rb') as f: | |
return f.read() | |
def _train_val_test_split(self, metadata): | |
"""Implement logic for splitting metadata into train, validation, and test sets.""" | |
keys = list(metadata.keys()) | |
train_ids = keys[:int(0.7 * len(keys))] | |
val_ids = keys[int(0.7 * len(keys)):int(0.85 * len(keys))] | |
test_ids = keys[int(0.85 * len(keys)):] | |
return train_ids, val_ids, test_ids |