|
import os |
|
import tarfile |
|
|
|
import h5py |
|
import numpy as np |
|
import torch |
|
import torch.utils.data as data |
|
from PIL import Image |
|
from scipy import io |
|
from torchvision.datasets.utils import download_url |
|
|
|
DATASET_YEAR_DICT = { |
|
"2012": { |
|
"url": "http://host.robots.ox.ac.uk/pascal/VOC/voc2012/VOCtrainval_11-May-2012.tar", |
|
"filename": "VOCtrainval_11-May-2012.tar", |
|
"md5": "6cd6e144f989b92b3379bac3b3de84fd", |
|
"base_dir": "VOCdevkit/VOC2012", |
|
}, |
|
"2011": { |
|
"url": "http://host.robots.ox.ac.uk/pascal/VOC/voc2011/VOCtrainval_25-May-2011.tar", |
|
"filename": "VOCtrainval_25-May-2011.tar", |
|
"md5": "6c3384ef61512963050cb5d687e5bf1e", |
|
"base_dir": "TrainVal/VOCdevkit/VOC2011", |
|
}, |
|
"2010": { |
|
"url": "http://host.robots.ox.ac.uk/pascal/VOC/voc2010/VOCtrainval_03-May-2010.tar", |
|
"filename": "VOCtrainval_03-May-2010.tar", |
|
"md5": "da459979d0c395079b5c75ee67908abb", |
|
"base_dir": "VOCdevkit/VOC2010", |
|
}, |
|
"2009": { |
|
"url": "http://host.robots.ox.ac.uk/pascal/VOC/voc2009/VOCtrainval_11-May-2009.tar", |
|
"filename": "VOCtrainval_11-May-2009.tar", |
|
"md5": "59065e4b188729180974ef6572f6a212", |
|
"base_dir": "VOCdevkit/VOC2009", |
|
}, |
|
"2008": { |
|
"url": "http://host.robots.ox.ac.uk/pascal/VOC/voc2008/VOCtrainval_14-Jul-2008.tar", |
|
"filename": "VOCtrainval_11-May-2012.tar", |
|
"md5": "2629fa636546599198acfcfbfcf1904a", |
|
"base_dir": "VOCdevkit/VOC2008", |
|
}, |
|
"2007": { |
|
"url": "http://host.robots.ox.ac.uk/pascal/VOC/voc2007/VOCtrainval_06-Nov-2007.tar", |
|
"filename": "VOCtrainval_06-Nov-2007.tar", |
|
"md5": "c52e279531787c972589f7e41ab4ae64", |
|
"base_dir": "VOCdevkit/VOC2007", |
|
}, |
|
} |
|
|
|
|
|
class VOCSegmentation(data.Dataset): |
|
"""`Pascal VOC <http://host.robots.ox.ac.uk/pascal/VOC/>`_ Segmentation Dataset. |
|
|
|
Args: |
|
root (string): Root directory of the VOC Dataset. |
|
year (string, optional): The dataset year, supports years 2007 to 2012. |
|
image_set (string, optional): Select the image_set to use, ``train``, ``trainval`` or ``val`` |
|
download (bool, optional): If true, downloads the dataset from the internet and |
|
puts it in root directory. If dataset is already downloaded, it is not |
|
downloaded again. |
|
transform (callable, optional): A function/transform that takes in an PIL image |
|
and returns a transformed version. E.g, ``transforms.RandomCrop`` |
|
target_transform (callable, optional): A function/transform that takes in the |
|
target and transforms it. |
|
""" |
|
|
|
CLASSES = 20 |
|
CLASSES_NAMES = [ |
|
"aeroplane", |
|
"bicycle", |
|
"bird", |
|
"boat", |
|
"bottle", |
|
"bus", |
|
"car", |
|
"cat", |
|
"chair", |
|
"cow", |
|
"diningtable", |
|
"dog", |
|
"horse", |
|
"motorbike", |
|
"person", |
|
"potted-plant", |
|
"sheep", |
|
"sofa", |
|
"train", |
|
"tvmonitor", |
|
"ambigious", |
|
] |
|
|
|
def __init__( |
|
self, |
|
root, |
|
year="2012", |
|
image_set="train", |
|
download=False, |
|
transform=None, |
|
target_transform=None, |
|
): |
|
self.root = os.path.expanduser(root) |
|
self.year = year |
|
self.url = DATASET_YEAR_DICT[year]["url"] |
|
self.filename = DATASET_YEAR_DICT[year]["filename"] |
|
self.md5 = DATASET_YEAR_DICT[year]["md5"] |
|
self.transform = transform |
|
self.target_transform = target_transform |
|
self.image_set = image_set |
|
base_dir = DATASET_YEAR_DICT[year]["base_dir"] |
|
voc_root = os.path.join(self.root, base_dir) |
|
image_dir = os.path.join(voc_root, "JPEGImages") |
|
mask_dir = os.path.join(voc_root, "SegmentationClass") |
|
|
|
if download: |
|
download_extract(self.url, self.root, self.filename, self.md5) |
|
|
|
if not os.path.isdir(voc_root): |
|
raise RuntimeError( |
|
"Dataset not found or corrupted." |
|
+ " You can use download=True to download it" |
|
) |
|
|
|
splits_dir = os.path.join(voc_root, "ImageSets/Segmentation") |
|
|
|
split_f = os.path.join(splits_dir, image_set.rstrip("\n") + ".txt") |
|
|
|
if not os.path.exists(split_f): |
|
raise ValueError( |
|
'Wrong image_set entered! Please use image_set="train" ' |
|
'or image_set="trainval" or image_set="val"' |
|
) |
|
|
|
with open(os.path.join(split_f), "r") as f: |
|
file_names = [x.strip() for x in f.readlines()] |
|
|
|
self.images = [os.path.join(image_dir, x + ".jpg") for x in file_names] |
|
self.masks = [os.path.join(mask_dir, x + ".png") for x in file_names] |
|
assert len(self.images) == len(self.masks) |
|
|
|
def __getitem__(self, index): |
|
""" |
|
Args: |
|
index (int): Index |
|
|
|
Returns: |
|
tuple: (image, target) where target is the image segmentation. |
|
""" |
|
img = Image.open(self.images[index]).convert("RGB") |
|
target = Image.open(self.masks[index]) |
|
|
|
if self.transform is not None: |
|
img = self.transform(img) |
|
|
|
if self.target_transform is not None: |
|
target = np.array(self.target_transform(target)).astype("int32") |
|
target[target == 255] = -1 |
|
target = torch.from_numpy(target).long() |
|
|
|
return img, target |
|
|
|
@staticmethod |
|
def _mask_transform(mask): |
|
target = np.array(mask).astype("int32") |
|
target[target == 255] = -1 |
|
return torch.from_numpy(target).long() |
|
|
|
def __len__(self): |
|
return len(self.images) |
|
|
|
@property |
|
def pred_offset(self): |
|
return 0 |
|
|
|
|
|
class VOCClassification(data.Dataset): |
|
"""`Pascal VOC <http://host.robots.ox.ac.uk/pascal/VOC/>`_ Segmentation Dataset. |
|
|
|
Args: |
|
root (string): Root directory of the VOC Dataset. |
|
year (string, optional): The dataset year, supports years 2007 to 2012. |
|
image_set (string, optional): Select the image_set to use, ``train``, ``trainval`` or ``val`` |
|
download (bool, optional): If true, downloads the dataset from the internet and |
|
puts it in root directory. If dataset is already downloaded, it is not |
|
downloaded again. |
|
transform (callable, optional): A function/transform that takes in an PIL image |
|
and returns a transformed version. E.g, ``transforms.RandomCrop`` |
|
""" |
|
|
|
CLASSES = 20 |
|
|
|
def __init__( |
|
self, root, year="2012", image_set="train", download=False, transform=None |
|
): |
|
self.root = os.path.expanduser(root) |
|
self.year = year |
|
self.url = DATASET_YEAR_DICT[year]["url"] |
|
self.filename = DATASET_YEAR_DICT[year]["filename"] |
|
self.md5 = DATASET_YEAR_DICT[year]["md5"] |
|
self.transform = transform |
|
self.image_set = image_set |
|
base_dir = DATASET_YEAR_DICT[year]["base_dir"] |
|
voc_root = os.path.join(self.root, base_dir) |
|
image_dir = os.path.join(voc_root, "JPEGImages") |
|
mask_dir = os.path.join(voc_root, "SegmentationClass") |
|
|
|
if download: |
|
download_extract(self.url, self.root, self.filename, self.md5) |
|
|
|
if not os.path.isdir(voc_root): |
|
raise RuntimeError( |
|
"Dataset not found or corrupted." |
|
+ " You can use download=True to download it" |
|
) |
|
|
|
splits_dir = os.path.join(voc_root, "ImageSets/Segmentation") |
|
|
|
split_f = os.path.join(splits_dir, image_set.rstrip("\n") + ".txt") |
|
|
|
if not os.path.exists(split_f): |
|
raise ValueError( |
|
'Wrong image_set entered! Please use image_set="train" ' |
|
'or image_set="trainval" or image_set="val"' |
|
) |
|
|
|
with open(os.path.join(split_f), "r") as f: |
|
file_names = [x.strip() for x in f.readlines()] |
|
|
|
self.images = [os.path.join(image_dir, x + ".jpg") for x in file_names] |
|
self.masks = [os.path.join(mask_dir, x + ".png") for x in file_names] |
|
assert len(self.images) == len(self.masks) |
|
|
|
def __getitem__(self, index): |
|
""" |
|
Args: |
|
index (int): Index |
|
|
|
Returns: |
|
tuple: (image, target) where target is the image segmentation. |
|
""" |
|
img = Image.open(self.images[index]).convert("RGB") |
|
target = Image.open(self.masks[index]) |
|
|
|
|
|
|
|
if self.transform is not None: |
|
img, target = self.transform(img, target) |
|
|
|
visible_classes = np.unique(target) |
|
labels = torch.zeros(self.CLASSES) |
|
for id in visible_classes: |
|
if id not in (0, 255): |
|
labels[id - 1].fill_(1) |
|
|
|
return img, labels |
|
|
|
def __len__(self): |
|
return len(self.images) |
|
|
|
|
|
class VOCSBDClassification(data.Dataset): |
|
"""`Pascal VOC <http://host.robots.ox.ac.uk/pascal/VOC/>`_ Segmentation Dataset. |
|
|
|
Args: |
|
root (string): Root directory of the VOC Dataset. |
|
year (string, optional): The dataset year, supports years 2007 to 2012. |
|
image_set (string, optional): Select the image_set to use, ``train``, ``trainval`` or ``val`` |
|
download (bool, optional): If true, downloads the dataset from the internet and |
|
puts it in root directory. If dataset is already downloaded, it is not |
|
downloaded again. |
|
transform (callable, optional): A function/transform that takes in an PIL image |
|
and returns a transformed version. E.g, ``transforms.RandomCrop`` |
|
""" |
|
|
|
CLASSES = 20 |
|
|
|
def __init__( |
|
self, |
|
root, |
|
sbd_root, |
|
year="2012", |
|
image_set="train", |
|
download=False, |
|
transform=None, |
|
): |
|
self.root = os.path.expanduser(root) |
|
self.sbd_root = os.path.expanduser(sbd_root) |
|
self.year = year |
|
self.url = DATASET_YEAR_DICT[year]["url"] |
|
self.filename = DATASET_YEAR_DICT[year]["filename"] |
|
self.md5 = DATASET_YEAR_DICT[year]["md5"] |
|
self.transform = transform |
|
self.image_set = image_set |
|
base_dir = DATASET_YEAR_DICT[year]["base_dir"] |
|
voc_root = os.path.join(self.root, base_dir) |
|
image_dir = os.path.join(voc_root, "JPEGImages") |
|
mask_dir = os.path.join(voc_root, "SegmentationClass") |
|
sbd_image_dir = os.path.join(sbd_root, "img") |
|
sbd_mask_dir = os.path.join(sbd_root, "cls") |
|
|
|
if download: |
|
download_extract(self.url, self.root, self.filename, self.md5) |
|
|
|
if not os.path.isdir(voc_root): |
|
raise RuntimeError( |
|
"Dataset not found or corrupted." |
|
+ " You can use download=True to download it" |
|
) |
|
|
|
splits_dir = os.path.join(voc_root, "ImageSets/Segmentation") |
|
|
|
split_f = os.path.join(splits_dir, image_set.rstrip("\n") + ".txt") |
|
sbd_split = os.path.join(sbd_root, "train.txt") |
|
|
|
if not os.path.exists(split_f): |
|
raise ValueError( |
|
'Wrong image_set entered! Please use image_set="train" ' |
|
'or image_set="trainval" or image_set="val"' |
|
) |
|
|
|
with open(os.path.join(split_f), "r") as f: |
|
voc_file_names = [x.strip() for x in f.readlines()] |
|
|
|
with open(os.path.join(sbd_split), "r") as f: |
|
sbd_file_names = [x.strip() for x in f.readlines()] |
|
|
|
self.images = [os.path.join(image_dir, x + ".jpg") for x in voc_file_names] |
|
self.images += [os.path.join(sbd_image_dir, x + ".jpg") for x in sbd_file_names] |
|
self.masks = [os.path.join(mask_dir, x + ".png") for x in voc_file_names] |
|
self.masks += [os.path.join(sbd_mask_dir, x + ".mat") for x in sbd_file_names] |
|
assert len(self.images) == len(self.masks) |
|
|
|
def __getitem__(self, index): |
|
""" |
|
Args: |
|
index (int): Index |
|
|
|
Returns: |
|
tuple: (image, target) where target is the image segmentation. |
|
""" |
|
img = Image.open(self.images[index]).convert("RGB") |
|
mask_path = self.masks[index] |
|
if mask_path[-3:] == "mat": |
|
target = io.loadmat(mask_path, struct_as_record=False, squeeze_me=True)[ |
|
"GTcls" |
|
].Segmentation |
|
target = Image.fromarray(target, mode="P") |
|
else: |
|
target = Image.open(self.masks[index]) |
|
|
|
if self.transform is not None: |
|
img, target = self.transform(img, target) |
|
|
|
visible_classes = np.unique(target) |
|
labels = torch.zeros(self.CLASSES) |
|
for id in visible_classes: |
|
if id not in (0, 255): |
|
labels[id - 1].fill_(1) |
|
|
|
return img, labels |
|
|
|
def __len__(self): |
|
return len(self.images) |
|
|
|
|
|
def download_extract(url, root, filename, md5): |
|
download_url(url, root, filename, md5) |
|
with tarfile.open(os.path.join(root, filename), "r") as tar: |
|
tar.extractall(path=root) |
|
|
|
|
|
class VOCResults(data.Dataset): |
|
CLASSES = 20 |
|
CLASSES_NAMES = [ |
|
"aeroplane", |
|
"bicycle", |
|
"bird", |
|
"boat", |
|
"bottle", |
|
"bus", |
|
"car", |
|
"cat", |
|
"chair", |
|
"cow", |
|
"diningtable", |
|
"dog", |
|
"horse", |
|
"motorbike", |
|
"person", |
|
"potted-plant", |
|
"sheep", |
|
"sofa", |
|
"train", |
|
"tvmonitor", |
|
"ambigious", |
|
] |
|
|
|
def __init__(self, path): |
|
super(VOCResults, self).__init__() |
|
|
|
self.path = os.path.join(path, "results.hdf5") |
|
self.data = None |
|
|
|
print("Reading dataset length...") |
|
with h5py.File(self.path, "r") as f: |
|
self.data_length = len(f["/image"]) |
|
|
|
def __len__(self): |
|
return self.data_length |
|
|
|
def __getitem__(self, item): |
|
if self.data is None: |
|
self.data = h5py.File(self.path, "r") |
|
|
|
image = torch.tensor(self.data["image"][item]) |
|
vis = torch.tensor(self.data["vis"][item]) |
|
target = torch.tensor(self.data["target"][item]) |
|
class_pred = torch.tensor(self.data["class_pred"][item]) |
|
|
|
return image, vis, target, class_pred |
|
|