Spaces:
Runtime error
Runtime error
File size: 2,101 Bytes
128757a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
import torch
import maskrcnn_benchmark.utils.dist as dist
def normalized_positive_map(positive_map):
positive_map = positive_map.float()
positive_map_num_pos = positive_map.sum(2)
positive_map_num_pos[positive_map_num_pos == 0] = 1e-6
positive_map = positive_map / positive_map_num_pos.unsqueeze(-1)
return positive_map
def pad_tensor_given_dim_length(tensor, dim, length, padding_value=0, batch_first=True):
new_size = list(tensor.size()[:dim]) + [length] + list(tensor.size()[dim + 1:])
out_tensor = tensor.data.new(*new_size).fill_(padding_value)
if batch_first:
out_tensor[:, :tensor.size(1), ...] = tensor
else:
out_tensor[:tensor.size(0), ...] = tensor
return out_tensor
def pad_random_negative_tensor_given_length(positive_tensor, negative_padding_tensor, length=None):
assert positive_tensor.shape[0] + negative_padding_tensor.shape[0] == length
return torch.cat((positive_tensor, negative_padding_tensor), dim=0)
def gather_tensors(tensor):
"""
Performs all_gather operation on the provided tensors.
*** Warning ***: torch.distributed.all_gather has no gradient.
"""
if not dist.is_dist_avail_and_initialized():
return torch.stack([tensor], dim=0)
total = dist.get_world_size()
rank = torch.distributed.get_rank()
# gathered_normalized_img_emb = [torch.zeros_like(normalized_img_emb) for _ in range(total)]
# torch.distributed.all_gather(gathered_normalized_img_emb, normalized_img_emb)
tensors_gather = [
torch.zeros_like(tensor)
for _ in range(total)
]
torch.distributed.all_gather(tensors_gather, tensor, async_op=False)
# need to do this to restore propagation of the gradients
tensors_gather[rank] = tensor
output = torch.stack(tensors_gather, dim=0)
return output
def convert_to_roi_format(boxes):
concat_boxes = boxes.bbox
device, dtype = concat_boxes.device, concat_boxes.dtype
ids = torch.full((len(boxes), 1), 0, dtype=dtype, device=device)
rois = torch.cat([ids, concat_boxes], dim=1)
return rois |