Spaces:
Running
on
L4
Running
on
L4
from typing import Optional | |
import numpy as np | |
import torch as th | |
import torch.nn.functional as F | |
import torch.nn as nn | |
from sklearn.neighbors import KDTree | |
import logging | |
logger = logging.getLogger(__name__) | |
# NOTE: we need pytorch3d primarily for UV rasterization things | |
from pytorch3d.renderer.mesh.rasterize_meshes import rasterize_meshes | |
from pytorch3d.structures import Meshes | |
from typing import Union, Optional, Tuple | |
import trimesh | |
from trimesh import Trimesh | |
from trimesh.triangles import points_to_barycentric | |
try: | |
# pyre-fixme[21]: Could not find module `igl`. | |
from igl import point_mesh_squared_distance # @manual | |
# pyre-fixme[3]: Return type must be annotated. | |
# pyre-fixme[2]: Parameter must be annotated. | |
def closest_point(mesh, points): | |
"""Helper function that mimics trimesh.proximity.closest_point but uses | |
IGL for faster queries.""" | |
v = mesh.vertices | |
vi = mesh.faces | |
dist, face_idxs, p = point_mesh_squared_distance(points, v, vi) | |
return p, dist, face_idxs | |
except ImportError: | |
from trimesh.proximity import closest_point | |
def closest_point_barycentrics(v, vi, points): | |
"""Given a 3D mesh and a set of query points, return closest point barycentrics | |
Args: | |
v: np.array (float) | |
[N, 3] mesh vertices | |
vi: np.array (int) | |
[N, 3] mesh triangle indices | |
points: np.array (float) | |
[M, 3] query points | |
Returns: | |
Tuple[approx, barys, interp_idxs, face_idxs] | |
approx: [M, 3] approximated (closest) points on the mesh | |
barys: [M, 3] barycentric weights that produce "approx" | |
interp_idxs: [M, 3] vertex indices for barycentric interpolation | |
face_idxs: [M] face indices for barycentric interpolation. interp_idxs = vi[face_idxs] | |
""" | |
mesh = Trimesh(vertices=v, faces=vi, process=False) | |
p, _, face_idxs = closest_point(mesh, points) | |
p = p.reshape((points.shape[0], 3)) | |
face_idxs = face_idxs.reshape((points.shape[0],)) | |
barys = points_to_barycentric(mesh.triangles[face_idxs], p) | |
b0, b1, b2 = np.split(barys, 3, axis=1) | |
interp_idxs = vi[face_idxs] | |
v0 = v[interp_idxs[:, 0]] | |
v1 = v[interp_idxs[:, 1]] | |
v2 = v[interp_idxs[:, 2]] | |
approx = b0 * v0 + b1 * v1 + b2 * v2 | |
return approx, barys, interp_idxs, face_idxs | |
def make_uv_face_index( | |
vt: th.Tensor, | |
vti: th.Tensor, | |
uv_shape: Union[Tuple[int, int], int], | |
flip_uv: bool = True, | |
device: Optional[Union[str, th.device]] = None, | |
): | |
"""Compute a UV-space face index map identifying which mesh face contains each | |
texel. For texels with no assigned triangle, the index will be -1.""" | |
if isinstance(uv_shape, int): | |
uv_shape = (uv_shape, uv_shape) | |
uv_max_shape_ind = uv_shape.index(max(uv_shape)) | |
uv_min_shape_ind = uv_shape.index(min(uv_shape)) | |
uv_ratio = uv_shape[uv_max_shape_ind] / uv_shape[uv_min_shape_ind] | |
if device is not None: | |
if isinstance(device, str): | |
dev = th.device(device) | |
else: | |
dev = device | |
assert dev.type == "cuda" | |
else: | |
dev = th.device("cuda") | |
vt = 1.0 - vt.clone() | |
if flip_uv: | |
vt = vt.clone() | |
vt[:, 1] = 1 - vt[:, 1] | |
vt_pix = 2.0 * vt.to(dev) - 1.0 | |
vt_pix = th.cat([vt_pix, th.ones_like(vt_pix[:, 0:1])], dim=1) | |
vt_pix[:, uv_min_shape_ind] *= uv_ratio | |
meshes = Meshes(vt_pix[np.newaxis], vti[np.newaxis].to(dev)) | |
with th.no_grad(): | |
face_index, _, _, _ = rasterize_meshes( | |
meshes, uv_shape, faces_per_pixel=1, z_clip_value=0.0, bin_size=0 | |
) | |
face_index = face_index[0, ..., 0] | |
return face_index | |
def make_uv_vert_index( | |
vt: th.Tensor, | |
vi: th.Tensor, | |
vti: th.Tensor, | |
uv_shape: Union[Tuple[int, int], int], | |
flip_uv: bool = True, | |
): | |
"""Compute a UV-space vertex index map identifying which mesh vertices | |
comprise the triangle containing each texel. For texels with no assigned | |
triangle, all indices will be -1. | |
""" | |
face_index_map = make_uv_face_index(vt, vti, uv_shape, flip_uv) | |
vert_index_map = vi[face_index_map.clamp(min=0)] | |
vert_index_map[face_index_map < 0] = -1 | |
return vert_index_map.long() | |
def bary_coords(points: th.Tensor, triangles: th.Tensor, eps: float = 1.0e-6): | |
"""Computes barycentric coordinates for a set of 2D query points given | |
coordintes for the 3 vertices of the enclosing triangle for each point.""" | |
x = points[:, 0] - triangles[2, :, 0] | |
x1 = triangles[0, :, 0] - triangles[2, :, 0] | |
x2 = triangles[1, :, 0] - triangles[2, :, 0] | |
y = points[:, 1] - triangles[2, :, 1] | |
y1 = triangles[0, :, 1] - triangles[2, :, 1] | |
y2 = triangles[1, :, 1] - triangles[2, :, 1] | |
denom = y2 * x1 - y1 * x2 | |
n0 = y2 * x - x2 * y | |
n1 = x1 * y - y1 * x | |
# Small epsilon to prevent divide-by-zero error. | |
denom = th.where(denom >= 0, denom.clamp(min=eps), denom.clamp(max=-eps)) | |
bary_0 = n0 / denom | |
bary_1 = n1 / denom | |
bary_2 = 1.0 - bary_0 - bary_1 | |
return th.stack((bary_0, bary_1, bary_2)) | |
def make_uv_barys( | |
vt: th.Tensor, | |
vti: th.Tensor, | |
uv_shape: Union[Tuple[int, int], int], | |
flip_uv: bool = True, | |
): | |
"""Compute a UV-space barycentric map where each texel contains barycentric | |
coordinates for that texel within its enclosing UV triangle. For texels | |
with no assigned triangle, all 3 barycentric coordinates will be 0. | |
""" | |
if isinstance(uv_shape, int): | |
uv_shape = (uv_shape, uv_shape) | |
if flip_uv: | |
# Flip here because texture coordinates in some of our topo files are | |
# stored in OpenGL convention with Y=0 on the bottom of the texture | |
# unlike numpy/torch arrays/tensors. | |
vt = vt.clone() | |
vt[:, 1] = 1 - vt[:, 1] | |
face_index_map = make_uv_face_index(vt, vti, uv_shape, flip_uv=False) | |
vti_map = vti.long()[face_index_map.clamp(min=0)] | |
uv_max_shape_ind = uv_shape.index(max(uv_shape)) | |
uv_min_shape_ind = uv_shape.index(min(uv_shape)) | |
uv_ratio = uv_shape[uv_max_shape_ind] / uv_shape[uv_min_shape_ind] | |
vt = vt.clone() | |
vt = vt * 2 - 1 | |
vt[:, uv_min_shape_ind] *= uv_ratio | |
uv_tri_uvs = vt[vti_map].permute(2, 0, 1, 3) | |
uv_grid = th.meshgrid( | |
th.linspace(0.5, uv_shape[0] - 0.5, uv_shape[0]) / uv_shape[0], | |
th.linspace(0.5, uv_shape[1] - 0.5, uv_shape[1]) / uv_shape[1], | |
) | |
uv_grid = th.stack(uv_grid[::-1], dim=2).to(uv_tri_uvs) | |
uv_grid = uv_grid * 2 - 1 | |
uv_grid[..., uv_min_shape_ind] *= uv_ratio | |
bary_map = bary_coords(uv_grid.view(-1, 2), uv_tri_uvs.view(3, -1, 2)) | |
bary_map = bary_map.permute(1, 0).view(uv_shape[0], uv_shape[1], 3) | |
bary_map[face_index_map < 0] = 0 | |
return face_index_map, bary_map | |
def index_image_impaint( | |
index_image: th.Tensor, | |
bary_image: Optional[th.Tensor] = None, | |
distance_threshold=100.0, | |
): | |
# getting the mask around the indexes? | |
if len(index_image.shape) == 3: | |
valid_index = (index_image != -1).any(dim=-1) | |
elif len(index_image.shape) == 2: | |
valid_index = index_image != -1 | |
else: | |
raise ValueError("`index_image` should be a [H,W] or [H,W,C] image") | |
invalid_index = ~valid_index | |
device = index_image.device | |
valid_ij = th.stack(th.where(valid_index), dim=-1) | |
invalid_ij = th.stack(th.where(invalid_index), dim=-1) | |
lookup_valid = KDTree(valid_ij.cpu().numpy()) | |
dists, idxs = lookup_valid.query(invalid_ij.cpu()) | |
# TODO: try average? | |
idxs = th.as_tensor(idxs, device=device)[..., 0] | |
dists = th.as_tensor(dists, device=device)[..., 0] | |
dist_mask = dists < distance_threshold | |
invalid_border = th.zeros_like(invalid_index) | |
invalid_border[invalid_index] = dist_mask | |
invalid_src_ij = valid_ij[idxs][dist_mask] | |
invalid_dst_ij = invalid_ij[dist_mask] | |
index_image_imp = index_image.clone() | |
index_image_imp[invalid_dst_ij[:, 0], invalid_dst_ij[:, 1]] = index_image[ | |
invalid_src_ij[:, 0], invalid_src_ij[:, 1] | |
] | |
if bary_image is not None: | |
bary_image_imp = bary_image.clone() | |
bary_image_imp[invalid_dst_ij[:, 0], invalid_dst_ij[:, 1]] = bary_image[ | |
invalid_src_ij[:, 0], invalid_src_ij[:, 1] | |
] | |
return index_image_imp, bary_image_imp | |
return index_image_imp | |
class GeometryModule(nn.Module): | |
def __init__( | |
self, | |
v, | |
vi, | |
vt, | |
vti, | |
uv_size, | |
v2uv: Optional[th.Tensor] = None, | |
flip_uv=False, | |
impaint=False, | |
impaint_threshold=100.0, | |
): | |
super().__init__() | |
self.register_buffer("v", th.as_tensor(v)) | |
self.register_buffer("vi", th.as_tensor(vi)) | |
self.register_buffer("vt", th.as_tensor(vt)) | |
self.register_buffer("vti", th.as_tensor(vti)) | |
if v2uv is not None: | |
self.register_buffer("v2uv", th.as_tensor(v2uv, dtype=th.int64)) | |
# TODO: should we just pass topology here? | |
# self.n_verts = v2uv.shape[0] | |
self.n_verts = vi.max() + 1 | |
self.uv_size = uv_size | |
# TODO: can't we just index face_index? | |
index_image = make_uv_vert_index( | |
self.vt, self.vi, self.vti, uv_shape=uv_size, flip_uv=flip_uv | |
).cpu() | |
face_index, bary_image = make_uv_barys( | |
self.vt, self.vti, uv_shape=uv_size, flip_uv=flip_uv | |
) | |
if impaint: | |
if min(uv_size) >= 1024: | |
logger.info( | |
"impainting index image might take a while for sizes >= 1024" | |
) | |
index_image, bary_image = index_image_impaint( | |
index_image, bary_image, impaint_threshold | |
) | |
# TODO: we can avoid doing this 2x | |
face_index = index_image_impaint( | |
face_index, distance_threshold=impaint_threshold | |
) | |
self.register_buffer("index_image", index_image.cpu()) | |
self.register_buffer("bary_image", bary_image.cpu()) | |
self.register_buffer("face_index_image", face_index.cpu()) | |
def render_index_images(self, uv_size, flip_uv=False, impaint=False): | |
index_image = make_uv_vert_index( | |
self.vt, self.vi, self.vti, uv_shape=uv_size, flip_uv=flip_uv | |
) | |
face_image, bary_image = make_uv_barys( | |
self.vt, self.vti, uv_shape=uv_size, flip_uv=flip_uv | |
) | |
if impaint: | |
index_image, bary_image = index_image_impaint( | |
index_image, | |
bary_image, | |
) | |
return index_image, face_image, bary_image | |
def vn(self, verts): | |
return vert_normals(verts, self.vi[np.newaxis].to(th.long)) | |
def to_uv(self, values): | |
return values_to_uv(values, self.index_image, self.bary_image) | |
def from_uv(self, values_uv): | |
# TODO: we need to sample this | |
return sample_uv(values_uv, self.vt, self.v2uv.to(th.long)) | |
def rand_sample_3d_uv(self, count, uv_img): | |
""" | |
Sample a set of 3D points on the surface of mesh, return corresponding interpolated values in UV space. | |
Args: | |
count - num of 3D points to be sampled | |
uv_img - the image in uv space to be sampled, e.g., texture | |
""" | |
_mesh = Trimesh(vertices=self.v.detach().cpu().numpy(), faces=self.vi.detach().cpu().numpy(), process=False) | |
points, _ = trimesh.sample.sample_surface(_mesh, count) | |
return self.sample_uv_from_3dpts(points, uv_img) | |
def sample_uv_from_3dpts(self, points, uv_img): | |
num_pts = points.shape[0] | |
approx, barys, interp_idxs, face_idxs = closest_point_barycentrics(self.v.detach().cpu().numpy(), self.vi.detach().cpu().numpy(), points) | |
interp_uv_coords = self.vt[interp_idxs, :] # [N, 3, 2] | |
# do bary interp first to get interp_uv_coord in high-reso uv space | |
target_uv_coords = th.sum(interp_uv_coords * th.from_numpy(barys)[..., None], dim=1).float() | |
# then directly sample from uv space | |
sampled_values = sample_uv(values_uv=uv_img.permute(2, 0, 1)[None, ...], uv_coords=target_uv_coords) # [1, count, c] | |
approx_values = sampled_values[0].reshape(num_pts, uv_img.shape[2]) | |
return approx_values.numpy(), points | |
def vert_sample_uv(self, uv_img): | |
count = self.v.shape[0] | |
points = self.v.detach().cpu().numpy() | |
approx_values, _ = self.sample_uv_from_3dpts(points, uv_img) | |
return approx_values | |
def sample_uv( | |
values_uv, | |
uv_coords, | |
v2uv: Optional[th.Tensor] = None, | |
mode: str = "bilinear", | |
align_corners: bool = True, | |
flip_uvs: bool = False, | |
): | |
batch_size = values_uv.shape[0] | |
if flip_uvs: | |
uv_coords = uv_coords.clone() | |
uv_coords[:, 1] = 1.0 - uv_coords[:, 1] | |
# uv_coords_norm is [1, N, 1, 2] afterwards | |
uv_coords_norm = (uv_coords * 2.0 - 1.0)[np.newaxis, :, np.newaxis].expand( | |
batch_size, -1, -1, -1 | |
) | |
# uv_shape = values_uv.shape[-2:] | |
# uv_max_shape_ind = uv_shape.index(max(uv_shape)) | |
# uv_min_shape_ind = uv_shape.index(min(uv_shape)) | |
# uv_ratio = uv_shape[uv_max_shape_ind] / uv_shape[uv_min_shape_ind] | |
# uv_coords_norm[..., uv_min_shape_ind] *= uv_ratio | |
values = ( | |
F.grid_sample(values_uv, uv_coords_norm, align_corners=align_corners, mode=mode) | |
.squeeze(-1) | |
.permute((0, 2, 1)) | |
) | |
if v2uv is not None: | |
values_duplicate = values[:, v2uv] | |
values = values_duplicate.mean(2) | |
return values | |
def values_to_uv(values, index_img, bary_img): | |
uv_size = index_img.shape | |
index_mask = th.all(index_img != -1, dim=-1) | |
idxs_flat = index_img[index_mask].to(th.int64) | |
bary_flat = bary_img[index_mask].to(th.float32) | |
# NOTE: here we assume | |
values_flat = th.sum(values[:, idxs_flat].permute(0, 3, 1, 2) * bary_flat, dim=-1) | |
values_uv = th.zeros( | |
values.shape[0], | |
values.shape[-1], | |
uv_size[0], | |
uv_size[1], | |
dtype=values.dtype, | |
device=values.device, | |
) | |
values_uv[:, :, index_mask] = values_flat | |
return values_uv | |
def face_normals(v, vi, eps: float = 1e-5): | |
pts = v[:, vi] | |
v0 = pts[:, :, 1] - pts[:, :, 0] | |
v1 = pts[:, :, 2] - pts[:, :, 0] | |
n = th.cross(v0, v1, dim=-1) | |
norm = th.norm(n, dim=-1, keepdim=True) | |
norm[norm < eps] = 1 | |
n /= norm | |
return n | |
def vert_normals(v, vi, eps: float = 1.0e-5): | |
fnorms = face_normals(v, vi) | |
fnorms = fnorms[:, :, None].expand(-1, -1, 3, -1).reshape(fnorms.shape[0], -1, 3) | |
vi_flat = vi.view(1, -1).expand(v.shape[0], -1) | |
vnorms = th.zeros_like(v) | |
for j in range(3): | |
vnorms[..., j].scatter_add_(1, vi_flat, fnorms[..., j]) | |
norm = th.norm(vnorms, dim=-1, keepdim=True) | |
norm[norm < eps] = 1 | |
vnorms /= norm | |
return vnorms | |
def compute_view_cos(verts, faces, camera_pos): | |
vn = F.normalize(vert_normals(verts, faces), dim=-1) | |
v2c = F.normalize(verts - camera_pos[:, np.newaxis], dim=-1) | |
return th.einsum("bnd,bnd->bn", vn, v2c) | |
def compute_tbn(geom, vt, vi, vti): | |
"""Computes tangent, bitangent, and normal vectors given a mesh. | |
Args: | |
geom: [N, n_verts, 3] th.Tensor | |
Vertex positions. | |
vt: [n_uv_coords, 2] th.Tensor | |
UV coordinates. | |
vi: [..., 3] th.Tensor | |
Face vertex indices. | |
vti: [..., 3] th.Tensor | |
Face UV indices. | |
Returns: | |
[..., 3] th.Tensors for T, B, N. | |
""" | |
v0 = geom[:, vi[..., 0]] | |
v1 = geom[:, vi[..., 1]] | |
v2 = geom[:, vi[..., 2]] | |
vt0 = vt[vti[..., 0]] | |
vt1 = vt[vti[..., 1]] | |
vt2 = vt[vti[..., 2]] | |
v01 = v1 - v0 | |
v02 = v2 - v0 | |
vt01 = vt1 - vt0 | |
vt02 = vt2 - vt0 | |
f = 1.0 / ( | |
vt01[None, ..., 0] * vt02[None, ..., 1] | |
- vt01[None, ..., 1] * vt02[None, ..., 0] | |
) | |
tangent = f[..., None] * th.stack( | |
[ | |
v01[..., 0] * vt02[None, ..., 1] - v02[..., 0] * vt01[None, ..., 1], | |
v01[..., 1] * vt02[None, ..., 1] - v02[..., 1] * vt01[None, ..., 1], | |
v01[..., 2] * vt02[None, ..., 1] - v02[..., 2] * vt01[None, ..., 1], | |
], | |
dim=-1, | |
) | |
tangent = F.normalize(tangent, dim=-1) | |
normal = F.normalize(th.cross(v01, v02, dim=3), dim=-1) | |
bitangent = F.normalize(th.cross(tangent, normal, dim=3), dim=-1) | |
return tangent, bitangent, normal | |
def compute_v2uv(n_verts, vi, vti, n_max=4): | |
"""Computes mapping from vertex indices to texture indices. | |
Args: | |
vi: [F, 3], triangles | |
vti: [F, 3], texture triangles | |
n_max: int, max number of texture locations | |
Returns: | |
[n_verts, n_max], texture indices | |
""" | |
v2uv_dict = {} | |
for i_v, i_uv in zip(vi.reshape(-1), vti.reshape(-1)): | |
v2uv_dict.setdefault(i_v, set()).add(i_uv) | |
assert len(v2uv_dict) == n_verts | |
v2uv = np.zeros((n_verts, n_max), dtype=np.int32) | |
for i in range(n_verts): | |
vals = sorted(list(v2uv_dict[i])) | |
v2uv[i, :] = vals[0] | |
v2uv[i, : len(vals)] = np.array(vals) | |
return v2uv | |
def compute_neighbours(n_verts, vi, n_max_values=10): | |
"""Computes first-ring neighbours given vertices and faces.""" | |
n_vi = vi.shape[0] | |
adj = {i: set() for i in range(n_verts)} | |
for i in range(n_vi): | |
for idx in vi[i]: | |
adj[idx] |= set(vi[i]) - set([idx]) | |
nbs_idxs = np.tile(np.arange(n_verts)[:, np.newaxis], (1, n_max_values)) | |
nbs_weights = np.zeros((n_verts, n_max_values), dtype=np.float32) | |
for idx in range(n_verts): | |
n_values = min(len(adj[idx]), n_max_values) | |
nbs_idxs[idx, :n_values] = np.array(list(adj[idx]))[:n_values] | |
nbs_weights[idx, :n_values] = -1.0 / n_values | |
return nbs_idxs, nbs_weights | |
def make_postex(v, idxim, barim): | |
return ( | |
barim[None, :, :, 0, None] * v[:, idxim[:, :, 0]] | |
+ barim[None, :, :, 1, None] * v[:, idxim[:, :, 1]] | |
+ barim[None, :, :, 2, None] * v[:, idxim[:, :, 2]] | |
).permute(0, 3, 1, 2) | |
def matrix_to_axisangle(r): | |
th = th.arccos(0.5 * (r[..., 0, 0] + r[..., 1, 1] + r[..., 2, 2] - 1.0))[..., None] | |
vec = ( | |
0.5 | |
* th.stack( | |
[ | |
r[..., 2, 1] - r[..., 1, 2], | |
r[..., 0, 2] - r[..., 2, 0], | |
r[..., 1, 0] - r[..., 0, 1], | |
], | |
dim=-1, | |
) | |
/ th.sin(th) | |
) | |
return th, vec | |
def axisangle_to_matrix(rvec): | |
theta = th.sqrt(1e-5 + th.sum(rvec**2, dim=-1)) | |
rvec = rvec / theta[..., None] | |
costh = th.cos(theta) | |
sinth = th.sin(theta) | |
return th.stack( | |
( | |
th.stack( | |
( | |
rvec[..., 0] ** 2 + (1.0 - rvec[..., 0] ** 2) * costh, | |
rvec[..., 0] * rvec[..., 1] * (1.0 - costh) - rvec[..., 2] * sinth, | |
rvec[..., 0] * rvec[..., 2] * (1.0 - costh) + rvec[..., 1] * sinth, | |
), | |
dim=-1, | |
), | |
th.stack( | |
( | |
rvec[..., 0] * rvec[..., 1] * (1.0 - costh) + rvec[..., 2] * sinth, | |
rvec[..., 1] ** 2 + (1.0 - rvec[..., 1] ** 2) * costh, | |
rvec[..., 1] * rvec[..., 2] * (1.0 - costh) - rvec[..., 0] * sinth, | |
), | |
dim=-1, | |
), | |
th.stack( | |
( | |
rvec[..., 0] * rvec[..., 2] * (1.0 - costh) - rvec[..., 1] * sinth, | |
rvec[..., 1] * rvec[..., 2] * (1.0 - costh) + rvec[..., 0] * sinth, | |
rvec[..., 2] ** 2 + (1.0 - rvec[..., 2] ** 2) * costh, | |
), | |
dim=-1, | |
), | |
), | |
dim=-2, | |
) | |
def rotation_interp(r0, r1, alpha): | |
r0a = r0.view(-1, 3, 3) | |
r1a = r1.view(-1, 3, 3) | |
r = th.bmm(r0a.permute(0, 2, 1), r1a).view_as(r0) | |
th, rvec = matrix_to_axisangle(r) | |
rvec = rvec * (alpha * th) | |
r = axisangle_to_matrix(rvec) | |
return th.bmm(r0a, r.view(-1, 3, 3)).view_as(r0) | |
def convert_camera_parameters(Rt, K): | |
R = Rt[:, :3, :3] | |
t = -R.permute(0, 2, 1).bmm(Rt[:, :3, 3].unsqueeze(2)).squeeze(2) | |
return dict( | |
campos=t, | |
camrot=R, | |
focal=K[:, :2, :2], | |
princpt=K[:, :2, 2], | |
) | |
def project_points_multi(p, Rt, K, normalize=False, size=None): | |
"""Project a set of 3D points into multiple cameras with a pinhole model. | |
Args: | |
p: [B, N, 3], input 3D points in world coordinates | |
Rt: [B, NC, 3, 4], extrinsics (where NC is the number of cameras to project to) | |
K: [B, NC, 3, 3], intrinsics | |
normalize: bool, whether to normalize coordinates to [-1.0, 1.0] | |
Returns: | |
tuple: | |
- [B, NC, N, 2] - projected points | |
- [B, NC, N] - their | |
""" | |
B, N = p.shape[:2] | |
NC = Rt.shape[1] | |
Rt = Rt.reshape(B * NC, 3, 4) | |
K = K.reshape(B * NC, 3, 3) | |
# [B, N, 3] -> [B * NC, N, 3] | |
p = p[:, np.newaxis].expand(-1, NC, -1, -1).reshape(B * NC, -1, 3) | |
p_cam = p @ Rt[:, :3, :3].transpose(-2, -1) + Rt[:, :3, 3][:, np.newaxis] | |
p_pix = p_cam @ K.transpose(-2, -1) | |
p_depth = p_pix[:, :, 2:] | |
p_pix = (p_pix[..., :2] / p_depth).reshape(B, NC, N, 2) | |
p_depth = p_depth.reshape(B, NC, N) | |
if normalize: | |
assert size is not None | |
h, w = size | |
p_pix = ( | |
2.0 * p_pix / th.as_tensor([w, h], dtype=th.float32, device=p.device) - 1.0 | |
) | |
return p_pix, p_depth | |