utils.py 6.96 KB
Newer Older
1
2
import numpy as np
import torch
3
from logging import warning
4
5
6


def limit_period(val, offset=0.5, period=np.pi):
zhangwenwei's avatar
zhangwenwei committed
7
8
9
    """Limit the value into a period for periodic function.

    Args:
liyinhao's avatar
liyinhao committed
10
        val (torch.Tensor): The value to be converted.
wangtai's avatar
wangtai committed
11
        offset (float, optional): Offset to set the value range. \
zhangwenwei's avatar
zhangwenwei committed
12
13
14
15
            Defaults to 0.5.
        period ([type], optional): Period of the value. Defaults to np.pi.

    Returns:
wangtai's avatar
wangtai committed
16
        torch.Tensor: Value in the range of \
zhangwenwei's avatar
zhangwenwei committed
17
18
            [-offset * period, (1-offset) * period]
    """
19
20
21
22
    return val - torch.floor(val / period + offset) * period


def rotation_3d_in_axis(points, angles, axis=0):
zhangwenwei's avatar
zhangwenwei committed
23
    """Rotate points by angles according to axis.
zhangwenwei's avatar
zhangwenwei committed
24
25
26
27
28
29
30

    Args:
        points (torch.Tensor): Points of shape (N, M, 3).
        angles (torch.Tensor): Vector of angles in shape (N,)
        axis (int, optional): The axis to be rotated. Defaults to 0.

    Raises:
zhangwenwei's avatar
zhangwenwei committed
31
        ValueError: when the axis is not in range [0, 1, 2], it will \
zhangwenwei's avatar
zhangwenwei committed
32
33
34
            raise value error.

    Returns:
zhangwenwei's avatar
zhangwenwei committed
35
        torch.Tensor: Rotated points in shape (N, M, 3)
zhangwenwei's avatar
zhangwenwei committed
36
    """
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
    rot_sin = torch.sin(angles)
    rot_cos = torch.cos(angles)
    ones = torch.ones_like(rot_cos)
    zeros = torch.zeros_like(rot_cos)
    if axis == 1:
        rot_mat_T = torch.stack([
            torch.stack([rot_cos, zeros, -rot_sin]),
            torch.stack([zeros, ones, zeros]),
            torch.stack([rot_sin, zeros, rot_cos])
        ])
    elif axis == 2 or axis == -1:
        rot_mat_T = torch.stack([
            torch.stack([rot_cos, -rot_sin, zeros]),
            torch.stack([rot_sin, rot_cos, zeros]),
            torch.stack([zeros, zeros, ones])
        ])
    elif axis == 0:
        rot_mat_T = torch.stack([
            torch.stack([zeros, rot_cos, -rot_sin]),
            torch.stack([zeros, rot_sin, rot_cos]),
            torch.stack([ones, zeros, zeros])
        ])
    else:
zhangwenwei's avatar
zhangwenwei committed
60
        raise ValueError(f'axis should in range [0, 1, 2], got {axis}')
61
62

    return torch.einsum('aij,jka->aik', (points, rot_mat_T))
63
64
65


def xywhr2xyxyr(boxes_xywhr):
zhangwenwei's avatar
zhangwenwei committed
66
67
68
69
70
71
72
73
    """Convert a rotated boxes in XYWHR format to XYXYR format.

    Args:
        boxes_xywhr (torch.Tensor): Rotated boxes in XYWHR format.

    Returns:
        torch.Tensor: Converted boxes in XYXYR format.
    """
74
75
76
77
78
79
80
81
82
83
    boxes = torch.zeros_like(boxes_xywhr)
    half_w = boxes_xywhr[:, 2] / 2
    half_h = boxes_xywhr[:, 3] / 2

    boxes[:, 0] = boxes_xywhr[:, 0] - half_w
    boxes[:, 1] = boxes_xywhr[:, 1] - half_h
    boxes[:, 2] = boxes_xywhr[:, 0] + half_w
    boxes[:, 3] = boxes_xywhr[:, 1] + half_h
    boxes[:, 4] = boxes_xywhr[:, 4]
    return boxes
zhangwenwei's avatar
zhangwenwei committed
84
85


wuyuefeng's avatar
Demo  
wuyuefeng committed
86
87
88
89
def get_box_type(box_type):
    """Get the type and mode of box structure.

    Args:
90
        box_type (str): The type of box structure.
wuyuefeng's avatar
Demo  
wuyuefeng committed
91
92
93
            The valid value are "LiDAR", "Camera", or "Depth".

    Returns:
wangtai's avatar
wangtai committed
94
        tuple: Box type and box mode.
wuyuefeng's avatar
Demo  
wuyuefeng committed
95
    """
zhangwenwei's avatar
zhangwenwei committed
96
97
    from .box_3d_mode import (Box3DMode, CameraInstance3DBoxes,
                              DepthInstance3DBoxes, LiDARInstance3DBoxes)
wuyuefeng's avatar
Demo  
wuyuefeng committed
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
    box_type_lower = box_type.lower()
    if box_type_lower == 'lidar':
        box_type_3d = LiDARInstance3DBoxes
        box_mode_3d = Box3DMode.LIDAR
    elif box_type_lower == 'camera':
        box_type_3d = CameraInstance3DBoxes
        box_mode_3d = Box3DMode.CAM
    elif box_type_lower == 'depth':
        box_type_3d = DepthInstance3DBoxes
        box_mode_3d = Box3DMode.DEPTH
    else:
        raise ValueError('Only "box_type" of "camera", "lidar", "depth"'
                         f' are supported, got {box_type}')

    return box_type_3d, box_mode_3d


115
def points_cam2img(points_3d, proj_mat, with_depth=False):
zhangwenwei's avatar
zhangwenwei committed
116
    """Project points from camera coordicates to image coordinates.
zhangwenwei's avatar
zhangwenwei committed
117
118

    Args:
119
        points_3d (torch.Tensor): Points in shape (N, 3).
zhangwenwei's avatar
zhangwenwei committed
120
        proj_mat (torch.Tensor): Transformation matrix between coordinates.
121
122
        with_depth (bool, optional): Whether to keep depth in the output.
            Defaults to False.
zhangwenwei's avatar
zhangwenwei committed
123
124
125
126

    Returns:
        torch.Tensor: Points in image coordinates with shape [N, 2].
    """
zhangwenwei's avatar
zhangwenwei committed
127
    points_num = list(points_3d.shape)[:-1]
128

zhangwenwei's avatar
zhangwenwei committed
129
    points_shape = np.concatenate([points_num, [1]], axis=0).tolist()
130
131
    assert len(proj_mat.shape) == 2, 'The dimension of the projection'\
        f' matrix should be 2 instead of {len(proj_mat.shape)}.'
132
133
    d1, d2 = proj_mat.shape[:2]
    assert (d1 == 3 and d2 == 3) or (d1 == 3 and d2 == 4) or (
134
        d1 == 4 and d2 == 4), 'The shape of the projection matrix'\
135
136
137
138
139
140
141
        f' ({d1}*{d2}) is not supported.'
    if d1 == 3:
        proj_mat_expanded = torch.eye(
            4, device=proj_mat.device, dtype=proj_mat.dtype)
        proj_mat_expanded[:d1, :d2] = proj_mat
        proj_mat = proj_mat_expanded

zhangwenwei's avatar
zhangwenwei committed
142
143
144
145
146
    # previous implementation use new_zeros, new_one yeilds better results
    points_4 = torch.cat(
        [points_3d, points_3d.new_ones(*points_shape)], dim=-1)
    point_2d = torch.matmul(points_4, proj_mat.t())
    point_2d_res = point_2d[..., :2] / point_2d[..., 2:3]
147
148
149

    if with_depth:
        return torch.cat([point_2d_res, point_2d[..., 2:3]], dim=-1)
zhangwenwei's avatar
zhangwenwei committed
150
    return point_2d_res
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170


def mono_cam_box2vis(cam_box):
    """This is a post-processing function on the bboxes from Mono-3D task. If
    we want to perform projection visualization, we need to:

        1. rotate the box along x-axis for np.pi / 2 (roll)
        2. change orientation from local yaw to global yaw
        3. convert yaw by (np.pi / 2 - yaw)

    After applying this function, we can project and draw it on 2D images.

    Args:
        cam_box (:obj:`CameraInstance3DBoxes`): 3D bbox in camera coordinate \
            system before conversion. Could be gt bbox loaded from dataset or \
                network prediction output.

    Returns:
        :obj:`CameraInstance3DBoxes`: Box after conversion.
    """
171
172
173
    warning.warn('DeprecationWarning: The hack of yaw and dimension in the '
                 'monocular 3D detection on nuScenes has been removed. The '
                 'function mono_cam_box2vis will be deprecated.')
174
175
176
177
178
179
180
181
182
183
184
185
186
187
    from . import CameraInstance3DBoxes
    assert isinstance(cam_box, CameraInstance3DBoxes), \
        'input bbox should be CameraInstance3DBoxes!'

    loc = cam_box.gravity_center
    dim = cam_box.dims
    yaw = cam_box.yaw
    feats = cam_box.tensor[:, 7:]
    # rotate along x-axis for np.pi / 2
    # see also here: https://github.com/open-mmlab/mmdetection3d/blob/master/mmdet3d/datasets/nuscenes_mono_dataset.py#L557  # noqa
    dim[:, [1, 2]] = dim[:, [2, 1]]
    # change local yaw to global yaw for visualization
    # refer to https://github.com/open-mmlab/mmdetection3d/blob/master/mmdet3d/datasets/nuscenes_mono_dataset.py#L164-L166  # noqa
    yaw += torch.atan2(loc[:, 0], loc[:, 2])
Ziyi Wu's avatar
Ziyi Wu committed
188
189
190
191
    # convert yaw by (-yaw - np.pi / 2)
    # this is because mono 3D box class such as `NuScenesBox` has different
    # definition of rotation with our `CameraInstance3DBoxes`
    yaw = -yaw - np.pi / 2
192
193
194
195
196
    cam_box = torch.cat([loc, dim, yaw[:, None], feats], dim=1)
    cam_box = CameraInstance3DBoxes(
        cam_box, box_dim=cam_box.shape[-1], origin=(0.5, 0.5, 0.5))

    return cam_box