|
|
|
import torch |
|
|
|
|
|
def fp16_clamp(x, min=None, max=None): |
|
if not x.is_cuda and x.dtype == torch.float16: |
|
|
|
return x.float().clamp(min, max).half() |
|
|
|
return x.clamp(min, max) |
|
|
|
|
|
def bbox_overlaps(bboxes1, bboxes2, mode='iou', is_aligned=False, eps=1e-6): |
|
"""Calculate overlap between two set of bboxes. |
|
|
|
FP16 Contributed by https://github.com/open-mmlab/mmdetection/pull/4889 |
|
Note: |
|
Assume bboxes1 is M x 4, bboxes2 is N x 4, when mode is 'iou', |
|
there are some new generated variable when calculating IOU |
|
using bbox_overlaps function: |
|
|
|
1) is_aligned is False |
|
area1: M x 1 |
|
area2: N x 1 |
|
lt: M x N x 2 |
|
rb: M x N x 2 |
|
wh: M x N x 2 |
|
overlap: M x N x 1 |
|
union: M x N x 1 |
|
ious: M x N x 1 |
|
|
|
Total memory: |
|
S = (9 x N x M + N + M) * 4 Byte, |
|
|
|
When using FP16, we can reduce: |
|
R = (9 x N x M + N + M) * 4 / 2 Byte |
|
R large than (N + M) * 4 * 2 is always true when N and M >= 1. |
|
Obviously, N + M <= N * M < 3 * N * M, when N >=2 and M >=2, |
|
N + 1 < 3 * N, when N or M is 1. |
|
|
|
Given M = 40 (ground truth), N = 400000 (three anchor boxes |
|
in per grid, FPN, R-CNNs), |
|
R = 275 MB (one times) |
|
|
|
A special case (dense detection), M = 512 (ground truth), |
|
R = 3516 MB = 3.43 GB |
|
|
|
When the batch size is B, reduce: |
|
B x R |
|
|
|
Therefore, CUDA memory runs out frequently. |
|
|
|
Experiments on GeForce RTX 2080Ti (11019 MiB): |
|
|
|
| dtype | M | N | Use | Real | Ideal | |
|
|:----:|:----:|:----:|:----:|:----:|:----:| |
|
| FP32 | 512 | 400000 | 8020 MiB | -- | -- | |
|
| FP16 | 512 | 400000 | 4504 MiB | 3516 MiB | 3516 MiB | |
|
| FP32 | 40 | 400000 | 1540 MiB | -- | -- | |
|
| FP16 | 40 | 400000 | 1264 MiB | 276MiB | 275 MiB | |
|
|
|
2) is_aligned is True |
|
area1: N x 1 |
|
area2: N x 1 |
|
lt: N x 2 |
|
rb: N x 2 |
|
wh: N x 2 |
|
overlap: N x 1 |
|
union: N x 1 |
|
ious: N x 1 |
|
|
|
Total memory: |
|
S = 11 x N * 4 Byte |
|
|
|
When using FP16, we can reduce: |
|
R = 11 x N * 4 / 2 Byte |
|
|
|
So do the 'giou' (large than 'iou'). |
|
|
|
Time-wise, FP16 is generally faster than FP32. |
|
|
|
When gpu_assign_thr is not -1, it takes more time on cpu |
|
but not reduce memory. |
|
There, we can reduce half the memory and keep the speed. |
|
|
|
If ``is_aligned`` is ``False``, then calculate the overlaps between each |
|
bbox of bboxes1 and bboxes2, otherwise the overlaps between each aligned |
|
pair of bboxes1 and bboxes2. |
|
|
|
Args: |
|
bboxes1 (Tensor): shape (B, m, 4) in <x1, y1, x2, y2> format or empty. |
|
bboxes2 (Tensor): shape (B, n, 4) in <x1, y1, x2, y2> format or empty. |
|
B indicates the batch dim, in shape (B1, B2, ..., Bn). |
|
If ``is_aligned`` is ``True``, then m and n must be equal. |
|
mode (str): "iou" (intersection over union), "iof" (intersection over |
|
foreground) or "giou" (generalized intersection over union). |
|
Default "iou". |
|
is_aligned (bool, optional): If True, then m and n must be equal. |
|
Default False. |
|
eps (float, optional): A value added to the denominator for numerical |
|
stability. Default 1e-6. |
|
|
|
Returns: |
|
Tensor: shape (m, n) if ``is_aligned`` is False else shape (m,) |
|
|
|
Example: |
|
>>> bboxes1 = torch.FloatTensor([ |
|
>>> [0, 0, 10, 10], |
|
>>> [10, 10, 20, 20], |
|
>>> [32, 32, 38, 42], |
|
>>> ]) |
|
>>> bboxes2 = torch.FloatTensor([ |
|
>>> [0, 0, 10, 20], |
|
>>> [0, 10, 10, 19], |
|
>>> [10, 10, 20, 20], |
|
>>> ]) |
|
>>> overlaps = bbox_overlaps(bboxes1, bboxes2) |
|
>>> assert overlaps.shape == (3, 3) |
|
>>> overlaps = bbox_overlaps(bboxes1, bboxes2, is_aligned=True) |
|
>>> assert overlaps.shape == (3, ) |
|
|
|
Example: |
|
>>> empty = torch.empty(0, 4) |
|
>>> nonempty = torch.FloatTensor([[0, 0, 10, 9]]) |
|
>>> assert tuple(bbox_overlaps(empty, nonempty).shape) == (0, 1) |
|
>>> assert tuple(bbox_overlaps(nonempty, empty).shape) == (1, 0) |
|
>>> assert tuple(bbox_overlaps(empty, empty).shape) == (0, 0) |
|
""" |
|
|
|
assert mode in ['iou', 'iof', 'giou'], f'Unsupported mode {mode}' |
|
|
|
assert (bboxes1.size(-1) == 4 or bboxes1.size(0) == 0) |
|
assert (bboxes2.size(-1) == 4 or bboxes2.size(0) == 0) |
|
|
|
|
|
|
|
assert bboxes1.shape[:-2] == bboxes2.shape[:-2] |
|
batch_shape = bboxes1.shape[:-2] |
|
|
|
rows = bboxes1.size(-2) |
|
cols = bboxes2.size(-2) |
|
if is_aligned: |
|
assert rows == cols |
|
|
|
if rows * cols == 0: |
|
if is_aligned: |
|
return bboxes1.new(batch_shape + (rows, )) |
|
else: |
|
return bboxes1.new(batch_shape + (rows, cols)) |
|
|
|
area1 = (bboxes1[..., 2] - bboxes1[..., 0]) * ( |
|
bboxes1[..., 3] - bboxes1[..., 1]) |
|
area2 = (bboxes2[..., 2] - bboxes2[..., 0]) * ( |
|
bboxes2[..., 3] - bboxes2[..., 1]) |
|
|
|
if is_aligned: |
|
lt = torch.max(bboxes1[..., :2], bboxes2[..., :2]) |
|
rb = torch.min(bboxes1[..., 2:], bboxes2[..., 2:]) |
|
|
|
wh = fp16_clamp(rb - lt, min=0) |
|
overlap = wh[..., 0] * wh[..., 1] |
|
|
|
if mode in ['iou', 'giou']: |
|
union = area1 + area2 - overlap |
|
else: |
|
union = area1 |
|
if mode == 'giou': |
|
enclosed_lt = torch.min(bboxes1[..., :2], bboxes2[..., :2]) |
|
enclosed_rb = torch.max(bboxes1[..., 2:], bboxes2[..., 2:]) |
|
else: |
|
lt = torch.max(bboxes1[..., :, None, :2], |
|
bboxes2[..., None, :, :2]) |
|
rb = torch.min(bboxes1[..., :, None, 2:], |
|
bboxes2[..., None, :, 2:]) |
|
|
|
wh = fp16_clamp(rb - lt, min=0) |
|
overlap = wh[..., 0] * wh[..., 1] |
|
|
|
if mode in ['iou', 'giou']: |
|
union = area1[..., None] + area2[..., None, :] - overlap |
|
else: |
|
union = area1[..., None] |
|
if mode == 'giou': |
|
enclosed_lt = torch.min(bboxes1[..., :, None, :2], |
|
bboxes2[..., None, :, :2]) |
|
enclosed_rb = torch.max(bboxes1[..., :, None, 2:], |
|
bboxes2[..., None, :, 2:]) |
|
|
|
eps = union.new_tensor([eps]) |
|
union = torch.max(union, eps) |
|
ious = overlap / union |
|
if mode in ['iou', 'iof']: |
|
return ious |
|
|
|
enclose_wh = fp16_clamp(enclosed_rb - enclosed_lt, min=0) |
|
enclose_area = enclose_wh[..., 0] * enclose_wh[..., 1] |
|
enclose_area = torch.max(enclose_area, eps) |
|
gious = ious - (enclose_area - union) / enclose_area |
|
return gious |
|
|