scispace - formally typeset
Open AccessJournal ArticleDOI

Oriented Bounding Boxes for Small and Freely Rotated Objects

Reads0
Chats0
TLDR
A novel object detection method is presented that handles freely rotated objects of arbitrary sizes, including tiny objects as small as $2 \times 2$ pixels, which has the added benefit of enabling oriented bounding box detection without any extra computation.
Abstract
A novel object detection method is presented that handles freely rotated objects of arbitrary sizes, including tiny objects as small as 2 x 2 pixels. Such tiny objects appear frequently in remotely sensed images, and present a challenge to recent object detection algorithms. More importantly, current object detection methods have been designed originally to accommodate axis-aligned bounding box detection, and therefore fail to accurately localize oriented boxes that best describe freely rotated objects. In contrast, the proposed convolutional neural network (CNN) -based approach uses potential pixel information at multiple scale levels without the need for any external resources, such as anchor boxes. The method encodes the precise location and orientation of features of the target objects at grid cell locations. Unlike existing methods that regress the bounding box location and dimension, the proposed method learns all the required information by classification, which has the added benefit of enabling oriented bounding box detection without any extra computation. It thus infers the bounding boxes only at inference time by finding the minimum surrounding box for every set of the same predicted class labels. Moreover, a rotation-invariant feature representation is applied to each scale, which imposes a regularization constraint to enforce covering the 360° range of in-plane rotation of the training samples to share similar features. Evaluations on the xView and dataset for object detection in aerial images (DOTA) data sets show that the proposed method uniformly improves performance over existing state-of-the-art methods.

read more

Citations
More filters
Proceedings Article

The KFIoU Loss for Rotated Object Detection

TL;DR: It is argued that one effective alternative is to devise an approximate loss who can achieve trend-level alignment with SkewIoU loss instead of the strict value-level identity and the resulting new loss called KFIoU is easier to implement and works better compared with exact Skew IoU, thanks to its full differentiability and ability to handle the non-overlapping cases.
Journal ArticleDOI

Towards Large-Scale Small Object Detection: Survey and Benchmarks

TL;DR: Two large-scale Small Object Detection dAtasets (SODA), SODA-D and S ODA-A, which focus on the Driving and Aerial scenarios respectively are constructed, and the performance of mainstream methods on SOD a is evaluated.
Journal ArticleDOI

A Guide to Image and Video based Small Object Detection using Deep Learning : Case Study of Maritime Surveillance

TL;DR: This paper investigates how to improve the performance of small object detection in maritime environments, where increasing performance is critical and establishes a connection between generic and maritime SOD research, future directions have been identified.
Journal ArticleDOI

Deep learning-based segmental analysis of fish for biomass estimation in an occulted environment

TL;DR: In this article , a deep learning-based segmental analysis technique was used to detect the fish head, body, and tail segments, and the detected segments were associated using sequence constrained nearest neighborhood (NN) association technique guided with fish head orientation.
Journal ArticleDOI

Rotation-Invariant Feature Learning via Convolutional Neural Network With Cyclic Polar Coordinates Convolutional Layer

TL;DR: In this paper , a cyclic polar coordinate convolutional layer (CPCCL) is proposed for CNNs to handle the problem of rotation invariance for feature learning, which can easily be handled by CNNs.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Book ChapterDOI

U-Net: Convolutional Networks for Biomedical Image Segmentation

TL;DR: Neber et al. as discussed by the authors proposed a network and training strategy that relies on the strong use of data augmentation to use the available annotated samples more efficiently, which can be trained end-to-end from very few images and outperforms the prior best method (a sliding-window convolutional network) on the ISBI challenge for segmentation of neuronal structures in electron microscopic stacks.
Proceedings ArticleDOI

Going deeper with convolutions

TL;DR: Inception as mentioned in this paper is a deep convolutional neural network architecture that achieves the new state of the art for classification and detection in the ImageNet Large-Scale Visual Recognition Challenge 2014 (ILSVRC14).
Proceedings ArticleDOI

You Only Look Once: Unified, Real-Time Object Detection

TL;DR: Compared to state-of-the-art detection systems, YOLO makes more localization errors but is less likely to predict false positives on background, and outperforms other detection methods, including DPM and R-CNN, when generalizing from natural images to other domains like artwork.
Related Papers (5)