scispace - formally typeset
Open AccessJournal ArticleDOI

INet: Convolutional Networks for Biomedical Image Segmentation

Weihao Weng, +1 more
- 21 Jan 2021 - 
- Vol. 9, pp 16591-16603
TLDR
In this article, the authors propose to enlarge receptive fields by increasing the kernel sizes of convolutional layers in steps (e.g., from $3\times 3$ to $7\times 7$ and then $15\times 15$ ) instead of downsampling.
Abstract
Encoder–decoder networks are state-of-the-art approaches to biomedical image segmentation, but have two problems: i.e., the widely used pooling operations may discard spatial information, and therefore low-level semantics are lost. Feature fusion methods can mitigate these problems but feature maps of different scales cannot be easily fused because down- and upsampling change the spatial resolution of feature map. To address these issues, we propose INet, which enlarges receptive fields by increasing the kernel sizes of convolutional layers in steps (e.g., from $3\times 3$ to $7\times 7$ and then $15\times 15$ ) instead of downsampling. Inspired by an Inception module, INet extracts features by kernels of different sizes through concatenating the output feature maps of all preceding convolutional layers. We also find that the large kernel makes the network feasible for biomedical image segmentation. In addition, INet uses two overlapping max-poolings, i.e., max-poolings with stride 1, to extract the sharpest features. Fixed-size and fixed-channel feature maps enable INet to concatenate feature maps and add multiple shortcuts across layers. In this way, INet can recover low-level semantics by concatenating the feature maps of all preceding layers and expedite the training by adding multiple shortcuts. Because INet has additional residual shortcuts, we compare INet with a UNet system that also has residual shortcuts (ResUNet). To confirm INet as a backbone architecture for biomedical image segmentation, we implement dense connections on INet (called DenseINet) and compare it to a DenseUNet system with residual shortcuts (ResDenseUNet). INet and DenseINet require 16.9% and 37.6% fewer parameters than ResUNet and ResDenseUNet, respectively. In comparison with six encoder–decoder approaches using nine public datasets, INet and DenseINet demonstrate efficient improvements in biomedical image segmentation. INet outperforms DeepLabV3, which implementing atrous convolution instead of downsampling to increase receptive fields. INet also outperforms two recent methods (named HRNet and MS-NAS) that maintain high-resolution representations and repeatedly exchange the information across resolutions.

read more

Citations
More filters
Journal ArticleDOI

Deep-learning-based visual data analytics for smart construction management

TL;DR: This in-depth review of state-of-the-art deep-learning applications on visual data analytics in the context of construction project management identifies six major fields and fifty-two subfields of construction management where deep- learning-based visual data Analytics methods have been applied and proposes a generalized workflow for applying these methods.
Journal ArticleDOI

Automatic polyp detection and segmentation using shuffle efficient channel attention network

TL;DR: In this paper, a new shuffle efficient channel attention network (sECANet) was proposed for colorectal polyp detection and segmentation, which achieved a precision of 94.9%, a recall of 96.9, and a F1 score of 95.9%.
Journal ArticleDOI

Deep Learning for automated phase segmentation in EBSD maps. A case study in Dual Phase steel microstructures

TL;DR: In this paper , U-Net has been used for phase transformation detection in steel microstructures, achieving Pixel-wise accuracies around 95% when inputting raw orientation data, while reaching 98% with orientation-derived parameters such as Kernel Average Misorientation (KAM) or pattern quality.
Journal ArticleDOI

Automatic polyp detection and segmentation using shuffle efficient channel attention network

TL;DR: In this paper , a new shuffle efficient channel attention network (sECANet) was proposed for colorectal polyp detection and segmentation, which achieved a precision of 94.9%, a recall of 96.9, and a F1 score of 95.9%.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Book ChapterDOI

U-Net: Convolutional Networks for Biomedical Image Segmentation

TL;DR: Neber et al. as discussed by the authors proposed a network and training strategy that relies on the strong use of data augmentation to use the available annotated samples more efficiently, which can be trained end-to-end from very few images and outperforms the prior best method (a sliding-window convolutional network) on the ISBI challenge for segmentation of neuronal structures in electron microscopic stacks.
Journal ArticleDOI

Deep learning

TL;DR: Deep learning is making major advances in solving problems that have resisted the best attempts of the artificial intelligence community for many years, and will have many more successes in the near future because it requires very little engineering by hand and can easily take advantage of increases in the amount of available computation and data.
Related Papers (5)