scispace - formally typeset
Open AccessProceedings ArticleDOI

IntraQ: Learning Synthetic Images with Intra-Class Heterogeneity for Zero-Shot Network Quantization

Reads0
Chats0
TLDR
Zhang et al. as mentioned in this paper proposed a local object reinforcement that locates the target objects at different scales and positions of the synthetic images and introduced a marginal distance constraint to form class-related features distributed in a coarse area.
Abstract
Learning to synthesize data has emerged as a promising direction in zero-shot quantization (ZSQ), which represents neural networks by low-bit integer without accessing any of the real data. In this paper, we observe an interesting phenomenon of intra-class heterogeneity in real data and show that existing methods fail to retain this property in their synthetic images, which causes a limited performance increase. To address this issue, we propose a novel zero-shot quantization method referred to as IntraQ. First, we propose a local object reinforcement that locates the target objects at different scales and positions of the synthetic images. Second, we introduce a marginal distance constraint to form class-related features distributed in a coarse area. Lastly, we devise a soft inception loss which injects a soft prior label to prevent the synthetic images from being over-fitting to a fixed object. Our IntraQ is demonstrated to well retain the intra-class heterogeneity in the synthetic images and also observed to perform state-of-the-art. For example, compared to the advanced ZSQ, our IntraQ obtains 9.17% increase of the top-1 accuracy on ImageNet when all layers of MobileNetV1 are quantized to 4-bit. Code is at https://github.com/zysxmu/IntraQ

read more

Content maybe subject to copyright    Report

Citations
More filters
Book ChapterDOI

Patch Similarity Aware Data-Free Quantization for Vision Transformers

TL;DR: PSAQ-ViT as mentioned in this paper is a patch similarity aware data-free quantization framework for vision transformers to enable the generation of "realistic" samples based on the vision transformer's unique properties for calibrating the quantization parameters.
Journal ArticleDOI

Fine-grained Data Distribution Alignment for Post-Training Quantization

TL;DR: Zhang et al. as mentioned in this paper proposed a fine-grained data distribution alignment (FDDA) method to boost the performance of post-training quantization, which is based on two important properties of batch normalization statistics (BNS) observed in deep layers of the trained network.
Journal ArticleDOI

Gradient distribution-aware INT8 training for neural networks

Shuai Wang
- 01 Apr 2023 - 
TL;DR: In this paper , the authors proposed two innovative techniques mainly for INT8 quantization training, including the Data-aware Dynamic Segmentation Quantization scheme to quantize various special gradient distributions and the Update Direction Periodic Search strategy to achieve lower quantization errors.
Proceedings ArticleDOI

A two-level architecture for deep learning applications in mobile edge computing

TL;DR: This research proposes a novel two-level inference architecture for deep learning applications in mobile edge computing that uses small models to do two levels of inferences on edge devices and edge servers respectively.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Journal ArticleDOI

ImageNet Large Scale Visual Recognition Challenge

TL;DR: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) as mentioned in this paper is a benchmark in object category classification and detection on hundreds of object categories and millions of images, which has been run annually from 2010 to present, attracting participation from more than fifty institutions.
Proceedings ArticleDOI

MobileNetV2: Inverted Residuals and Linear Bottlenecks

TL;DR: MobileNetV2 as mentioned in this paper is based on an inverted residual structure where the shortcut connections are between the thin bottleneck layers and intermediate expansion layer uses lightweight depthwise convolutions to filter features as a source of non-linearity.
Proceedings ArticleDOI

Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration

TL;DR: He et al. as discussed by the authors proposed a filter pruning via geometric median (FPGM) method to compress CNN models by pruning filters with redundancy, rather than those with relatively less importance.
Proceedings ArticleDOI

HRank: Filter Pruning Using High-Rank Feature Map

TL;DR: This paper proposes a novel filter pruning method by exploring the High Rank of feature maps (HRank), inspired by the discovery that the average rank of multiple feature maps generated by a single filter is always the same, regardless of the number of image batches CNNs receive.
Related Papers (5)
Trending Questions (1)
How does class heterogeneity affect the performance of deep learning models?

The provided paper does not directly discuss the impact of class heterogeneity on the performance of deep learning models.