scispace - formally typeset
L

Lei Wang

Researcher at University of Science and Technology Beijing

Publications -  8
Citations -  298

Lei Wang is an academic researcher from University of Science and Technology Beijing. The author has contributed to research in topics: Computer science & Complex network. The author has an hindex of 4, co-authored 6 publications receiving 32 citations.

Papers
More filters
Journal ArticleDOI

Pruning and quantization for deep neural network acceleration: A survey

TL;DR: A survey on two types of network compression: pruning and quantization is provided, which compare current techniques, analyze their strengths and weaknesses, provide guidance for compressing networks, and discuss possible future compression techniques.
Posted Content

Pruning and Quantization for Deep Neural Network Acceleration: A Survey

TL;DR: In this article, the authors provide a survey on two types of network compression: pruning and quantization, and compare current techniques, analyze their strengths and weaknesses, present compressed network accuracy results on a number of frameworks, and provide practical guidance for compressing networks.
Posted Content

Dynamic Runtime Feature Map Pruning

TL;DR: This work analyzes parameter sparsity of six popular convolutional neural networks and introduces dynamic runtime pruning of feature maps, showing that 10% of dynamic feature map execution can be removed without loss of accuracy.
Patent

Variable translation-lookaside buffer (TLB) indexing

TL;DR: A processor includes a translation lookaside buffer (TLB) comprising a plurality of ways, wherein each way is associated with a respective page size, and a processing core, communicatively coupled to the TLB, can execute an instruction associated with virtual memory page to a first physical memory page as mentioned in this paper.
Proceedings ArticleDOI

Heterogeneous Edge CNN Hardware Accelerator

TL;DR: In this article, the authors describe a programmable and scalable Convolutional Neural Network (CNN) hardware accelerator optimized for mobile and edge inference computing, which is comprised of four heterogeneous engines -input engine, filter engine, post processing engine and output engine.