L
Lei Wang
Researcher at University of Science and Technology Beijing
Publications - 8
Citations - 298
Lei Wang is an academic researcher from University of Science and Technology Beijing. The author has contributed to research in topics: Computer science & Complex network. The author has an hindex of 4, co-authored 6 publications receiving 32 citations.
Papers
More filters
Journal ArticleDOI
Pruning and quantization for deep neural network acceleration: A survey
TL;DR: A survey on two types of network compression: pruning and quantization is provided, which compare current techniques, analyze their strengths and weaknesses, provide guidance for compressing networks, and discuss possible future compression techniques.
Posted Content
Pruning and Quantization for Deep Neural Network Acceleration: A Survey
TL;DR: In this article, the authors provide a survey on two types of network compression: pruning and quantization, and compare current techniques, analyze their strengths and weaknesses, present compressed network accuracy results on a number of frameworks, and provide practical guidance for compressing networks.
Posted Content
Dynamic Runtime Feature Map Pruning
TL;DR: This work analyzes parameter sparsity of six popular convolutional neural networks and introduces dynamic runtime pruning of feature maps, showing that 10% of dynamic feature map execution can be removed without loss of accuracy.
Patent
Variable translation-lookaside buffer (TLB) indexing
Mayan Moudgill,A. Joseph Hoane,Lei Wang,Gary Nacer,Aaron G. Milbury,Enrique A. Barria,Hurtley Paul +6 more
TL;DR: A processor includes a translation lookaside buffer (TLB) comprising a plurality of ways, wherein each way is associated with a respective page size, and a processing core, communicatively coupled to the TLB, can execute an instruction associated with virtual memory page to a first physical memory page as mentioned in this paper.
Proceedings ArticleDOI
Heterogeneous Edge CNN Hardware Accelerator
Mayan Moudgill,John Glossner,Huang Wei,Chaoyang Tian,Chunxia Xu,Nianliang Yang,Lei Wang,Tailin Liang,Shaobo Shi,Xiaodong Zhang,Daniel Iancu,Gary Nacer,Kerry Li +12 more
TL;DR: In this article, the authors describe a programmable and scalable Convolutional Neural Network (CNN) hardware accelerator optimized for mobile and edge inference computing, which is comprised of four heterogeneous engines -input engine, filter engine, post processing engine and output engine.