scispace - formally typeset
T

Tuan Nghia Nguyen

Researcher at Seoul National University

Publications -  7
Citations -  261

Tuan Nghia Nguyen is an academic researcher from Seoul National University. The author has contributed to research in topics: Computer science & Dynamic random-access memory. The author has an hindex of 1, co-authored 4 publications receiving 92 citations.

Papers
More filters
Journal ArticleDOI

A High-Throughput and Power-Efficient FPGA Implementation of YOLO CNN for Object Detection

TL;DR: This paper presents a Tera-OPS streaming hardware accelerator implementing a you-only-look-once (YOLO) CNN, which outperforms the “one-size-fits-all” designs in both performance and power efficiency.
Proceedings ArticleDOI

A Real-time Super-resolution Accelerator Using a big. LITTLE Core Architecture

TL;DR: This work proposes an SR accelerator using a big.LITTLE core architecture which is able to execute various networks in real-time and achieves an inference speed of 36.63 frames per second and a throughput of 221.79 GOPs at 200MHz for ×2 SR while using only 1,280 eight-bit multipliers and 330 KB on-chip SRAM.
Proceedings ArticleDOI

A Lightweight YOLOv2 Object Detector Using a Dilated Convolution

TL;DR: This study proposes a simple yet effective variant of YOLOv2 by using dilated convolution to reduce its complexity, and results show that the proposed method achieves up to 36% of memory access reduction and 9% of computation reduction on a network with a negligible performance loss.
Posted Content

ShortcutFusion: From Tensorflow to FPGA-based accelerator with reuse-aware memory allocation for shortcut data.

TL;DR: ShortcutFusion as discussed by the authors is an optimization tool for FPGA-based accelerator with a reuse-aware static memory allocation for shortcut data, to maximize on-chip data reuse given resource constraints.
Proceedings ArticleDOI

Computation-Skipping Mask Generation for Super-Resolution Networks

TL;DR: This study introduces a computation-skipping mask (CSM) generation framework to reduce redundant computations of Super-Resolution (SR) neural networks, which reduces the computational cost up to 58%, with negligible performance degradation for various SR models.