J
Joo-Young Kim
Researcher at KAIST
Publications - 91
Citations - 3740
Joo-Young Kim is an academic researcher from KAIST. The author has contributed to research in topics: Computer science & Network on a chip. The author has an hindex of 20, co-authored 68 publications receiving 3415 citations. Previous affiliations of Joo-Young Kim include Microsoft.
Papers
More filters
Journal ArticleDOI
A reconfigurable fabric for accelerating large-scale datacenter services
Andrew Putnam,Adrian M. Caulfield,Eric S. Chung,Derek Chiou,Kypros Constantinides,John Demme,Hadi Esmaeilzadeh,Jeremy Fowers,Gopi Prashanth Gopal,Jan Gray,Michael Haselman,Scott Hauck,Stephen F. Heil,Amir Hormati,Joo-Young Kim,Sitaram Lanka,James R. Larus,Eric C. Peterson,Simon Pope,Aaron L. Smith,Jason Thong,Phillip Yi Xiao,Doug Burger +22 more
TL;DR: The authors deployed the reconfigurable fabric in a bed of 1,632 servers and FPGAs in a production datacenter and successfully used it to accelerate the ranking portion of the Bing Web search engine by nearly a factor of two.
Journal ArticleDOI
A reconfigurable fabric for accelerating large-scale datacenter services
Andrew Putnam,Adrian M. Caulfield,Eric S. Chung,Derek Chiou,Kypros Constantinides,John Demme,Hadi Esmaeilzadeh,Jeremy Fowers,Gopi Prashanth Gopal,Jan Gray,Michael Haselman,Scott Hauck,Stephen F. Heil,Amir Hormati,Joo-Young Kim,Sitaram Lanka,James R. Larus,Eric C. Peterson,Simon Pope,Aaron L. Smith,Jason Thong,Phillip Yi Xiao,Doug Burger +22 more
TL;DR: The requirements and architecture of the fabric are described, the critical engineering challenges and solutions needed to make the system robust in the presence of failures are detailed, and the performance, power, and resilience of the system when ranking candidate documents are measured.
Proceedings ArticleDOI
A cloud-scale acceleration architecture
Adrian M. Caulfield,Eric S. Chung,Andrew Putnam,Hari Angepat,Jeremy Fowers,Michael Haselman,Stephen F. Heil,Matt Humphrey,Puneet Kaur,Joo-Young Kim,Lo Daniel,Todd Massengill,Kalin Ovtcharov,Michael K. Papamichael,Lisa Woods,Sitaram Lanka,Derek Chiou,Doug Burger +17 more
TL;DR: A new cloud architecture that uses reconfigurable logic to accelerate both network plane functions and applications, and is much more scalable than prior work which used secondary rack-scale networks for inter-FPGA communication.
Accelerating Deep Convolutional Neural Networks Using Specialized Hardware
TL;DR: Hardware specialization in the form of GPGPUs, FPGAs, and ASICs offers a promising path towards major leaps in processing capability while achieving high energy efficiency, and combining multiple FPGA over a low-latency communication fabric offers further opportunity to train and evaluate models of unprecedented size and quality.
Journal ArticleDOI
A 201.4 GOPS 496 mW Real-Time Multi-Object Recognition Processor With Bio-Inspired Neural Perception Engine
Joo-Young Kim,Minsu Kim,Seungjin Lee,Jinwook Oh,Kwanho Kim,Sejong Oh,Jeong-Ho Woo,Dong-Hyun Kim,Hoi-Jun Yoo +8 more
TL;DR: In the proposed hardware architecture, three recognition tasks (visual perception, descriptor generation, and object decision) are directly mapped to the neural perception engine, 16 SIMD processors including 128 processing elements, and decision processor and executed in the pipeline to maximize throughput of the object recognition.