scispace - formally typeset
C

Changwan Hong

Researcher at Massachusetts Institute of Technology

Publications -  26
Citations -  484

Changwan Hong is an academic researcher from Massachusetts Institute of Technology. The author has contributed to research in topics: Speedup & Sparse matrix. The author has an hindex of 10, co-authored 25 publications receiving 279 citations. Previous affiliations of Changwan Hong include Ohio State University.

Papers
More filters
Proceedings ArticleDOI

Adaptive sparse tiling for sparse matrix multiplication

TL;DR: This paper devise an adaptive tiling strategy and apply it to enhance the performance of two primitives: SpMM (product of sparse matrix and dense matrix) and SDDMM (sampled dense-dense matrix multiplication).
Proceedings ArticleDOI

Efficient sparse-matrix multi-vector product on GPUs

TL;DR: An in-depth analysis is presented to contrast SpMV and SpMM, and a new sparse-matrix representation and computation approach suited to achieving high data-movement efficiency and effective GPU parallelization of SpMM is developed.
Journal ArticleDOI

Static and Dynamic Frequency Scaling on Multicore CPUs

TL;DR: This article proposes a lightweight runtime approach that can exploit the properties of the power profile specific to a processor, outperforming classical Linux governors such as powersave or on-demand for computational kernels and demonstrates that it systematically outperforms the powersave Linux governor while also improving overall performance.
Journal ArticleDOI

A sparse iteration space transformation framework for sparse tensor algebra

TL;DR: The results show that the sparse transformations are sufficient to generate code with competitive performance to hand-optimized implementations from the literature, while generalizing to all of the tensor algebra.
Proceedings ArticleDOI

A novel data transformation and execution strategy for accelerating sparse matrix multiplication on GPUs

TL;DR: This work proposes a novel row-reordering technique to improve data locality for SpMM and SDDMM on GPUs by using a hierarchical clustering procedure optimized by locality-sensitive hashing.