scispace - formally typeset
S

Shaden Smith

Researcher at University of Minnesota

Publications -  35
Citations -  1846

Shaden Smith is an academic researcher from University of Minnesota. The author has contributed to research in topics: Speedup & Tensor. The author has an hindex of 17, co-authored 32 publications receiving 792 citations. Previous affiliations of Shaden Smith include Microsoft & University of Kentucky.

Papers
More filters
Journal ArticleDOI

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Teven Le Scao, +386 more
- 09 Nov 2022 - 
TL;DR: BLOOM as discussed by the authors is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total).
Journal Article

Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model

TL;DR: The infrastructure as well as the 3D parallelism methodology used to train the largest monolithic transformer based language model, Megatron-Turing NLG 530B (MT-NLG), with 530 billion parameters is presented.
Proceedings ArticleDOI

SPLATT: Efficient and Parallel Sparse Tensor-Matrix Multiplication

TL;DR: SPLATT as discussed by the authors is a C library with shared-memory parallelism for three-mode tensors that uses a data structure that exploits the sparsity patterns of tensors.
Proceedings ArticleDOI

Tensor-matrix products with a compressed sparse tensor

TL;DR: The compressed sparse fiber (CSF) a data structure for sparse tensors along with a novel parallel algorithm for tensor-matrix multiplication is introduced and offers similar operation reductions as existing compressed methods while using only a single tensor structure.
Proceedings ArticleDOI

ZeRO-infinity: breaking the GPU memory wall for extreme scale deep learning

TL;DR: The ZeRO-Infinity project as mentioned in this paper leverages GPU, CPU, and NVMe memory to allow for unprecedented model scale on limited resources without requiring model code refactoring.