scispace - formally typeset
J

Jared Casper

Researcher at Nvidia

Publications -  28
Citations -  8773

Jared Casper is an academic researcher from Nvidia. The author has contributed to research in topics: Transactional memory & Software transactional memory. The author has an hindex of 18, co-authored 25 publications receiving 6648 citations. Previous affiliations of Jared Casper include Massachusetts Institute of Technology & Stanford University.

Papers
More filters
Posted Content

Deep Speech: Scaling up end-to-end speech recognition

TL;DR: Deep Speech, a state-of-the-art speech recognition system developed using end-to-end deep learning, outperforms previously published results on the widely studied Switchboard Hub5'00, achieving 16.0% error on the full test set.
Posted Content

Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism

TL;DR: A simple, efficient intra-layer model parallel approach that enables training transformer models with billions of parameters and shows that careful attention to the placement of layer normalization in BERT-like models is critical to achieving increased performance as the model size grows.
Journal ArticleDOI

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Teven Le Scao, +386 more
- 09 Nov 2022 - 
TL;DR: BLOOM as discussed by the authors is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total).