scispace - formally typeset
M

Mohammad Shoeybi

Researcher at Nvidia

Publications -  53
Citations -  3377

Mohammad Shoeybi is an academic researcher from Nvidia. The author has contributed to research in topics: Computer science & Jet (fluid). The author has an hindex of 15, co-authored 42 publications receiving 1589 citations. Previous affiliations of Mohammad Shoeybi include Sharif University of Technology & Stanford University.

Papers
More filters
Posted Content

Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism

TL;DR: A simple, efficient intra-layer model parallel approach that enables training transformer models with billions of parameters and shows that careful attention to the placement of layer normalization in BERT-like models is critical to achieving increased performance as the model size grows.
Proceedings Article

Deep Voice: Real-time Neural Text-to-Speech

TL;DR: Deep Voice lays the groundwork for truly end-to-end neural speech synthesis and shows that inference with the system can be performed faster than real time and describes optimized WaveNet inference kernels on both CPU and GPU that achieve up to 400x speedups over existing implementations.
Journal ArticleDOI

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Teven Le Scao, +386 more
- 09 Nov 2022 - 
TL;DR: BLOOM as discussed by the authors is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total).
Journal Article

Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model

TL;DR: The infrastructure as well as the 3D parallelism methodology used to train the largest monolithic transformer based language model, Megatron-Turing NLG 530B (MT-NLG), with 530 billion parameters is presented.
Posted Content

Deep Voice: Real-time Neural Text-to-Speech

TL;DR: Deep Voice as discussed by the authors proposes a segmentation model for locating phoneme boundaries, a grapheme-to-phoneme conversion model, a phoneme duration prediction model, and an audio synthesis model.