scispace - formally typeset
Z

Zihang Dai

Researcher at Google

Publications -  57
Citations -  14783

Zihang Dai is an academic researcher from Google. The author has contributed to research in topics: Language model & Computer science. The author has an hindex of 28, co-authored 52 publications receiving 9340 citations. Previous affiliations of Zihang Dai include Baidu & Carnegie Mellon University.

Papers
More filters
Posted Content

Large-scale Cloze Test Dataset Created by Teachers

TL;DR: This article proposed the first large-scale human-created cloze test dataset CLOTH, containing questions used in middle-school and high-school language exams, which requires a deeper language understanding and a wider attention span than previously automatically-generated cloze datasets.
Posted Content

SimVLM: Simple Visual Language Model Pretraining with Weak Supervision

TL;DR: SimVLM as discussed by the authors reduces the training complexity by exploiting large-scale weak supervision, and is trained end-to-end with a single prefix language modeling objective, achieving state-of-the-art results on a wide range of discriminative and generative vision-language benchmarks.
Journal ArticleDOI

Prediction of the sulfate resistance for recycled aggregate concrete based on ensemble learning algorithms

TL;DR: In this paper , four ensemble learning methods (random forest, adaptive boosting, gradient boosting, and extreme gradient boosting) were employed to establish the predictive model for sulfate resistance of aggregate concrete.
Proceedings Article

Pay Attention to MLPs

TL;DR: This paper proposed a simple network architecture, gMLP, based on MLPs with gating, and showed that it can perform as well as Transformers in key language and vision applications.
Posted Content

SwitchOut: an Efficient Data Augmentation Algorithm for Neural Machine Translation

TL;DR: In this paper, a simple data augmentation strategy called SwitchOut was proposed, which randomly replaces words in both the source sentence and the target sentence with other random words from their corresponding vocabularies.