scispace - formally typeset
W

Wenhui Wang

Researcher at Microsoft

Publications -  81
Citations -  4273

Wenhui Wang is an academic researcher from Microsoft. The author has contributed to research in topics: Computer science & Medicine. The author has an hindex of 15, co-authored 29 publications receiving 2424 citations. Previous affiliations of Wenhui Wang include Peking University.

Papers
More filters
Proceedings Article

Unified Language Model Pre-training for Natural Language Understanding and Generation

TL;DR: UniLM as mentioned in this paper is a unified pre-trained language model that can be fine-tuned for both natural language understanding and generation tasks, achieving state-of-the-art results on five natural language generation datasets, including improving the CNN/DailyMail abstractive summarization ROUGE-L to 40.51 (2.04 absolute improvement).
Proceedings ArticleDOI

Gated Self-Matching Networks for Reading Comprehension and Question Answering

TL;DR: The gated self-matching networks for reading comprehension style question answering, which aims to answer questions from a given passage, are presented and holds the first place on the SQuAD leaderboard for both single and ensemble model.
Posted Content

Unified Language Model Pre-training for Natural Language Understanding and Generation

TL;DR: A new Unified pre-trained Language Model (UniLM) that can be fine-tuned for both natural language understanding and generation tasks that compares favorably with BERT on the GLUE benchmark, and the SQuAD 2.0 and CoQA question answering tasks.
Proceedings Article

MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers

TL;DR: The authors distill the self-attention module of the last Transformer layer of the teacher, which is effective and flexible for the student, and introduce the scaled dot-product between values in the selfatt attention module as the new deep selfattention knowledge, in addition to the attention distributions.
Journal ArticleDOI

Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks

TL;DR: This work introduces a general- Purpose multimodal foundation model BE I T-3, which achieves state-of-the-art transfer performance on both vision and vision-language tasks and introduces Multiway Transformers for general-purpose modeling.