scispace - formally typeset
B

Bohan Li

Researcher at Carnegie Mellon University

Publications -  10
Citations -  568

Bohan Li is an academic researcher from Carnegie Mellon University. The author has contributed to research in topics: Autoencoder & Language model. The author has an hindex of 5, co-authored 10 publications receiving 180 citations. Previous affiliations of Bohan Li include Zhejiang University.

Papers
More filters
Posted Content

On the Sentence Embeddings from Pre-trained Language Models

TL;DR: This paper proposes to transform the anisotropic sentence embedding distribution to a smooth and isotropic Gaussian distribution through normalizing flows that are learned with an unsupervised objective and achieves significant performance gains over the state-of-the-art sentence embeddings on a variety of semantic textual similarity tasks.
Proceedings ArticleDOI

On the Sentence Embeddings from Pre-trained Language Models

TL;DR: BERT-flow as mentioned in this paper transforms the anisotropic sentence embedding distribution to a smooth and isotropic Gaussian distribution through normalizing flows that are learned with an unsupervised objective.
Proceedings ArticleDOI

A Surprisingly Effective Fix for Deep Latent Variable Modeling of Text

TL;DR: In this paper, the authors investigate a simple fix for posterior collapse which yields surprisingly effective results, and demonstrate that the typical surrogate objective for VAEs may not be sufficient or necessarily appropriate for balancing the goals of representation learning and data distribution modeling.
Posted Content

An Adversarial Approach to High-Quality, Sentiment-Controlled Neural Dialogue Generation.

TL;DR: Experimental results using automatic and human evaluation both demonstrate that the proposed framework for neural dialogue response generation is able to generate both semantically reasonable and sentiment-controlled dialogue responses.
Posted Content

A Surprisingly Effective Fix for Deep Latent Variable Modeling of Text

TL;DR: A simple fix for posterior collapse is investigated which yields surprisingly effective results and is used to argue that the typical surrogate objective for VAEs may not be sufficient or necessarily appropriate for balancing the goals of representation learning and data distribution modeling.