scispace - formally typeset
N

Najoung Kim

Researcher at Johns Hopkins University

Publications -  29
Citations -  1619

Najoung Kim is an academic researcher from Johns Hopkins University. The author has contributed to research in topics: Computer science & Language model. The author has an hindex of 10, co-authored 19 publications receiving 776 citations. Previous affiliations of Najoung Kim include KAIST.

Papers
More filters
Journal ArticleDOI

BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

Teven Le Scao, +386 more
- 09 Nov 2022 - 
TL;DR: BLOOM as discussed by the authors is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total).
Posted Content

What do you learn from context? Probing for sentence structure in contextualized word representations

TL;DR: The authors investigate word-level contextual representations from four recent models and investigate how they encode sentence structure across a range of syntactic, semantic, local, and long-range phenomena, finding that existing models trained on language modeling and translation produce strong representations for syntactic phenomena, but only offer comparably small improvements on semantic tasks over a non-contextual baseline.
Proceedings Article

What do you learn from context? Probing for sentence structure in contextualized word representations

TL;DR: A novel edge probing task design is introduced and a broad suite of sub-sentence tasks derived from the traditional structured NLP pipeline are constructed to investigate how sentence structure is encoded across a range of syntactic, semantic, local, and long-range phenomena.
Proceedings ArticleDOI

COGS: A compositional generalization challenge based on semantic interpretation

TL;DR: In experiments with Transformers and LSTMs, it is found that in-distribution accuracy on the COGS test set was near-perfect, but generalization accuracy was substantially lower, and the dataset showed high sensitivity to random seed.
Proceedings ArticleDOI

Probing What Different NLP Tasks Teach Machines about Function Word Comprehension

TL;DR: The results show that pretraining on CCG—the authors' most syntactic objective—performs the best on average across their probing tasks, suggesting that syntactic knowledge helps function word comprehension.