scispace - formally typeset
A

Alan Ansell

Publications -  5
Citations -  20

Alan Ansell is an academic researcher. The author has contributed to research in topics: Language model & Conflation. The author has an hindex of 2, co-authored 5 publications receiving 16 citations.

Papers
More filters
Proceedings Article

MAD-G: Multilingual Adapter Generation for Efficient Cross-Lingual Transfer.

TL;DR: The authors propose MAD-G (Multilingual ADapter Generation) which generates language adapters from language representations based on typological features. But this approach is not viable for the vast majority of languages, due to limitations in their corpus size or compute budgets.
Proceedings Article

An ELMo-inspired approach to SemDeep-5’s Word-in-Context task

TL;DR: This paper took an ELMo-inspired approach similar to the baseline model in the task description paper, where contextualized representations are obtained for the focus words and a classification is made according to the degree of similarity between these representations.
Proceedings ArticleDOI

PolyLM: Learning about Polysemy through Language Modeling

TL;DR: Ansell et al. as mentioned in this paper introduced PolyLM, a method which formulates the task of learning sense embeddings as a language modeling problem, allowing contextualization techniques to be applied.
Posted Content

PolyLM: Learning about Polysemy through Language Modeling

TL;DR: PolyLM as mentioned in this paper formulates the task of learning sense embeddings as a language modeling problem, allowing contextualization techniques to be applied, and is based on two underlying assumptions about word senses: first, that the probability of a word occurring in a given context is equal to the sum of the probabilities of its individual senses occurring; and secondly, for a given occurrence of the word, one of its senses tends to be much more plausible in the context than others.
Posted Content

Composable Sparse Fine-Tuning for Cross-Lingual Transfer.

TL;DR: In this article, the authors propose to learn sparse, real-valued masks based on a simple variant of the Lottery Ticket Hypothesis and then combine these masks with the pre-trained model.