V
Vedanuj Goswami
Researcher at Facebook
Publications - 27
Citations - 1115
Vedanuj Goswami is an academic researcher from Facebook. The author has contributed to research in topics: Computer science & Question answering. The author has an hindex of 8, co-authored 18 publications receiving 387 citations. Previous affiliations of Vedanuj Goswami include National Institute of Technology, Silchar & Georgia Institute of Technology.
Papers
More filters
Proceedings ArticleDOI
12-in-1: Multi-Task Vision and Language Representation Learning
TL;DR: This paper investigated the relationship between vision and language tasks by developing a large-scale, multi-task model, which culminates in a single model on 12 datasets from four broad categories of task including visual question answering, caption-based image retrieval, grounding referring expressions, and multimodal verification.
Journal ArticleDOI
No Language Left Behind: Scaling Human-Centered Machine Translation
Nllb team,Marta R. Costa-jussà,James Cross,Onur cCelebi,Maha Elbayad,Kenneth Heafield,Kevin Heffernan,Elahe Kalbassi,Janice Si-Man Lam,Daniel Licht,Jean Maillard,Anna Sun,Skyler Wang,Guillaume Wenzek,Alison Youngblood,Bapi Akula,Loïc Barrault,Gabriel Mejia Gonzalez,Prangthip Hansanti,John Hoffman,Semarley Jarrett,Kaushik Ram Sadagopan,Dirk Rowe,Shannon Spruit,Chau Tran,Pierre Andrews,Necip Fazil Ayan,Shruti Bhosale,Sergey Edunov,Angela Fan,Cynthia Gao,Vedanuj Goswami,Francisco Guzm'an,Philipp Koehn,Alexandre Mourachko,Christophe Ropers,Safiyyah Saleem,Holger Schwenk,Jeff Wang +38 more
TL;DR: A conditional compute model based on Sparsely Gated Mixture of Experts that is trained on data obtained with novel and effective data mining techniques tailored for low-resource languages is developed, laying important groundwork towards realizing a universal translation system.
Proceedings Article
The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes
Douwe Kiela,Hamed Firooz,Aravind Mohan,Vedanuj Goswami,Amanpreet Singh,Pratik Ringshia,Davide Testuggine +6 more
TL;DR: This work proposes a new challenge set for multimodal classification, focusing on detecting hate speech in multi-modal memes, constructed such that unimodal models struggle and only multimodAL models can succeed.
Posted Content
12-in-1: Multi-Task Vision and Language Representation Learning
TL;DR: This work develops a large-scale, multi-task model that culminates in a single model on 12 datasets from four broad categories of task including visual question answering, caption-based image retrieval, grounding referring expressions, and multimodal verification and shows that finetuning task-specific models from this model can lead to further improvements, achieving performance at or above the state-of-the-art.
Posted Content
The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes
Douwe Kiela,Hamed Firooz,Aravind Mohan,Vedanuj Goswami,Amanpreet Singh,Pratik Ringshia,Davide Testuggine +6 more
TL;DR: The authors proposed a new challenge set for multimodal classification, focusing on detecting hate speech in multi-modal memes, where difficult examples are added to the dataset to make it hard to rely on unimodal signals.