scispace - formally typeset
M

Mohammad Taghi Saffar

Researcher at Google

Publications -  20
Citations -  547

Mohammad Taghi Saffar is an academic researcher from Google. The author has contributed to research in topics: Context (language use) & Computer science. The author has an hindex of 4, co-authored 16 publications receiving 190 citations. Previous affiliations of Mohammad Taghi Saffar include University of Tehran & University of Nevada, Reno.

Papers
More filters
Posted Content

Efficient Content-Based Sparse Attention with Routing Transformers

TL;DR: This work proposes to learn dynamic sparse attention patterns that avoid allocating computation and memory to attend to content unrelated to the query of interest, and shows that this model outperforms comparable sparse attention models on language modeling on Wikitext-103, as well as on image generation on ImageNet-64 while using fewer self-attention layers.
Journal ArticleDOI

Efficient Content-Based Sparse Attention with Routing Transformers

TL;DR: The Routing Transformer as discussed by the authors proposes to learn dynamic sparse attention patterns that avoid allocating computation and memory to attend to content unrelated to the query of interest by combining self-attention with a sparse routing module based on online k-means.
Proceedings ArticleDOI

Phenaki: Variable Length Video Generation From Open Domain Textual Description

TL;DR: A new model for learning video representation which compresses the video to a small representation of discrete tokens, which results in better spatio-temporal consistency and joint training on a large corpus of image-text pairs as well as a smaller number of video-text examples can result in generalization beyond what is available in the video datasets.
Journal ArticleDOI

A Scale and Translation Invariant Approach for Early Classification of Spatio-Temporal Patterns Using Spiking Neural Networks

TL;DR: The results show that the proposed approach significantly outperforms these methods, it is invariant to scale and translation, and it has the ability to recognize patterns from only partial information.
Journal ArticleDOI

Answer-Me: Multi-Task Open-Vocabulary Visual Question Answering

TL;DR: This work proposes a novel and simple recipe to pre-train a vision-language joint model, which is multi-task as well, and observes that the proposed approach is able to generalize to unseen tasks and that more diverse mixtures lead to higher accuracy in both known and novel tasks.