scispace - formally typeset
A

Alex Tamkin

Researcher at Stanford University

Publications -  27
Citations -  348

Alex Tamkin is an academic researcher from Stanford University. The author has contributed to research in topics: Computer science & Markov decision process. The author has an hindex of 8, co-authored 15 publications receiving 177 citations.

Papers
More filters
Posted Content

On the Opportunities and Risks of Foundation Models.

Rishi Bommasani, +113 more
- 16 Aug 2021 - 
TL;DR: The authors provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (e.g., language, vision, robotics, reasoning, human interaction) and technical principles(e. g.g. model architectures, training procedures, data, systems, security, evaluation, theory) to their applications.
Proceedings ArticleDOI

Investigating Transferability in Pretrained Language Models

TL;DR: This technique reveals that in BERT, layers with high probing performance on downstream GLUE tasks are neither necessary nor sufficient for high accuracy on those tasks, and the benefit of using pretrained parameters for a layer varies dramatically with dataset size.
Proceedings ArticleDOI

drone.io: a gestural and visual interface for human-drone interaction

TL;DR: The design process of drone.io is described, a projected body-centric graphical user interface for human-drone interaction embedded on a drone to provide both input and output capabilities, and it is reported that people were able to use the interface with little prior training.
Posted Content

Viewmaker Networks: Learning Views for Unsupervised Representation Learning

TL;DR: This work proposes viewmaker networks: generative models that learn to produce input-dependent views for contrastive learning, and demonstrates that learned views are a promising way to reduce the amount of expertise and effort needed for unsupervised learning, potentially extending its benefits to a much wider set of domains.
Posted Content

Being Optimistic to Be Conservative: Quickly Learning a CVaR Policy

TL;DR: This paper presents the first algorithm for sample-efficient learning of CVaR-optimal policies in Markov decision processes based on the optimism in the face of uncertainty principle by relying on a novel optimistic version of the distributional Bellman operator that moves probability mass from the lower to the upper tail of the return distribution.