scispace - formally typeset
D

Dan Alistarh

Researcher at Institute of Science and Technology Austria

Publications -  213
Citations -  4887

Dan Alistarh is an academic researcher from Institute of Science and Technology Austria. The author has contributed to research in topics: Computer science & Stochastic gradient descent. The author has an hindex of 27, co-authored 175 publications receiving 3761 citations. Previous affiliations of Dan Alistarh include ETH Zurich & Microsoft.

Papers
More filters
Posted Content

Asynchronous Stochastic Subgradient Methods for General Nonsmooth Nonconvex Optimization

TL;DR: This paper introduces the first convergence analysis covering asynchronous methods in the case of general non-smooth, non-convex objectives and shows their overall successful asymptotic convergence as well as exploring how momentum, synchronization, and partitioning all affect performance.
Journal ArticleDOI

Inherent limitations of hybrid transactional memory

TL;DR: A general model for HyTM implementations, which captures the ability of hardware transactions to buffer memory accesses and captures for the first time the trade-off between the degree of hardware-software TM concurrency and the amount of instrumentation overhead.
Posted Content

Fast Graphical Population Protocols.

TL;DR: In this article, the authors consider the more general setting where a graph is an arbitrary graph, and present a technique for simulating protocols designed for fully-connected networks in any connected regular graph.
Posted Content

On the Sample Complexity of Adversarial Multi-Source PAC Learning

TL;DR: In this paper, the problem of learning from multiple untrusted data sources has been studied and a generalization of PAC-learnability has been shown to the multi-source setting, where the adversary can arbitrarily corrupt a fixed fraction of the data sources.
Proceedings ArticleDOI

Efficiency Guarantees for Parallel Incremental Algorithms under Relaxed Schedulers

TL;DR: This paper analyzes the efficiency guarantees provided by a range of incremental algorithms when parallelized via relaxed schedulers, and provides lower bounds showing that certain algorithms will inherently incur a non-trivial amount of wasted work due to scheduler relaxation.