scispace - formally typeset
R

Ruslan Salakhutdinov

Researcher at Carnegie Mellon University

Publications -  457
Citations -  142495

Ruslan Salakhutdinov is an academic researcher from Carnegie Mellon University. The author has contributed to research in topics: Computer science & Artificial neural network. The author has an hindex of 107, co-authored 410 publications receiving 115921 citations. Previous affiliations of Ruslan Salakhutdinov include Carnegie Learning & University of Toronto.

Papers
More filters
Proceedings Article

Matrix reconstruction with the local max norm

TL;DR: In this paper, a new family of matrix norms, the local max norm (local max) is introduced, which generalizes existing methods such as the max norm, the nuclear norm, and the weighted or smoothed weighted trace norm.
Posted Content

AutoLoss: Learning Discrete Schedules for Alternate Optimization

TL;DR: The authors proposed AutoLoss, a meta-learning framework that automatically learns and determines the optimization schedule for machine learning problems, which allows for a dynamic and data-driven schedule in ML problems that involve alternating updates of different parameters or loss objectives.
Journal ArticleDOI

Effective Data Augmentation With Diffusion Models

TL;DR: In this article , a text-to-image diffusion model is used to transform images to change their semantics using an off-the-shelf diffusion model, and generalizes to novel visual concepts from a few labeled examples.
Posted Content

How Many Samples are Needed to Estimate a Convolutional or Recurrent Neural Network

TL;DR: It is shown that the sample-complexity to learn CNNs and RNNs scales linearly with their intrinsic dimension and this sample- Complexity is much smaller than for their FNN counterparts.
Posted Content

Deep Neural Networks with Multi-Branch Architectures Are Less Non-Convex

TL;DR: It is shown that the duality gap of both population and empirical risks shrinks to zero as the number of branches increases, and sheds light on better understanding the power of over-parametrization where increasing the network width tends to make the loss surface less non-convex.