scispace - formally typeset
C

Christopher De Sa

Researcher at Cornell University

Publications -  134
Citations -  4050

Christopher De Sa is an academic researcher from Cornell University. The author has contributed to research in topics: Computer science & Gibbs sampling. The author has an hindex of 27, co-authored 107 publications receiving 3104 citations. Previous affiliations of Christopher De Sa include Stanford University.

Papers
More filters
Proceedings Article

Data Programming: Creating Large Training Sets, Quickly

TL;DR: Data programming as discussed by the authors proposes a paradigm for the programmatic creation of training sets called data programming in which users provide a set of labeling functions, which are programs that heuristically label subsets of the data, but that are noisy and may conflict.
Posted Content

Data Programming: Creating Large Training Sets, Quickly

TL;DR: A paradigm for the programmatic creation of training sets called data programming is proposed in which users express weak supervision strategies or domain heuristics as labeling functions, which are programs that label subsets of the data, but that are noisy and may conflict.
Journal ArticleDOI

Incremental knowledge base construction using DeepDive

TL;DR: This work describes DeepDive, a system that combines database and machine learning ideas to help develop KBC systems, and it presents techniques to make the KBC process more efficient, and proposes two methods for incremental inference, based respectively on sampling and variational techniques.
Proceedings Article

Taming the wild: a unified analysis of HOG WILD! -style algorithms

TL;DR: This work uses a martingale-based analysis to derive convergence rates for the convex case (Hogwild!) with relaxed assumptions on the sparsity of the problem and designs and analyzes an asynchronous SGD algorithm, called Buckwild!, that uses lower-precision arithmetic.
Posted Content

Incremental Knowledge Base Construction Using DeepDive

TL;DR: This work describes DeepDive, a system that combines database and machine learning ideas to help develop KBC systems, and presents techniques to make the KBC process more efficient, and proposes two methods for incremental inference, based, respectively, on sampling and variational techniques.