J
Joseph E. Gonzalez
Researcher at University of California, Berkeley
Publications - 229
Citations - 19900
Joseph E. Gonzalez is an academic researcher from University of California, Berkeley. The author has contributed to research in topics: Computer science & Reinforcement learning. The author has an hindex of 49, co-authored 209 publications receiving 15003 citations. Previous affiliations of Joseph E. Gonzalez include UCB & Facebook.
Papers
More filters
Journal ArticleDOI
Apache Spark: a unified engine for big data processing
Matei Zaharia,Reynold Xin,Patrick Wendell,Tathagata Das,Michael Armbrust,Ankur Dave,Xiangrui Meng,Josh Rosen,Shivaram Venkataraman,Michael J. Franklin,Ali Ghodsi,Joseph E. Gonzalez,Scott Shenker,Ion Stoica +13 more
TL;DR: This open source computing framework unifies streaming, batch, and interactive big data workloads to unlock new applications.
Proceedings ArticleDOI
PowerGraph: distributed graph-parallel computation on natural graphs
TL;DR: This paper describes the challenges of computation on natural graphs in the context of existing graph-parallel abstractions and introduces the PowerGraph abstraction which exploits the internal structure of graph programs to address these challenges.
Journal ArticleDOI
Distributed GraphLab: a framework for machine learning and data mining in the cloud
Yucheng Low,Danny Bickson,Joseph E. Gonzalez,Carlos Guestrin,Aapo Kyrola,Joseph M. Hellerstein +5 more
TL;DR: GraphLab as discussed by the authors extends the GraphLab framework to the substantially more challenging distributed setting while preserving strong data consistency guarantees to reduce network congestion and mitigate the effect of network latency in the shared-memory setting.
Proceedings ArticleDOI
GraphX: graph processing in a distributed dataflow framework
TL;DR: This paper introduces GraphX, an embedded graph processing framework built on top of Apache Spark, a widely used distributed dataflow system and demonstrates that GraphX achieves an order of magnitude performance gain over the base dataflow framework and matches the performance of specialized graph processing systems while enabling a wider range of computation.
Proceedings Article
GraphLab: a new framework for parallel machine learning
Yucheng Low,Joseph E. Gonzalez,Aapo Kyrola,Danny Bickson,Carlos Guestrin,Joseph M. Hellerstein +5 more
TL;DR: The expressiveness of the GraphLab framework is demonstrated by designing and implementing parallel versions of belief propagation, Gibbs sampling, Co-EM, Lasso and Compressed Sensing and it is shown that using GraphLab the authors can achieve excellent parallel performance on large scale real-world problems.