scispace - formally typeset
Journal ArticleDOI

Apache Spark: a unified engine for big data processing

Reads0
Chats0
TLDR
This open source computing framework unifies streaming, batch, and interactive big data workloads to unlock new applications.
Abstract
This open source computing framework unifies streaming, batch, and interactive big data workloads to unlock new applications

read more

Citations
More filters
Proceedings ArticleDOI

Efficient shuffle management with SCache for DAG computing frameworks

TL;DR: Evaluated evaluations have demonstrated that, by incorporating SCache, the shuffle overhead of Spark can be reduced by nearly 89%, and the overall completion time of TPC-DS queries improves 40% on average.
Posted Content

AFrame: Extending DataFrames for Large-Scale Modern Data Analysis (Extended Version).

TL;DR: AFrame as mentioned in this paper is a new scalable data analysis package powered by a big data management system that extends the data scientists' familiar DataFrame operations to efficiently operate on managed data at scale, which can result in serious productivity issues for "normal" data scientists.
Proceedings ArticleDOI

Living on the Edge: Serverless Computing and the Cost of Failure Resiliency

TL;DR: The latency, throughput, and resource costs of providing different reliability guarantees, with a focus on emerging edge cloud platforms and applications, are evaluated.
Journal ArticleDOI

20 GB in 10 minutes: A case for linking major biodiversity databases using an open socio-technical infrastructure and a pragmatic, cross-institutional collaboration

TL;DR: A technical solution developed by the GUODA collaboration for faster linking across databases with a use case linking Wikidata and the Global Biotic Interactions database (GloBI).
Posted Content

Artificial Intelligence versus Maya Angelou: Experimental evidence that people cannot differentiate AI-generated from human-written poetry

TL;DR: What these results convey about the performance of NLG algorithms to produce human-like text are discussed and methodologies to study such learning algorithms in human-agent experimental settings are proposed.
References
More filters
Journal ArticleDOI

MapReduce: simplified data processing on large clusters

TL;DR: This paper presents the implementation of MapReduce, a programming model and an associated implementation for processing and generating large data sets that runs on a large cluster of commodity machines and is highly scalable.
Proceedings Article

Resilient distributed datasets: a fault-tolerant abstraction for in-memory cluster computing

TL;DR: Resilient Distributed Datasets is presented, a distributed memory abstraction that lets programmers perform in-memory computations on large clusters in a fault-tolerant manner and is implemented in a system called Spark, which is evaluated through a variety of user applications and benchmarks.
Journal ArticleDOI

A bridging model for parallel computation

TL;DR: The bulk-synchronous parallel (BSP) model is introduced as a candidate for this role, and results quantifying its efficiency both in implementing high-level language features and algorithms, as well as in being implemented in hardware.
Proceedings ArticleDOI

Pregel: a system for large-scale graph processing

TL;DR: A model for processing large graphs that has been designed for efficient, scalable and fault-tolerant implementation on clusters of thousands of commodity computers, and its implied synchronicity makes reasoning about programs easier.
Proceedings ArticleDOI

Dryad: distributed data-parallel programs from sequential building blocks

TL;DR: The Dryad execution engine handles all the difficult problems of creating a large distributed, concurrent application: scheduling the use of computers and their CPUs, recovering from communication or computer failures, and transporting data between vertices.
Related Papers (5)