scispace - formally typeset
Open AccessJournal Article

MLlib: machine learning in apache spark

Reads0
Chats0
TLDR
MLlib as mentioned in this paper is an open-source distributed machine learning library for Apache Spark that provides efficient functionality for a wide range of learning settings and includes several underlying statistical, optimization, and linear algebra primitives.
Abstract
Apache Spark is a popular open-source platform for large-scale data processing that is well-suited for iterative machine learning tasks. In this paper we present MLlib, Spark's open-source distributed machine learning library. MLLIB provides efficient functionality for a wide range of learning settings and includes several underlying statistical, optimization, and linear algebra primitives. Shipped with Spark, MLLIB supports several languages and provides a high-level API that leverages Spark's rich ecosystem to simplify the development of end-to-end machine learning pipelines. MLLIB has experienced a rapid growth due to its vibrant open-source community of over 140 contributors, and includes extensive documentation to support further growth and to let users quickly get up to speed.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Cluster-size optimization within a cloud-based ETL framework for Big Data

TL;DR: A cloud-based ETL framework where a general cluster-size optimization algorithm is used, while providing implementation details, and is able to perform the required job within a predefined, and thus known, time.
Posted Content

A Layered Aggregate Engine for Analytics Workloads

TL;DR: LMFAO as discussed by the authors is an in-memory optimization and execution engine for batches of aggregates over the input database, which consists of several layers of logical and code optimizations that systematically exploit sharing of computation, parallelism, and code specialization.
Journal ArticleDOI

Leveraging resource management for efficient performance of Apache Spark

TL;DR: This study uses the Machine Learning library of Spark to implement different machine learning algorithms, then it is used to manage the resources (CPU, memory, and Disk) in order to assess the performance of Apache Spark, and investigates the tuning of the resource allocation in Spark.
Journal ArticleDOI

A utilization model for optimization of checkpoint intervals in distributed stream processing systems

TL;DR: A rigorous derivation of utilization that incorporates checkpoint interval, failure rate, checkpoint cost, failure detection and restart cost, depth of the system topology and message delay yields an elegant expression for utilization and provides an optimal checkpoint interval given these parameters.
Journal ArticleDOI

Big Data: Controlling Fraud by Using Machine Learning Libraries on Spark

TL;DR: This article has used k-Means method from the Machine Learning libraries on Spark to determine whether the incoming network values are normal behavior and detected 10 abnormal behaviors from 400 thousand network data with k-means method.
References
More filters
Journal Article

Scikit-learn: Machine Learning in Python

TL;DR: Scikit-learn is a Python module integrating a wide range of state-of-the-art machine learning algorithms for medium-scale supervised and unsupervised problems, focusing on bringing machine learning to non-specialists using a general-purpose high-level language.
Journal ArticleDOI

Latent dirichlet allocation

TL;DR: This work proposes a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams, and Hofmann's aspect model.
Proceedings Article

Latent Dirichlet Allocation

TL;DR: This paper proposed a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams, and Hof-mann's aspect model, also known as probabilistic latent semantic indexing (pLSI).
Journal ArticleDOI

MapReduce: simplified data processing on large clusters

TL;DR: This paper presents the implementation of MapReduce, a programming model and an associated implementation for processing and generating large data sets that runs on a large cluster of commodity machines and is highly scalable.
Related Papers (5)