scispace - formally typeset
Proceedings ArticleDOI

Odessa: enabling interactive perception applications on mobile devices

TLDR
Odessa is developed, a novel, lightweight, runtime that automatically and adaptively makes offloading and parallelism decisions for mobile interactive perception applications and provides more than a 3x improvement in application performance over partitioning suggested by domain experts.
Abstract
Resource constrained mobile devices need to leverage computation on nearby servers to run responsive applications that recognize objects, people, or gestures from real-time video. The two key questions that impact performance are what computation to offload, and how to structure the parallelism across the mobile device and server. To answer these questions, we develop and evaluate three interactive perceptual applications. We find that offloading and parallelism choices should be dynamic, even for a given application, as performance depends on scene complexity as well as environmental factors such as the network and device capabilities. To this end we develop Odessa, a novel, lightweight, runtime that automatically and adaptively makes offloading and parallelism decisions for mobile interactive perception applications. Our evaluation shows that the incremental greedy strategy of Odessa converges to an operating point that is close to an ideal offline partitioning. It provides more than a 3x improvement in application performance over partitioning suggested by domain experts. Odessa works well across a variety of execution environments, and is agile to changes in the network, device and application inputs.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

ThinkAir: Dynamic resource allocation and parallel execution in the cloud for mobile code offloading

TL;DR: This paper proposes ThinkAir, a framework that makes it simple for developers to migrate their smartphone applications to the cloud and enhances the power of mobile cloud computing by parallelizing method execution using multiple virtual machine (VM) images.
Proceedings ArticleDOI

Neurosurgeon: Collaborative Intelligence Between the Cloud and Mobile Edge

TL;DR: Neurosurgeon, a lightweight scheduler to automatically partition DNN computation between mobile devices and datacenters at the granularity of neural network layers is designed, finding that a fine-grained, layer-level computation partitioning strategy based on the data and computation variations of each layer within a DNN has significant latency and energy advantages over the status quo approach.
Journal ArticleDOI

Deep Learning With Edge Computing: A Review

TL;DR: This paper will provide an overview of applications where deep learning is used at the network edge, discuss various approaches for quickly executing deep learning inference across a combination of end devices, edge servers, and the cloud, and describe the methods for training deep learning models across multiple edge devices.
Journal ArticleDOI

Offloading in Mobile Edge Computing: Task Allocation and Computational Frequency Scaling

TL;DR: This paper proposes an optimization framework of offloading from a single mobile device (MD) to multiple edge devices and proposes a linear relaxation-based approach and a semidefinite relaxation (SDR)-based approach for the fixed CPU frequency case, and an exhaustive search- based approach and an SDR-based approaches for the elasticCPU frequency case.
Journal ArticleDOI

Future Edge Cloud and Edge Computing for Internet of Things Applications

TL;DR: This survey paper investigates the key rationale, the state-of-the-art efforts, the key enabling technologies and research topics, and typical IoT applications benefiting from edge cloud.
References
More filters
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Book

Computers and Intractability: A Guide to the Theory of NP-Completeness

TL;DR: The second edition of a quarterly column as discussed by the authors provides a continuing update to the list of problems (NP-complete and harder) presented by M. R. Garey and myself in our book "Computers and Intractability: A Guide to the Theory of NP-Completeness,” W. H. Freeman & Co., San Francisco, 1979.
Journal ArticleDOI

MapReduce: simplified data processing on large clusters

TL;DR: This paper presents the implementation of MapReduce, a programming model and an associated implementation for processing and generating large data sets that runs on a large cluster of commodity machines and is highly scalable.
Journal ArticleDOI

MapReduce: simplified data processing on large clusters

TL;DR: This presentation explains how the underlying runtime system automatically parallelizes the computation across large-scale clusters of machines, handles machine failures, and schedules inter-machine communication to make efficient use of the network and disks.
Related Papers (5)