scispace - formally typeset
Search or ask a question

Showing papers by "Sasu Tarkoma published in 2020"


Journal ArticleDOI
TL;DR: IoT-Keeper is a lightweight system which secures the communication of IoT and uses the proposed anomaly detection technique to perform traffic analysis at edge gateways, and can detect and mitigate various network attacks—without requiring explicit attack signatures or sophisticated hardware.
Abstract: IoT devices are notoriously vulnerable even to trivial attacks and can be easily compromised. In addition, resource constraints and heterogeneity of IoT devices make it impractical to secure IoT installations using traditional endpoint and network security solutions. To address this problem, we present IoT-Keeper, a lightweight system which secures the communication of IoT. IoT-Keeper uses our proposed anomaly detection technique to perform traffic analysis at edge gateways. It uses a combination of fuzzy C-means clustering and fuzzy interpolation scheme to analyze network traffic and detect malicious network activity. Once malicious activity is detected, IoT-Keeper automatically enforces network access restrictions against IoT device generating this activity, and prevents it from attacking other devices or services. We have evaluated IoT-Keeper using a comprehensive dataset, collected from a real-world testbed, containing popular IoT devices. Using this dataset, our proposed technique achieved high accuracy (≈0.98) and low false positive rate (≈0.02) for detecting malicious network activity. Our evaluation also shows that IoT-Keeper has low resource footprint, and it can detect and mitigate various network attacks—without requiring explicit attack signatures or sophisticated hardware.

96 citations


Posted Content
TL;DR: This survey article provides a comprehensive introduction to edge intelligence and its application areas and presents a systematic classification of the state of the solutions by examining research results and observations for each of the four components.
Abstract: Edge intelligence refers to a set of connected systems and devices for data collection, caching, processing, and analysis in locations close to where data is captured based on artificial intelligence. The aim of edge intelligence is to enhance the quality and speed of data processing and protect the privacy and security of the data. Although recently emerged, spanning the period from 2011 to now, this field of research has shown explosive growth over the past five years. In this paper, we present a thorough and comprehensive survey on the literature surrounding edge intelligence. We first identify four fundamental components of edge intelligence, namely edge caching, edge training, edge inference, and edge offloading, based on theoretical and practical results pertaining to proposed and deployed systems. We then aim for a systematic classification of the state of the solutions by examining research results and observations for each of the four components and present a taxonomy that includes practical problems, adopted techniques, and application goals. For each category, we elaborate, compare and analyse the literature from the perspectives of adopted techniques, objectives, performance, advantages and drawbacks, etc. This survey article provides a comprehensive introduction to edge intelligence and its application areas. In addition, we summarise the development of the emerging research field and the current state-of-the-art and discuss the important open issues and possible theoretical and technical solutions.

65 citations


Journal ArticleDOI
TL;DR: This strategy generalizes the models of calibration and virtual sensing which then allows LCS to be deployed in field independently with a high accuracy, and enables scaling-up accurate air pollution mapping appropriate for smart cities.
Abstract: This paper presents the development of air quality low-cost sensors (LCS) with improved accuracy features. The LCS features integrate machine learning based calibration models and virtual sensors. LCS performances are analyzed and some LCS variables with low performance are improved through intelligent field-calibrations. Meteorological variables are calibrated using linear dynamic models.While, due to the non-linear relationship to reference instruments, fine particulate matter (PM2.5) are calibrated using non-linear machine learning models. However, due to sensor drifts or faults, carbon dioxide (CO2) does not present correlation to reference instrument. As a result, the LCS for CO2 is not feasible to be calibrated. Hence, to estimate the CO2 concentration,mathematicalmodels are developed to be integrated in the calibrated LCS, known as a virtual sensor. In addition, another virtual sensor is developed to demonstrate the capability of estimating air pollutant concentrations, e.g. black carbon, when the physical sensor devices are not available. In our paper, calibration models and virtual sensors are established using corresponding reference instruments that are installed on two reference stations. This strategy generalizes the models of calibration and virtual sensing which then allows LCS to be deployed in field independently with a high accuracy. Our proposed methodology enables scaling-up accurate air pollution mapping appropriate for smart cities.

63 citations


Posted Content
TL;DR: In this paper, the authors provide a vision for 6G Edge Intelligence and present edge computing along with other 6G enablers as a key component to establish the future intelligent Internet technologies as shown in this series of 6G White Papers.
Abstract: In this white paper we provide a vision for 6G Edge Intelligence. Moving towards 5G and beyond the future 6G networks, intelligent solutions utilizing data-driven machine learning and artificial intelligence become crucial for several real-world applications including but not limited to, more efficient manufacturing, novel personal smart device environments and experiences, urban computing and autonomous traffic settings. We present edge computing along with other 6G enablers as a key component to establish the future 2030 intelligent Internet technologies as shown in this series of 6G White Papers. In this white paper, we focus in the domains of edge computing infrastructure and platforms, data and edge network management, software development for edge, and real-time and distributed training of ML/AI algorithms, along with security, privacy, pricing, and end-user aspects. We discuss the key enablers and challenges and identify the key research questions for the development of the Intelligent Edge services. As a main outcome of this white paper, we envision a transition from Internet of Things to Intelligent Internet of Intelligent Things and provide a roadmap for development of 6G Intelligent Edge.

42 citations


Proceedings ArticleDOI
23 Mar 2020
TL;DR: This paper formalises the requirements of trustworthy AI systems through an ethics perspective, and focuses on the aspects that can be integrated into the design and development of AI systems.
Abstract: The era of pervasive computing has resulted in countless devices that continuously monitor users and their environment, generating an abundance of user behavioural data. Such data may support improving the quality of service, but may also lead to adverse usages such as surveillance and advertisement. In parallel, Artificial Intelligence (AI) systems are being applied to sensitive fields such as healthcare, justice, or human resources, raising multiple concerns on the trustworthiness of such systems. Trust in AI systems is thus intrinsically linked to ethics, including the ethics of algorithms, the ethics of data, or the ethics of practice. In this paper, we formalise the requirements of trustworthy AI systems through an ethics perspective. We specifically focus on the aspects that can be integrated into the design and development of AI systems. After discussing the state of research and the remaining challenges, we show how a concrete use-case in smart cities can benefit from these methods.

41 citations


Posted Content
26 Mar 2020
TL;DR: This survey article provides a comprehensive introduction to edge intelligence and its application areas and presents a systematic classification of the state of the solutions by examining research results and observations for each of the four components.
Abstract: Edge intelligence refers to a set of connected systems and devices for data collection, caching, processing, and analysis in locations close to where data is captured based on artificial intelligence. The aim of edge intelligence is to enhance the quality and speed of data processing and protect the privacy and security of the data. Although recently emerged, spanning the period from 2011 to now, this field of research has shown explosive growth over the past five years. In this paper, we present a thorough and comprehensive survey on the literature surrounding edge intelligence. We first identify four fundamental components of edge intelligence, namely edge caching, edge training, edge inference, and edge offloading, based on theoretical and practical results pertaining to proposed and deployed systems. We then aim for a systematic classification of the state of the solutions by examining research results and observations for each of the four components and present a taxonomy that includes practical problems, adopted techniques, and application goals. For each category, we elaborate, compare and analyse the literature from the perspectives of adopted techniques, objectives, performance, advantages and drawbacks, etc. This survey article provides a comprehensive introduction to edge intelligence and its application areas. In addition, we summarise the development of the emerging research field and the current state-of-the-art and discuss the important open issues and possible theoretical and technical solutions.

25 citations


Proceedings ArticleDOI
20 Apr 2020
TL;DR: This paper introduces an app usage collection platform named carat, and conducts the first study on the long-term evolution processes on a macro-level and micro-level of mobile app usage, finding that there is a growth stage enabled by the introduction of new technologies and a plateau stage caused by high correlations between app categories and a pareto effect in individual app usage.
Abstract: The prevalence of smartphones has promoted the popularity of mobile apps in recent years. Although significant effort has been made to understand mobile app usage, existing studies are based primarily on short-term datasets with limited time span, e.g., a few months. Therefore, many basic facts about the long-term evolution of mobile app usage are unknown. In this paper, we study how mobile app usage evolves over a long-term period. We first introduce an app usage collection platform named carat, from which we have gathered app usage records of 1,465 users from 2012 to 2017. We then conduct the first study on the long-term evolution processes on a macro-level, i.e., app-category, and micro-level, i.e., individual app. We discover that, on both levels, there is a growth stage enabled by the introduction of new technologies. Then there is a plateau stage caused by high correlations between app categories and a pareto effect in individual app usage, respectively. Additionally, the evolution of individual app usage undergoes an elimination stage due to fierce intra-category competition. Nevertheless, the diverseness of app-category and individual app usage exhibit opposing trends: app-category usage assimilates while individual app usage diversifies. Our study provides useful implications for app developers, market intermediaries, and service providers.

25 citations


Journal ArticleDOI
TL;DR: Matryoshka, the divide-and-conquer approach for finding a sequence of feasible configurations that can be used to reconfigure 5G network slices, is presented and it is observed that it yields close-to-optimal reconfiguration sequences 10X faster than existing approaches.
Abstract: The virtual resources of 5G networks are expected to scale and support migration to other locations within the substrate. In this context, a configuration for 5G network slices details the instantaneous mapping of the virtual resources across all slices on the substrate, and a feasible configuration satisfies the Service-Level Objectives (SLOs) without overloading the substrate. Reconfiguring a network from a given source configuration to the desired target configuration involves identifying an ordered sequence of feasible configurations from the source to the target. The proposed solutions for finding such a sequence are optimized for data centers and cannot be used as-is for reconfiguring 5G network slices. We present Matryoshka , our divide-and-conquer approach for finding a sequence of feasible configurations that can be used to reconfigure 5G network slices. Unlike previous approaches, Matryoshka also considers the bandwidth and latency constraints between the network functions of network slices. Evaluating Matryoshka required a dataset of pairs of source and target configurations. Because such a dataset is currently unavailable, we analyze proof of concept roll-outs, trends in standardization bodies, and research sources to compile an input dataset. On using Matryoshka on our dataset, we observe that it yields close-to-optimal reconfiguration sequences 10X faster than existing approaches.

23 citations


Journal ArticleDOI
TL;DR: A framework to discover daily cyber activity patterns across people's mobile app usage is proposed, which shows that people usually follow yesterday's activity patterns, but the patterns tend to deviate as the time-lapse increases.
Abstract: With the prevalence of smartphones, people have left abundant behavior records in cyberspace. Discovering and understanding individuals' cyber activities can provide useful implications for policymakers, service providers, and app developers. In this paper, we propose a framework to discover daily cyber activity patterns across people's mobile app usage. We first segment app usage traces into small time windows and then design a probabilistic topic model to infer users' cyber activities of each window. By exploring the coherence of users' activity sequences, the daily patterns of individuals are identified. Next, we recognize the common patterns across diverse groups of individuals using a hierarchical clustering algorithm. We then apply our framework on a large-scale and real-world dataset, consisting of 653,092 users with 971,818,946 usage records of 2,000 popular mobile apps. Our analysis shows that people usually obey yesterday's activity patterns, but the patterns tend to deviate as the time-lapse increases. We also discover five common daily cyber activity patterns, including afternoon reading, nightly entertainment, pervasive socializing, commuting, and nightly socializing. Our findings have profound implications on identifying the demographics of users and their lifestyles, habits, service requirements, and further detecting other disrupting trends such as working overtime and addiction to the game and social media.

20 citations


Journal ArticleDOI
19 Jun 2020-Sensors
TL;DR: It is argued that Software-defined Networking (SDN) together with the data generated by IoT applications can enhance the control and management of IoT in terms of flexibility and intelligence.
Abstract: The Internet of Things (IoT) connects smart devices to enable various intelligent services The deployment of IoT encounters several challenges, such as difficulties in controlling and managing IoT applications and networks, problems in programming existing IoT devices, long service provisioning time, underused resources, as well as complexity, isolation and scalability, among others One fundamental concern is that current IoT networks lack flexibility and intelligence A network-wide flexible control and management are missing in IoT networks In addition, huge numbers of devices and large amounts of data are involved in IoT, but none of them have been tuned for supporting network management and control In this paper, we argue that Software-defined Networking (SDN) together with the data generated by IoT applications can enhance the control and management of IoT in terms of flexibility and intelligence We present a review for the evolution of SDN and IoT and analyze the benefits and challenges brought by the integration of SDN and IoT with the help of IoT data We discuss the perspectives of knowledge-driven SDN for IoT through a new IoT architecture and illustrate how to realize Industry IoT by using the architecture We also highlight the challenges and future research works toward realizing IoT with the knowledge-driven SDN

17 citations


Journal ArticleDOI
TL;DR: In addition to the availability and accessibility of data and digital tools in pandemic responses, it is important to improve health equity by promoting access to connectivity and digital health literacy skills and to consider the interaction between digital health technologies and society, culture, and the economy.

Proceedings ArticleDOI
09 Nov 2020
TL;DR: The design and development of the MegaSense Cyber-Physical System (CPS) for spatially distributed IoT-based monitoring of urban air quality and its applications have applications in policy intervention management mechanisms and design of clean air routing and healthier navigation applications to reduce pollution exposure.
Abstract: Air pollution is a contributor to approximately one in every nine deaths annually. To counteract health issues resulting from air pollution, air quality monitoring is being carried out extensively in urban environments. Currently, however, city air quality monitoring stations are expensive to maintain, resulting in sparse coverage. In this paper, we introduce the design and development of the MegaSense Cyber-Physical System (CPS) for spatially distributed IoT-based monitoring of urban air quality. MegaSense is able to produce aggregated, privacy-aware maps and history graphs of collected pollution data. It provides a feedback loop in the form of personal outdoor and indoor air pollution exposure information, allowing citizens to take measures to avoid future exposure. We present a battery-powered, portable low-cost air quality sensor design for sampling PM 2.5 and air pollutant gases in different micro-environments. We validate the approach with a use case in Helsinki, deploying MegaSense with citizens carrying low-maintenance portable sensors, and using smart phone exposure apps. We demonstrate daily air pollution exposure profiles and the air pollution hot-spot profile of a district. Our contributions have applications in policy intervention management mechanisms and design of clean air routing and healthier navigation applications to reduce pollution exposure.

Proceedings ArticleDOI
01 Dec 2020
TL;DR: In this paper, a blockchain empowered chatbot for financial transactions, called BONIK, is presented, and the developed Proof-of-Concept (PoC) is evaluated.
Abstract: A Chatbot is a popular platform to enable users to interact with a software or website to gather information or execute actions in an automated fashion. In recent years, chatbots are being used for executing financial transactions, however, there are a number of security issues, such as secure authentication, data integrity, system availability and transparency, that must be carefully handled for their wide-scale adoption. Recently, the blockchain technology, with a number of security advantages, has emerged as one of the foundational technologies with the potential to disrupt a number of application domains, particularly in the financial sector. In this paper, we forward the idea of integrating a chatbot with blockchain technology in the view to improve the security issues in financial chatbots. More specifically, we present BONIK, a blockchain empowered chatbot for financial transactions, and discuss its architecture and design choices. Furthermore, we explore the developed Proof-of-Concept (PoC), evaluate its performance, analyse how different security and privacy issues are mitigated using BONIK.

Proceedings ArticleDOI
09 Nov 2020
TL;DR: In this paper, the authors present an air quality sensing process needed for low-cost sensors which are planned for long-term use, including design and production, laboratory tests, field tests, deployment, and maintenance.
Abstract: Air pollution is a main challenge in societies with particulate matter PM 2.5 as the major air pollutant causing serious health implications. Due to health and economic impacts of air pollution, low-cost and portable air quality sensors can be vastly deployed to gain personal air pollutant exposure. In this paper, we present an air quality sensing process needed for low-cost sensors which are planned for long-term use. The steps of this process include design and production, laboratory tests, field tests, deployment, and maintenance. As a case study we focus on the field test, where we use two generations of a portable air quality sensor (capable of measuring meteorological variables and PM 2.5 ) to perform an indoor-outdoor measurement. The study found that all of the measurements shown to be consistent through validation among themselves. The sensors accuracy also demonstrate to be adequate by showing similar readings compared to the nearest air quality reference station.

Proceedings ArticleDOI
19 Jun 2020
TL;DR: PENGUIN has been designed as the first system that can both recognize pollutants and classify them according to specifics of the material, and it is demonstrated that simple PPG sensors provide a low-cost and energy-efficient solution for classifying different plastics.
Abstract: Underwater plastic pollution is a significant global concern, affecting everything from marine ecosystems to climate change and even human health. Currently, obtaining accurate information about aquatic plastic pollutants at high spatial and temporal resolution is difficult as existing methods are laborious (e.g., dive surveys), restricted to a subset of plastics (e.g., aerial imaging for floating debris), have limited resolution (e.g., beach surveys), or are unsuited for aquatic environments (e.g., wireless sensing or Fourier-transform infrared spectroscopy). We propose PENGUIN, a work-in-progress AUV-based solution for identifying and classifying aquatic plastic pollutants. PENGUIN has been designed as the first system that can both recognize pollutants and classify them according to specifics of the material. We present the overall design of PENGUIN, introducing the different components of the architecture, and presenting current status of development. We also present results of plastic classification experiments using optical sensing, demonstrating that simple PPG sensors provide a low-cost and energy-efficient solution for classifying different plastics. Our solution can easily monitor larger underwater areas than what current techniques offer while at the same time capturing a wider range of pollutants.

Posted Content
TL;DR: This work sketches guidelines for a new AI diffusion method based on a decentralized online marketplace that considers the technical, economic, and regulatory aspects of such a marketplace including a discussion of solutions for problems in these areas.
Abstract: Artificial intelligence shows promise for solving many practical societal problems in areas such as healthcare and transportation. However, the current mechanisms for AI model diffusion such as Github code repositories, academic project webpages, and commercial AI marketplaces have some limitations; for example, a lack of monetization methods, model traceability, and model auditabilty. In this work, we sketch guidelines for a new AI diffusion method based on a decentralized online marketplace. We consider the technical, economic, and regulatory aspects of such a marketplace including a discussion of solutions for problems in these areas. Finally, we include a comparative analysis of several current AI marketplaces that are already available or in development. We find that most of these marketplaces are centralized commercial marketplaces with relatively few models.

Book ChapterDOI
18 Dec 2020
TL;DR: In this paper, the authors proposed a statistically sound, Bayesian inference federated learning for heart rate prediction with autoregression with exogenous variable (ARX) model, which achieves accurate and robust heart rate predictions.
Abstract: The advances of sensing and computing technologies pave the way to develop novel applications and services for wearable devices. For example, wearable devices measure heart rate, which accurately reflects the intensity of physical exercise. Therefore, heart rate prediction from wearable devices benefits users with optimization of the training process. Conventionally, Cloud collects user data from wearable devices and conducts inference. However, this paradigm introduces significant privacy concerns. Federated learning is an emerging paradigm that enhances user privacy by remaining the majority of personal data on users’ devices. In this paper, we propose a statistically sound, Bayesian inference federated learning for heart rate prediction with autoregression with exogenous variable (ARX) model. The proposed privacy-preserving method achieves accurate and robust heart rate prediction. To validate our method, we conduct extensive experiments with real-world outdoor running exercise data collected from wearable devices.

Journal ArticleDOI
TL;DR: A design of regulated efficient/bounded inefficient economic mechanisms for oligopoly data trading markets using a novel preference function bidding approach on a simplified sellers-broker market is proposed.
Abstract: In the modern era of the mobile apps (the era of surveillance capitalism - as termed by Shoshana Zuboff) huge quantities of surveillance data about consumers and their activities offer a wave of opportunities for economic and societal value creation. ln-app advertising - a multi-billion dollar industry, is an essential part of the current digital ecosystem driven by free mobile applications, where the ecosystem entities usually comprise consumer apps, their clients (consumers), ad-networks, and advertisers. Sensitive consumer information is often being sold downstream in this ecosystem without the knowledge of consumers, and in many cases to their annoyance. While this practice, in cases, may result in long-term benefits for the consumers, it can result in serious information privacy breaches of very significant impact (e.g., breach of genetic data) in the short term. The question we raise through this paper is: Is it economically feasible to trade consumer personal information with their formal consent (permission) and in return provide them incentives (monetary or otherwise)?. In view of (a) the behavioral assumption that humans are `compromising' beings and have privacy preferences, (b) privacy as a good not having strict boundaries, and (c) the practical inevitability of inappropriate data leakage by data holders downstream in the data-release supply-chain, we propose a design of regulated efficient/bounded inefficient economic mechanisms for oligopoly data trading markets using a novel preference function bidding approach on a simplified sellers-broker market. Our methodology preserves the heterogeneous privacy preservation constraints (at a grouped consumer, i.e., app, level) upto certain compromise levels, and at the same time satisfies information demand (via the broker) of agencies (e.g., advertising organizations) that collect client data for the purpose of targeted behavioral advertising.

Proceedings ArticleDOI
23 Mar 2020
TL;DR: Developing COSINE as a novel approach for selecting collaborators in multi-device computing scenarios that identifies and recommends collaborators based on a novel information theoretic measure based on Markov trajectory entropy shows significant improvement in collaboration benefits.
Abstract: Pervasive availability of programmable smart de-vices is giving rise to sensing and computing scenarios that involve collaboration between multiple devices. Maximizing the benefits of collaboration requires careful selection of devices with whom to collaborate as otherwise collaboration may be interrupted prematurely or be sub-optimal for the characteristics of the task at hand. Existing research on collaborative scenarios has mostly focused on providing mechanisms that can establish and harness collaboration, without considering how to maximally benefit from it. In this paper, we contribute by developing COSINE as a novel approach for selecting collaborators in multi-device computing scenarios. COSINE identifies and recommends collaborators based on a novel information theoretic measure based on Markov trajectory entropy. Rigorous experimental benchmarks carried out using a large-scale dataset of device-to-device encounters demonstrate that COSINE can significantly improve collaboration benefits compared to current state-of-the-art solutions, increasing expected duration of collaboration and reducing variability of collaborations.

Proceedings ArticleDOI
01 Nov 2020
TL;DR: Agora as mentioned in this paper is the first privacy-aware data marketplace that enables parties to get compensated for contributing data without relying on a trusted third party, leveraging cryptographic techniques to achieve three security properties: (i) data privacy, (ii) output verifiability, and (iii) atomicity of payments.
Abstract: We propose Agora, the first privacy-aware data marketplace that enables parties to get compensated for contributing data, without relying on a trusted third party. We leverage cryptographic techniques to achieve three security properties: (i) data privacy—raw data remain private except for a function output, (ii) output verifiability—the output is proven to be correct, and (iii) atomicity of payments—parties cannot avoid paying for provided services. Agora is designed as a decentralized blockchain application via smart contracts. We implement a prototype on Ethereum and evaluate its performance in terms of computation overhead and monetary cost.

Journal ArticleDOI
TL;DR: This study finds that IoT device mobility exhibits significantly different patterns compared with smartphones in multiple aspects, and finds the gap mobility predictability and predictability limit between IoT and human is not as big as people expected.
Abstract: Internet of Thing (IoT) devices are rapidly becoming an indispensable part of our life with their increasing deployment in many promising areas, including tele-health, smart city, intelligent agriculture. In this article, we aim to answer three research questions: (i) what are the mobility patterns of IoT device (ii) what are the differences between IoT device and smartphone mobility patterns (iii) how the IoT device mobility patterns differ among device types and usage scenarios We present a comprehensive characterization of IoT device mobility patterns from the perspective of cellular data networks, using a 36-days long signal trace, including 1.5 million IoT devices and 0.425 million smartphones, collected from a nation-wide cellular network in China. We first investigate the basic patterns of IoT devices from two perspectives: temporal and spatial characteristics. Our study finds that IoT device mobility exhibits significantly different patterns compared with smartphones in multiple aspects. For instance, IoT devices move more frequently and have larger radius of gyration. Then we explore the essential mobility of IoT devices by utilizing two models that reveal the nature of human mobility, i.e., exploration and preferential return (EPR) model and entropy based predictability model.

Posted Content
TL;DR: This article presents a research vision of large-scale autonomous marine pollution monitoring that uses coordinated groups of autonomous underwater vehicles (AUV)s to monitor extent and characteristics of marine pollutants and addresses the feasibility of this vision.
Abstract: Marine pollution is a growing worldwide concern, affecting health of marine ecosystems, human health, climate change, and weather patterns. To reduce underwater pollution, it is critical to have access to accurate information about the extent of marine pollutants as otherwise appropriate countermeasures and cleaning measures cannot be chosen. Currently such information is difficult to acquire as existing monitoring solutions are highly laborious or costly, limited to specific pollutants, and have limited spatial and temporal resolution. In this article, we present a research vision of large-scale autonomous marine pollution monitoring that uses coordinated groups of autonomous underwater vehicles (AUV)s to monitor extent and characteristics of marine pollutants. We highlight key requirements and reference technologies to establish a research roadmap for realizing this vision. We also address the feasibility of our vision, carrying out controlled experiments that address classification of pollutants and collaborative underwater processing, two key research challenges for our vision.

Book ChapterDOI
TL;DR: This chapter briefly introduces self-driving vehicles and gives an overview of validation frameworks for testing them in a simulated environment, and discusses what an ideal validation framework at the state of the art should be and what could benefit validated frameworks for self- driving vehicles in the future.
Abstract: As a part of the digital transformation, we interact with more and more intelligent gadgets. Today, these gadgets are often mobile devices, but in the advent of smart cities, more and more infrastructure—such as traffic and buildings—in our surroundings becomes intelligent. The intelligence, however, does not emerge by itself. Instead, we need both design techniques to create intelligent systems, as well as approaches to validate their correct behavior. An example of intelligent systems that could benefit smart cities are self-driving vehicles. Self-driving vehicles are continuously becoming both commercially available and common on roads. Accidents involving self-driving vehicles, however, have raised concerns about their reliability. Due to these concerns, the safety of self-driving vehicles should be thoroughly tested before they can be released into traffic. To ensure that self-driving vehicles encounter all possible scenarios, several millions of hours of testing must be carried out; therefore, testing self-driving vehicles in the real world is impractical. There is also the issue that testing self-driving vehicles directly in the traffic poses a potential safety hazard to human drivers. To tackle this challenge, validation frameworks for testing self-driving vehicles in simulated scenarios are being developed by academia and industry. In this chapter, we briefly introduce self-driving vehicles and give an overview of validation frameworks for testing them in a simulated environment. We conclude by discussing what an ideal validation framework at the state of the art should be and what could benefit validation frameworks for self-driving vehicles in the future.

Journal ArticleDOI
TL;DR: The multiple set matching problem is solved by proposing two efficient Bloom Multifilters called Bloom Matrix and Bloom Vector, which generalize the standard Bloom Filter and are much more space-efficient compared with the state-of-the-art, Bloofi.
Abstract: Bloom Filter is a space-efficient probabilistic data structure for checking the membership of elements in a set. Given multiple sets, a standard Bloom Filter is not sufficient when looking for the items to which an element or a set of input elements belong. An example case is searching for documents with keywords in a large text corpus, which is essentially a multiple set matching problem where the input is single or multiple keywords, and the result is a set of possible candidate documents. This article solves the multiple set matching problem by proposing two efficient Bloom Multifilters called Bloom Matrix and Bloom Vector, which generalize the standard Bloom Filter. Both structures are space-efficient and answer queries with a set of identifiers for multiple set matching problems. The space efficiency can be optimized according to the distribution of labels among multiple sets: Uniform and Zipf. Bloom Vector efficiently exploits the Zipf distribution of data for further space reduction. Indeed, both structures are much more space-efficient compared with the state-of-the-art, Bloofi. The results also highlight that a Lookup operation on Bloom Matrix is significantly faster than on Bloom Vector and Bloofi.

Proceedings ArticleDOI
27 May 2020
TL;DR: This paper systematically measure the performance of six popular mobile VoIP applications with controlled human conversation and acoustic setup and demonstrates that significant savings can indeed be achieved - with the best performing silence suppression technique being effective on 75% of silent pauses in the conversation in a quiet place.
Abstract: Human conversation is characterized by brief pauses and so-called turn-taking behavior between the speakers. In the context of VoIP, this means that there are frequent periods where the microphone captures only background noise - or even silence whenever the microphone is muted. The bits transmitted from such silence periods introduce overhead in terms of data usage, energy consumption, and network infrastructure costs. In this paper, we contribute by shedding light on these costs for VoIP applications. We systematically measure the performance of six popular mobile VoIP applications with controlled human conversation and acoustic setup. Our analysis demonstrates that significant savings can indeed be achieved - with the best performing silence suppression technique being effective on 75% of silent pauses in the conversation in a quiet place. This results in 2-5 times data savings, and 50-90% lower energy consumption compared to the next best alternative. Even then, the effectiveness of silence suppression can be sensitive to the amount of background noise, underlying speech codec, and the device being used. The codec characteristics and performance do not depend on the network type. However, silence suppression makes VoIP traffic network friendly as much as VoLTE traffic. Our results provide new insights into VoIP performance and offer a motivation for further enhancements to a wide variety of voice assisted applications, such as home assistants and other IoT devices.

Proceedings ArticleDOI
03 Mar 2020
TL;DR: This paper develops a new approach for facilitating the life-cycle management of large-scale sensor deployments through online estimation of battery health through so-called V-edge dynamics which capture and characterize instantaneous voltage drops.
Abstract: Deployments of battery-powered IoT devices have become ubiquitous, monitoring everything from environmental conditions in smart cities to wildlife movements in remote areas. How to manage the life-cycle of sensors in such large-scale deployments is currently an open issue. Indeed, most deployments let sensors operate until they fail and fix or replace the sensors post-hoc. In this paper, we contribute by developing a new approach for facilitating the life-cycle management of large-scale sensor deployments through online estimation of battery health. Our approach relies on so-called V-edge dynamics which capture and characterize instantaneous voltage drops. Experiments carried out on a dataset of battery discharge measurements demonstrate that our approach is capable of estimating battery health with up to $80%$ accuracy, depending on the characteristics of the devices and the processing load they undergo. Our method is particularly well-suited for the sensor devices, operating dedicated tasks, that they have constant discharge during their operation.

Proceedings ArticleDOI
10 Jun 2020
TL;DR: This paper investigates the utilization of different media input/output devices, e.g., camera, microphone, and speaker, by different types of multimedia applications, and introduces the notion of multimedia context.
Abstract: We use various multimedia applications on smart devices to consume multimedia content, to communicate with our peers, and to broadcast our events live. This paper investigates the utilization of different media input/output devices, e.g., camera, microphone, and speaker, by different types of multimedia applications, and introduces the notion of multimedia context. Our measurements lead to a sensing algorithm called MediaSense, which senses the states of multiple I/O devices and identifies eleven multimedia contexts of a mobile device in real time. The algorithm distinguishes stored content playback from streaming, live broadcasting from local recording, and conversational multimedia sessions from GSM/VoLTE calls on mobile devices.

Journal ArticleDOI
08 Jun 2020
TL;DR: This article shows somewhat surprisingly that, following a cyber-attack, the effect of a network interconnection topology and a wide range of loss distributions on the probability of a Cyber-blackout and the increase in total service-related monetary losses across all organizations are mostly very small.
Abstract: Service liability interconnections among globally networked IT- and IoT-driven service organizations create potential channels for cascading service disruptions worth billions of dollars, due to modern cyber-crimes such as DDoS, APT, and ransomware attacks. A natural question that arises in this context is: What is the likelihood of a cyber-blackout?, where the latter term is defined as the probability that all (or a major subset of) organizations in a service chain become dysfunctional in a certain manner due to a cyber-attack at some or all points in the chain. The answer to this question has major implications to risk management businesses such as cyber-insurance when it comes to designing policies by risk-averse insurers for providing coverage to clients in the aftermath of such catastrophic network events. In this article, we investigate this question in general as a function of service chain networks and different cyber-loss distribution types. We show somewhat surprisingly (and discuss the potential practical implications) that, following a cyber-attack, the effect of (a) a network interconnection topology and (b) a wide range of loss distributions on the probability of a cyber-blackout and the increase in total service-related monetary losses across all organizations are mostly very small. The primary rationale behind these results are attributed to degrees of heterogeneity in the revenue base among organizations and the Increasing Failure Rate property of popular (i.i.d/non-i.i.d) loss distributions, i.e., log-concave cyber-loss distributions. The result will enable risk-averse cyber-risk managers to safely infer the impact of cyber-attacks in a worst-case network and distribution oblivious setting.

Posted Content
TL;DR: Agora, the first privacy-aware data marketplace that enables parties to get compensated for contributing data, without relying on a trusted third party is proposed, designed as a decentralized blockchain application via smart contracts.
Abstract: We propose Agora, the first blockchain-based data marketplace that enables multiple privacy-concerned parties to get compensated for contributing and exchanging data, without relying on a trusted third party during the exchange. Agora achieves data privacy, output verifiability, and atomicity of payments by leveraging cryptographic techniques, and is designed as a decentralized application via smart contracts. Particularly, data generators provide encrypted data to data brokers who use a functional secret key to learn nothing but the output of a specific, agreed upon, function over the raw data. Data consumers can purchase decrypted outputs from the brokers, accompanied by corresponding proofs of correctness. We implement a working prototype of Agora on Ethereum and experimentally evaluate its performance and deployment costs. As a core building block of Agora, we propose a new functional encryption scheme with additional public parameters that operate as a trust anchor for verifying decrypted results.

Posted Content
TL;DR: This study demonstrates that mobile systems employ energy-aware system hardware tuning, which affects application performance and network throughput, and shows that the VPN-based application performance measurement tools, such as Lumen, PrivacyGuard, and Video Optimizer, aid in ambiguous network performance measurements and degrade the application performance.
Abstract: Understanding network and application performance are essential for debugging, improving user experience, and performance comparison. Meanwhile, modern mobile systems are optimized for energy-efficient computation and communications that may limit the performance of network and applications. In recent years, several tools have emerged that analyze network performance of mobile applications in~situ with the help of the VPN service. There is a limited understanding of how these measurement tools and system optimizations affect the network and application performance. In this study, we first demonstrate that mobile systems employ energy-aware system hardware tuning, which affects application performance and network throughput. We next show that the VPN-based application performance measurement tools, such as Lumen, PrivacyGuard, and Video Optimizer, aid in ambiguous network performance measurements and degrade the application performance. Our findings suggest that sound application and network performance measurement on Android devices requires a good understanding of the device, networks, measurement tools, and applications.