scispace - formally typeset
Search or ask a question
Author

Alessandra Mileo

Bio: Alessandra Mileo is an academic researcher from Dublin City University. The author has contributed to research in topics: Answer set programming & Logic programming. The author has an hindex of 18, co-authored 71 publications receiving 1182 citations. Previous affiliations of Alessandra Mileo include University of Milan & National University of Ireland, Galway.


Papers
More filters
Journal ArticleDOI
TL;DR: The CityPulse framework supports smart city service creation by means of a distributed system for semantic discovery, data analytics, and interpretation of large-scale (near-)real-time Internet of Things data and social media data streams to break away from silo applications and enable cross-domain data integration.
Abstract: Our world and our lives are changing in many ways. Communication, networking, and computing technologies are among the most influential enablers that shape our lives today. Digital data and connected worlds of physical objects, people, and devices are rapidly changing the way we work, travel, socialize, and interact with our surroundings, and they have a profound impact on different domains, such as healthcare, environmental monitoring, urban systems, and control and management applications, among several other areas. Cities currently face an increasing demand for providing services that can have an impact on people’s everyday lives. The CityPulse framework supports smart city service creation by means of a distributed system for semantic discovery, data analytics, and interpretation of large-scale (near-)real-time Internet of Things data and social media data streams. To goal is to break away from silo applications and enable cross-domain data integration. The CityPulse framework integrates multimodal, mixed quality, uncertain and incomplete data to create reliable, dependable information and continuously adapts data processing techniques to meet the quality of information requirements from end users. Different than existing solutions that mainly offer unified views of the data, the CityPulse framework is also equipped with powerful data analytics modules that perform intelligent data aggregation, event detection, quality assessment, contextual filtering, and decision support. This paper presents the framework, describes its components, and demonstrates how they interact to support easy development of custom-made applications for citizens. The benefits and the effectiveness of the framework are demonstrated in a use-case scenario implementation presented in this paper.

199 citations

Book ChapterDOI
11 Oct 2015
TL;DR: Performance, correctness and technical soundness of few existing RSP engines have been evaluated in controlled settings using existing benchmarks like LSBench and SRBench, but these benchmarks focus merely on features of the RSP query languages and engines and do not consider dynamic application requirements and data-dependent properties.
Abstract: With the growing popularity of Internet of Things (IoT) and IoT-enabled smart city applications, RDF stream processing (RSP) is gaining increasing attention in the Semantic Web community. As a result, several RSP engines have emerged, which are capable of processing semantically annotated data streams on the fly. Performance, correctness and technical soundness of few existing RSP engines have been evaluated in controlled settings using existing benchmarks like LSBench and SRBench. However, these benchmarks focus merely on features of the RSP query languages and engines, and do not consider dynamic application requirements and data-dependent properties such as changes in streaming rate during query execution or changes in application requirements over a period of time. This hinders wide adoption of RSP engines for real-time applications where data properties and application requirements play a key role and need to be characterised in their dynamic setting, such as in the smart city domain.

145 citations

01 Jan 2014
TL;DR: Examples of data that can be collected from cities are presented, issues around this data are discussed, and some preliminary thoughts for creating a semantic description model to describe and help discover, index and query smart city data are put forward.
Abstract: Recent advancements in communication technologies for providing ubiquitous Internet access as well as advancements on reduction of cost and form-factor of mobile devices and sensors are seen as an enabler for the Internet of Things (IoT). The industry predicts an interconnected world of 50 billion devices by 20201. The Web of Things (WoT) relies on the connectivity service of IoT to create services and applications exploiting the IoT data [1]. Cities present an opportunity for rendering WoT-enabled services. According to the World Health Organization, population in cities will double by the middle of this century2, while cities deal with increasingly pressing issues such as environmental sustainability, economic growth and citizen mobility. In this paper, we propose a discussion around the need for common semantic descriptions for smart city data to facilitate future services in “smart cities”. We present examples of data that can be collected from cities, discuss issues around this data and put forward some preliminary thoughts for creating a semantic description model to describe and help discover, index and query smart city data.

83 citations

Proceedings ArticleDOI
24 Feb 2014
TL;DR: This work uses an existing Linked Data knowledge-base to find pre-existing relations between entities in Wikipedia tables, suggesting the same relations as holding for other entities in analogous columns on different rows, and extracts RDF triples from Wikipedia's tables at a raw precision of 40%.
Abstract: The tables embedded in Wikipedia articles contain rich, semi-structured encyclopaedic content. However, the cumulative content of these tables cannot be queried against. We thus propose methods to recover the semantics of Wikipedia tables and, in particular, to extract facts from them in the form of RDF triples. Our core method uses an existing Linked Data knowledge-base to find pre-existing relations between entities in Wikipedia tables, suggesting the same relations as holding for other entities in analogous columns on different rows. We find that such an approach extracts RDF triples from Wikipedia's tables at a raw precision of 40%. To improve the raw precision, we define a set of features for extracted triples that are tracked during the extraction phase. Using a manually labelled gold standard, we then test a variety of machine learning methods for classifying correct/incorrect triples. One such method extracts 7.9 million unique and novel RDF triples from over one million Wikipedia tables at an estimated precision of 81.5%.

81 citations

Journal ArticleDOI
TL;DR: This work proposes a novel framework with an efficient semantic data processing pipeline, allowing for real-time observation of the pulse of a city and investigates the optimization of the semantic data discovery and integration based on the proposed stream quality analysis and data aggregation techniques.
Abstract: An increasing number of cities are confronted with challenges resulting from the rapid urbanization and new demands that a rapidly growing digital economy imposes on current applications and information systems. Smart city applications enable city authorities to monitor, manage, and provide plans for public resources and infrastructures in city environments, while offering citizens and businesses to develop and use intelligent services in cities. However, providing such smart city applications gives rise to several issues, such as semantic heterogeneity and trustworthiness of data sources, and extracting up-to-date information in real time from large-scale dynamic data streams. In order to address these issues, we propose a novel framework with an efficient semantic data processing pipeline, allowing for real-time observation of the pulse of a city. The proposed framework enables efficient semantic integration of data streams, and complex event processing on top of real-time data aggregation and quality analysis in a semantic Web environment. To evaluate our system, we use real-time sensor observations that have been published via an open platform called Open Data Aarhus by the City of Aarhus. We examine the framework utilizing symbolic aggregate approximation to reduce the size of data streams, and perform quality analysis taking into account both single and multiple data streams. We also investigate the optimization of the semantic data discovery and integration based on the proposed stream quality analysis and data aggregation techniques.

55 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Abstract: Machine Learning is the study of methods for programming computers to learn. Computers are applied to a wide range of tasks, and for most of these it is relatively easy for programmers to design and implement the necessary software. However, there are many tasks for which this is difficult or impossible. These can be divided into four general categories. First, there are problems for which there exist no human experts. For example, in modern automated manufacturing facilities, there is a need to predict machine failures before they occur by analyzing sensor readings. Because the machines are new, there are no human experts who can be interviewed by a programmer to provide the knowledge necessary to build a computer system. A machine learning system can study recorded data and subsequent machine failures and learn prediction rules. Second, there are problems where human experts exist, but where they are unable to explain their expertise. This is the case in many perceptual tasks, such as speech recognition, hand-writing recognition, and natural language understanding. Virtually all humans exhibit expert-level abilities on these tasks, but none of them can describe the detailed steps that they follow as they perform them. Fortunately, humans can provide machines with examples of the inputs and correct outputs for these tasks, so machine learning algorithms can learn to map the inputs to the outputs. Third, there are problems where phenomena are changing rapidly. In finance, for example, people would like to predict the future behavior of the stock market, of consumer purchases, or of exchange rates. These behaviors change frequently, so that even if a programmer could construct a good predictive computer program, it would need to be rewritten frequently. A learning program can relieve the programmer of this burden by constantly modifying and tuning a set of learned prediction rules. Fourth, there are applications that need to be customized for each computer user separately. Consider, for example, a program to filter unwanted electronic mail messages. Different users will need different filters. It is unreasonable to expect each user to program his or her own rules, and it is infeasible to provide every user with a software engineer to keep the rules up-to-date. A machine learning system can learn which mail messages the user rejects and maintain the filtering rules automatically. Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis. Statistics focuses on understanding the phenomena that have generated the data, often with the goal of testing different hypotheses about those phenomena. Data mining seeks to find patterns in the data that are understandable by people. Psychological studies of human learning aspire to understand the mechanisms underlying the various learning behaviors exhibited by people (concept learning, skill acquisition, strategy change, etc.).

13,246 citations

21 Jun 2010

1,966 citations

Journal ArticleDOI
TL;DR: A survey of such knowledge graph refinement approaches, with a dual look at both the methods being proposed as well as the evaluation methodologies used.
Abstract: In the recent years, different Web knowledge graphs, both free and commercial, have been created. While Google coined the term "Knowledge Graph" in 2012, there are also a few openly available knowledge graphs, with DBpedia, YAGO, and Freebase being among the most prominent ones. Those graphs are often constructed from semi-structured knowledge, such as Wikipedia, or harvested from the web with a combination of statistical and linguistic methods. The result are large-scale knowledge graphs that try to make a good trade-off between completeness and correctness. In order to further increase the utility of such knowledge graphs, various refinement methods have been proposed, which try to infer and add missing knowledge to the graph, or identify erroneous pieces of information. In this article, we provide a survey of such knowledge graph refinement approaches, with a dual look at both the methods being proposed as well as the evaluation methodologies used.

915 citations

Journal ArticleDOI
TL;DR: A combined IoT-based system for smart city development and urban planning using Big Data analytics, consisting of various types of sensor deployment, including smart home sensors, vehicular networking, weather and water sensors, smart parking sensors, and surveillance objects is proposed.

701 citations