scispace - formally typeset
Search or ask a question

Showing papers on "Ontology-based data integration published in 2005"


Book ChapterDOI
TL;DR: This paper presents a new classification of schema-based matching techniques that builds on the top of state of the art in both schema and ontology matching and distinguishes between approximate and exact techniques at schema-level; and syntactic, semantic, and external techniques at element- and structure-level.
Abstract: Schema and ontology matching is a critical problem in many application domains, such as semantic web, schema/ontology integration, data warehouses, e-commerce, etc. Many different matching solutions have been proposed so far. In this paper we present a new classification of schema-based matching techniques that builds on the top of state of the art in both schema and ontology matching. Some innovations are in introducing new criteria which are based on (i) general properties of matching techniques, (ii) interpretation of input information, and (iii) the kind of input information. In particular, we distinguish between approximate and exact techniques at schema-level; and syntactic, semantic, and external techniques at element- and structure-level. Based on the classification proposed we overview some of the recent schema/ontology matching systems pointing which part of the solution space they cover. The proposed classification provides a common conceptual basis, and, hence, can be used for comparing different existing schema/ontology matching techniques and systems as well as for designing new ones, taking advantages of state of the art solutions.

1,285 citations


Proceedings Article
01 Jan 2005
TL;DR: This article comprehensively reviews and provides insights on the pragmatics of ontology mapping and elaborate on a theoretical approach for defining ontology mapped.
Abstract: Ontology mapping is seen as a solution provider in today's landscape of ontology research. As the number of ontologies that are made publicly available and accessible on the Web increases steadily, so does the need for applications to use them. A single ontology is no longer enough to support the tasks envisaged by a distributed environment like the Semantic Web. Multiple ontologies need to be accessed from several applications. Mapping could provide a common layer from which several ontologies could be accessed and hence could exchange information in semantically sound manners. Developing such mapping has beeb the focus of a variety of works originating from diverse communities over a number of years. In this article we comprehensively review and present these works. We also provide insights on the pragmatics of ontology mapping and elaborate on a theoretical approach for defining ontology mapping.

748 citations


01 Jan 2005
TL;DR: A survey of the state of the art in ontology evaluation is presented, typically in order to determine which of several ontologies would best suit a particular purpose.
Abstract: An ontology is an explicit formal conceptualization of some domain of interest Ontologies are increasingly used in various fields such as knowledge management, information extraction, and the semantic web Ontology evaluation is the problem of assessing a given ontology from the point of view of a particular criterion of application, typically in order to determine which of several ontologies would best suit a particular purpose This paper presents a survey of the state of the art in ontology evaluation

641 citations


Book ChapterDOI
06 Nov 2005
TL;DR: In this article, the authors present a framework for introducing design patterns that facilitate or improve the techniques used during ontology lifecycle, and some distinctions are drawn between kinds of ontology design patterns.
Abstract: The paper presents a framework for introducing design patterns that facilitate or improve the techniques used during ontology lifecycle. Some distinctions are drawn between kinds of ontology design patterns. Some content-oriented patterns are presented in order to illustrate their utility at different degrees of abstraction, and how they can be specialized or composed. The proposed framework and the initial set of patterns are designed in order to function as a pipeline connecting domain modelling, user requirements, and ontology-driven tasks/queries to be executed.

502 citations


Journal ArticleDOI
TL;DR: This article first reviews database applications that require semantic integration and discusses the difficulties underlying the integration process, then describes recent progress and identifies open research issues.
Abstract: Semantic integration has been a long-standing challenge for the database community. It has received steady attention over the past two decades, and has now become a prominent area of database research. In this article, we first review database applications that require semantic integration and discuss the difficulties underlying the integration process. We then describe recent progress and identify open research issues. We focus in particular on schema matching, a topic that has received much attention in the database community, but also discuss data matching (for example, tuple deduplication) and open issues beyond the match discovery context (for example, reasoning with matches, match verification and repair, and reconciling inconsistent data values). For previous surveys of database research on semantic integration, see Rahm and Bernstein (2001); Ouksel and Seth (1999); and Batini, Lenzerini, and Navathe (1986).

499 citations


Book
01 Jul 2005
TL;DR: This volume presents current research in ontology learning, addressing three perspectives, including methodologies that have been proposed to automatically extract information from texts and to give a structured organization to such knowledge, including approaches based on machine learning techniques.
Abstract: This volume brings together ontology learning, knowledge acquisition and other related topics It presents current research in ontology learning, addressing three perspectives The first perspective looks at methodologies that have been proposed to automatically extract information from texts and to give a structured organization to such knowledge, including approaches based on machine learning techniques Then there are evaluation methods for ontology learning, aiming at defining procedures and metrics for a quantitative evaluation of the ontology learning task; and finally application scenarios that make ontology learning a challenging area in the context of real applications such as bio-informatics According to the three perspectives mentioned above, the book is divided into three sections, each including a selection of papers addressing respectively the methods, the applications and the evaluation of ontology learning approaches

488 citations


Journal ArticleDOI
TL;DR: An ontology for cell types that covers the prokaryotic, fungal, animal and plant worlds and is designed to be used in the context of model organism genome and other biological databases.
Abstract: We describe an ontology for cell types that covers the prokaryotic, fungal, animal and plant worlds. It includes over 680 cell types. These cell types are classified under several generic categories and are organized as a directed acyclic graph. The ontology is available in the formats adopted by the Open Biological Ontologies umbrella and is designed to be used in the context of model organism genome and other biological databases. The ontology is freely available at http://obo.sourceforge.net/ and can be viewed using standard ontology visualization tools such as OBO-Edit and COBrA.

437 citations


Journal ArticleDOI
01 Oct 2005
TL;DR: The experimental results show that the news agent based on the fuzzy ontology can effectively operate for news summarization and an experimental website is constructed to test the approach.
Abstract: In this paper, a fuzzy ontology and its application to news summarization are presented. The fuzzy ontology with fuzzy concepts is an extension of the domain ontology with crisp concepts. It is more suitable to describe the domain knowledge than domain ontology for solving the uncertainty reasoning problems. First, the domain ontology with various events of news is predefined by domain experts. The document preprocessing mechanism will generate the meaningful terms based on the news corpus and the Chinese news dictionary defined by the domain expert. Then, the meaningful terms will be classified according to the events of the news by the term classifier. The fuzzy inference mechanism will generate the membership degrees for each fuzzy concept of the fuzzy ontology. Every fuzzy concept has a set of membership degrees associated with various events of the domain ontology. In addition, a news agent based on the fuzzy ontology is also developed for news summarization. The news agent contains five modules, including a retrieval agent, a document preprocessing mechanism, a sentence path extractor, a sentence generator, and a sentence filter to perform news summarization. Furthermore, we construct an experimental website to test the proposed approach. The experimental results show that the news agent based on the fuzzy ontology can effectively operate for news summarization.

377 citations


Book ChapterDOI
24 Jul 2005
TL;DR: Gumo as mentioned in this paper is a general user model ontology for the uniform interpretation of distributed user models in intelligent semantic web enriched environments, and it supports ubiquitous applications with the u2m.org user model service.
Abstract: We introduce the general user model ontology Gumo for the uniform interpretation of distributed user models in intelligent semantic web enriched environments. We discuss design decisions, show the relation to the user model markup language UserML and present the integration of ubiquitous applications with the u2m.org user model service.

370 citations


Journal ArticleDOI
01 Oct 2005
TL;DR: An initial validation of the Ontology Auditor on the DARPA Agent Markup Language (DAML) library of domain ontologies indicates that the metrics are feasible and highlights the wide variation in quality among ontologies in the library.
Abstract: A suite of metrics is proposed to assess the quality of an ontology. Drawing upon semiotic theory, the metrics assess the syntactic, semantic, pragmatic, and social aspects of ontology quality. We operationalize the metrics and implement them in a prototype tool called the Ontology Auditor. An initial validation of the Ontology Auditor on the DARPA Agent Markup Language (DAML) library of domain ontologies indicates that the metrics are feasible and highlights the wide variation in quality among ontologies in the library. The contribution of the research is to provide a theory-based framework that developers can use to develop high quality ontologies and that applications can use to choose appropriate ontologies for a given task.

330 citations


Proceedings ArticleDOI
10 May 2005
TL;DR: A number of simple debugging cues generated from the description logic reasoner, Pellet, are integrated in the hypertextual ontology development environment, Swoop, to significantly improve the OWL debugging experience, and point the way to more general improvements in the presentation of an ontology to new users.
Abstract: As an increasingly large number of OWL ontologies become available on the Semantic Web and the descriptions in the ontologies become more complicated, finding the cause of errors becomes an extremely hard task even for experts. Existing ontology development environments provide some limited support, in conjunction with a reasoner, for detecting and diagnosing errors in OWL ontologies. Typically these are restricted to the mere detection of, for example, unsatisfiable concepts. We have integrated a number of simple debugging cues generated from our description logic reasoner, Pellet, in our hypertextual ontology development environment, Swoop. These cues, in conjunction with extensive undo/redo and Annotea based collaboration support in Swoop, significantly improve the OWL debugging experience, and point the way to more general improvements in the presentation of an ontology to new users.

Book ChapterDOI
29 May 2005
TL;DR: A model for the semantics of change for OWL ontologies, considering structural, logical, and user-defined consistency is presented, and resolution strategies to ensure that consistency is maintained as the ontology evolves are introduced.
Abstract: Support for ontology evolution is extremely important in ontology engineering and application of ontologies in dynamic environments. A core aspect in the evolution process is the to guarantee consistency of the ontology when changes occur. In this paper we discuss the consistent evolution of OWL ontologies. We present a model for the semantics of change for OWL ontologies, considering structural, logical, and user-defined consistency. We introduce resolution strategies to ensure that consistency is maintained as the ontology evolves.

Journal ArticleDOI
TL;DR: A number of debugging cues generated from the authors' reasoner, Pellet, are integrated in their hypertextual ontology development environment, Swoop, and it is demonstrated that these debugging cues significantly improve the OWL debugging experience, and point the way to more general improvements in the presentation of an ontology to users.

Proceedings Article
01 Jan 2005
TL;DR: The approach to building OntoSensor is described, a prototype sensor knowledge repository compatible with evolving Semantic Web infrastructure that includes definitions of concepts and properties adopted in part from SensorML, extensions to IEEE SUMO and references to ISO 19115.
Abstract: - This paper describes the approach to building OntoSensor: a prototype sensor knowledge repository compatible with evolving Semantic Web infrastructure. OntoSensor includes definitions of concepts and properties adopted in part from SensorML, extensions to IEEE SUMO and references to ISO 19115. Simple queries have been developed and tested using Protege 2000 and Prolog. Although OntoSensor is in the early development stage, it presents a practical approach to building a sensor knowledge repository. It is proposed that OntoSensor may serve as a component in comprehensive applications that include more advanced inference mechanisms. Such comprehensive applications will be used for synergistic fusion of heterogeneous data in a network-centric environment. Keywords: Data Fusion, Semantic Web, Sensor Ontology 1.0 Introduction The assessment of situations and strategies in complex environments requires the fusion of information from heterogeneous data sources including sensors [1]. Synergistic fusion of data from multiple sensors ranging in complexity from simple acoustic sensors to sophisticated imaging equipment such as Forward Looking Infrared (FLIR) sensors will lead to the extraction of knowledge that cannot be perceived or inferred using individual sensors alone. For example, information from diverse sources such as sensors on an Unmanned Aerial Vehicle (UAV), reconnaissance reports and satellite imagery can be integrated to obtain high-level knowledge of objects in an area under surveillance including their spatial and temporal interrelationships; and to generate predictions of their intentions, positions and alignments in future states.

Book ChapterDOI
TL;DR: OntoMerge, an online system for ontology merging and automated reasoning, can implement ontology translation with inputs and outputs in OWL or other web languages.
Abstract: Ontologies are a crucial tool for formally specifying the vocabulary and relationship of concepts used on the Semantic Web. In order to share information, agents that use different vocabularies must be able to translate data from one ontological framework to another. Ontology translation is required when translating datasets, generating ontology extensions, and querying through different ontologies. OntoMerge, an online system for ontology merging and automated reasoning, can implement ontology translation with inputs and outputs in OWL or other web languages. Ontology translation can be thought of in terms of formal inference in a merged ontology. The merge of two related ontologies is obtained by taking the union of the concepts and the axioms defining them, and then adding bridging axioms that relate their concepts. The resulting merged ontology then serves as an inferential medium within which translation can occur. Our internal representation, Web-PDDL, is a strong typed first-order logic language for web application. Using a uniform notation for all problems allows us to factor out syntactic and semantic translation problems, and focus on the latter. Syntactic translation is done by an automatic translator between Web-PDDL and OWL or other web languages. Semantic translation is implemented using an inference engine (OntoEngine) which processes assertions and queries in Web-PDDL syntax, running in either a data-driven (forward chaining) or demand-driven (backward chaining) way.

Book ChapterDOI
01 Jan 2005
TL;DR: This paper presents how to build an ontology in the legal domain following the ontology development methodology METHONTOLOGY and using the ontological engineering workbench WebODE.
Abstract: This paper presents how to build an ontology in the legal domain following the ontology development methodology METHONTOLOGY and using the ontology engineering workbench WebODE. Both of them have been widely used to develop ontologies in many other domains. The ontology used to illustrate this paper has been extracted from an existing class taxonomy proposed by Breuker, and adapted to the Spanish legal domain.

Book ChapterDOI
31 Oct 2005
TL;DR: The NRL Security Ontology is more comprehensive and better organized than existing security ontologies, capable of representing more types of security statements and can be applied to any electronic resource.
Abstract: Annotation with security-related metadata enables discovery of resources that meet security requirements. This paper presents the NRL Security Ontology, which complements existing ontologies in other domains that focus on annotation of functional aspects of resources. Types of security information that could be described include mechanisms, protocols, objectives, algorithms, and credentials in various levels of detail and specificity. The NRL Security Ontology is more comprehensive and better organized than existing security ontologies. It is capable of representing more types of security statements and can be applied to any electronic resource. The class hierarchy of the ontology makes it both easy to use and intuitive to extend. We applied this ontology to a Service Oriented Architecture to annotate security aspects of Web service descriptions and queries. A refined matching algorithm was developed to perform requirement-capability matchmaking that takes into account not only the ontology concepts, but also the properties of the concepts.

Journal Article
TL;DR: The intention of this essay is to give an overview of different methods that learn ontologies or ontology-like structures from unstructured text.
Abstract: After the vision of the Semantic Web was broadcasted at the turn of the millennium, ontology became a synonym for the solution to many problems concerning the fact that computers do not understand human language: if there were an ontology and every document were marked up with it and we had agents that would understand the markup, then computers would finally be able to process our queries in a really sophisticated way. Some years later, the success of Google shows us that the vision has not come true, being hampered by the incredible amount of extra work required for the intellectual encoding of semantic mark-up – as compared to simply uploading an HTML page. To alleviate this acquisition bottleneck, the field of ontology learning has since emerged as an important sub-field of ontology engineering. It is widely accepted that ontologies can facilitate text understanding and automatic processing of textual resources. Moving from words to concepts not only mitigates data sparseness issues, but also promises appealing solutions to polysemy and homonymy by finding non-ambiguous concepts that may map to various realizations in – possibly ambiguous – words. Numerous applications using lexical-semantic databases like WordNet (Miller, 1990) and its non-English counterparts, e.g. EuroWordNet (Vossen, 1997) or CoreNet (Choi and Bae, 2004) demonstrate the utility of semantic resources for natural language processing. Learning semantic resources from text instead of manually creating them might be dangerous in terms of correctness, but has undeniable advantages: Creating resources for text processing from the texts to be processed will fit the semantic component neatly and directly to them, which will never be possible with general-purpose resources. Further, the cost per entry is greatly reduced, giving rise to much larger resources than an advocate of a manual approach could ever afford. On the other hand, none of the methods used today are good enough for creating semantic resources of any kind in a completely unsupervised fashion, albeit automatic methods can facilitate manual construction to a large extent. The term ontology is understood in a variety of ways and has been used in philosophy for many centuries. In contrast, the notion of ontology in the field of computer science is younger – but almost used as inconsistently, when it comes to the details of the definition. The intention of this essay is to give an overview of different methods that learn ontologies or ontology-like structures from unstructured text. Ontology learning from other sources, issues in description languages, ontology editors, ontology merging and ontology evolving transcend the scope of this article. Surveys on ontology learning from text and other sources can be found in Ding and Foo (2002) and Gomez-Perez

Book ChapterDOI
06 Nov 2005
TL;DR: Omen, an Ontology Mapping ENhancer, is based on a set of meta-rules that captures the influence of the ontology structure and the existing matches to match nodes that are neighbours to matched nodes in the two ontologies.
Abstract: Most existing ontology mapping tools are inexact. Inexact ontology mapping rules, if not rectified, result in imprecision in the applications that use them. We describe a framework to probabilistically improve existing ontology mappings using a Bayesian Network. Omen, an Ontology Mapping ENhancer, is based on a set of meta-rules that captures the influence of the ontology structure and the existing matches to match nodes that are neighbours to matched nodes in the two ontologies. We have implemented a protype ontology matcher that can either map concepts across two input ontologies or enhance existing matches between ontology concepts. Preliminary experiments demonstrate that Omen enhances existing ontology mappings in our test cases.

Journal ArticleDOI
TL;DR: The cohesion metrics examine the fundamental quality of cohesion as it relates to ontologies in order to effectively make use of domain specific ontology development.
Abstract: Recently, domain specific ontology development has been driven by research on the Semantic Web. Ontologies have been suggested for use in many application areas targeted by the Semantic Web, such as dynamic web service composition and general web service matching. Fundamental characteristics of these ontologies must be determined in order to effectively make use of them: for example, Sirin, Hendler and Parsia have suggested that determining fundamental characteristics of ontologies is important for dynamic web service composition. Our research examines cohesion metrics for ontologies. The cohesion metrics examine the fundamental quality of cohesion as it relates to ontologies.

Book ChapterDOI
06 Nov 2005
TL;DR: In this paper, the source and target ontologies are first translated into Bayesian networks (BN) and the concept mapping between the two ontologies is treated as evidential reasoning between the translated BNs.
Abstract: This paper presents our ongoing effort on developing a principled methodology for automatic ontology mapping based on BayesOWL, a probabilistic framework we developed for modeling uncertainty in semantic web. In this approach, the source and target ontologies are first translated into Bayesian networks (BN); the concept mapping between the two ontologies are treated as evidential reasoning between the two translated BNs. Probabilities needed for constructing conditional probability tables (CPT) during translation and for measuring semantic similarity during mapping are learned using text classification techniques where each concept in an ontology is associated with a set of semantically relevant text documents, which are obtained by ontology guided web mining. The basic ideas of this approach are validated by positive results from computer experiments on two small real-world ontologies.

Proceedings ArticleDOI
02 Oct 2005
TL;DR: The results show that AKTiveRank will have great utility although there is potential for improvement, and a number of metrics are applied in an attempt to investigate their appropriateness for ranking ontologies.
Abstract: In view of the need to provide tools to facilitate the re-use of existing knowledge structures such as ontologies, we present in this paper a system, AKTiveRank, for the ranking of ontologies AKTiveRank uses as input the search terms provided by a knowledge engineer and, using the output of an ontology search engine, ranks the ontologies We apply a number of metrics in an attempt to investigate their appropriateness for ranking ontologies, and compare the results with a questionnaire-based human study Our results show that AKTiveRank will have great utility although there is potential for improvement

Proceedings ArticleDOI
19 Sep 2005
TL;DR: A software requirements analysis method based ondomain ontology technique, where a mapping between a software requirements specification and the domain ontology that represents semantic components is established, which allows requirements engineers to analyze a requirements specification with respect to the semantics of the application domain.
Abstract: We propose a software requirements analysis method based on domain ontology technique, where we can establish a mapping between a software requirements specification and the domain ontology that represents semantic components. Our ontology system consists of a thesaurus and inference rules and the thesaurus part comprises domain specific concepts and relationships suitable for semantic processing. It allows requirements engineers to analyze a requirements specification with respect to the semantics of the application domain. More concretely, we demonstrate following three kinds of semantic processing through a case study, (1) detecting incompleteness and inconsistency included in a requirements specification, (2) measuring the quality of a specification with respect to its meaning and (3) predicting requirements changes based on semantic analysis on a change history.

Proceedings ArticleDOI
07 Nov 2005
TL;DR: Compared with existing methods, the approach can acquire ontology from relational database automatically by using a group of learning rules instead of using a middle model and can obtain OWL ontology, including the classes, properties, properties characteristics, cardinality and instances, while none of existing methods can acquire all of them.
Abstract: Ontology provides a shared and reusable piece of knowledge about a specific domain, and has been applied in many fields, such as semantic Web, e-commerce and information retrieval, etc. However, building ontology by hand is a very hard and error-prone task. Learning ontology from existing resources is a good solution. Because relational database is widely used for storing data and OWL is the latest standard recommended by W3C, this paper proposes an approach of learning OWL ontology from data in relational database. Compared with existing methods, the approach can acquire ontology from relational database automatically by using a group of learning rules instead of using a middle model. In addition, it can obtain OWL ontology, including the classes, properties, properties characteristics, cardinality and instances, while none of existing methods can acquire all of them. The proposed learning rules have been proven to be correct by practice.

Proceedings ArticleDOI
10 May 2005
TL;DR: With APFEL (Alignment Process Feature Estimation and Learning), this work presents a machine learning approach that explores the user validation of initial alignments for optimizing alignment methods based on extensional and intensional ontology definitions.
Abstract: Ontology alignment is a prerequisite in order to allow for interoperation between different ontologies and many alignment strategies have been proposed to facilitate the alignment task by (semi-)automatic means. Due to the complexity of the alignment task, manually defined methods for (semi-)automatic alignment rarely constitute an optimal configuration of substrategies from which they have been built. In fact, scrutinizing current ontology alignment methods, one may recognize that most are not optimized for given ontologies. Some few include machine learning for automating the task, but their optimization by machine learning means is mostly restricted to the extensional definition of ontology concepts. With APFEL (Alignment Process Feature Estimation and Learning) we present a machine learning approach that explores the user validation of initial alignments for optimizing alignment methods. The methods are based on extensional and intensional ontology definitions. Core to APFEL is the idea of a generic alignment process, the steps of which may be represented explicitly. APFEL then generates new hypotheses for what might be useful features and similarity assessments and weights them by machine learning approaches. APFEL compares favorably in our experiments to competing approaches.

Proceedings ArticleDOI
10 May 2005
TL;DR: The proposed extraction method is a helpful tool to support the process of building domain ontologies for web service descriptions and is conducted in the field of bioinformatics by learning an ontology from the documentation of the web services used in myGrid, a project that supports biology experiments on the Grid.
Abstract: The reasoning tasks that can be performed with semantic web service descriptions depend on the quality of the domain ontologies used to create these descriptions. However, building such domain ontologies is a time consuming and difficult task.We describe an automatic extraction method that learns domain ontologies for web service descriptions from textual documentations attached to web services. We conducted our experiments in the field of bioinformatics by learning an ontology from the documentation of the web services used in myGrid, a project that supports biology experiments on the Grid. Based on the evaluation of the extracted ontology in the context of the project, we conclude that the proposed extraction method is a helpful tool to support the process of building domain ontologies for web service descriptions.

Journal Article
TL;DR: Five different cases studies that illustrate the use of ontologies in metadata representation, in global conceptualization, in high-level querying, in declarative mediation, and in mapping support are discussed.
Abstract: In this paper, we discuss the use of ontologies for data integration. We consider two different settings depending on the system architecture: central and peer-to-peer data integration. Within those settings, we discuss five different cases studies that illustrate the use of ontologies in metadata representation, in global conceptualization, in high-level querying, in declarative mediation, and in mapping support. Each case study is described in detail and accompanied by examples.

Proceedings Article
01 Jan 2005
TL;DR: This paper looks at existing ontology-evaluation methods from the perspective of their integration in one single framework based on a catalogue of qualitative and quantitative measures for ontologies, and sets up a formal model for ontology.
Abstract: The need for evaluation-methodologies emerged very early in the field of ontology development and reuse and it has grown steadily. Yet, no comprehensive and global approach to this problem has been proposed to date. This situation may become a serious obstacle for the success of ontology-based Knowledge Technology, especially in the industrial and commercial sectors. In this paper we look at existing ontology-evaluation methods from the perspective of their integration in one single framework. Based on a catalogue of qualitative and quantitative measures for ontologies, we set up a formal model for ontology. The proposed formal model consists of a meta-ontology Othat characterizes ontologies as semiotic objects. The meta-ontology is complemented with an ontology of ontology evaluation and validation oQual. Based on O and oQual, we identify three main types of measures for ontology evaluation: structural measures, that are typical of ontologies represented as graphs; functional measures, that are related to the intended use of an ontology and of its components, i.e. their function; usability-related measures, that depend on the level of annotation of the considered ontology.

Proceedings Article
30 Jul 2005
TL;DR: An in-depth analysis of the output of the system shows that the model is accurate and has good potentials for text mining and ontology building applications.
Abstract: In this paper we present an unsupervised model for learning arbitrary relations between concepts of a molecular biology ontology for the purpose of supporting text mining and manual ontology building. Relations between named-entities are learned from the GENIA corpus by means of several standard natural language processing techniques. An in-depth analysis of the output of the system shows that the model is accurate and has good potentials for text mining and ontology building applications.

01 Jan 2005
TL;DR: This paper demonstrates how spreading activation improves the result by naturally integrating the mentioned methods to semi-automatically extend and refine ontologies by mining textual data from the Web sites of international online media.
Abstract: This paper describes a system to semi-automatically extend and refine ontologies by mining textual data from the Web sites of international online media. Expanding a seed ontology creates a semantic network through co-occurrence analysis, trigger phrase analysis, and disambiguation based on the WordNet lexical dictionary. Spreading activation then processes this semantic network to find the most probable candidates for inclusion in an extended ontology. Approaches to identifying hierarchical relationships such as subsumption, head noun analysis and WordNet consultation are used to confirm and classify the found relationships. Using a seed ontology on "climate change" as an example, this paper demonstrates how spreading activation improves the result by naturally integrating the mentioned methods.