scispace - formally typeset
Search or ask a question

Showing papers on "Tree-adjoining grammar published in 2009"


Proceedings ArticleDOI
06 Aug 2009
TL;DR: A novel method for learning a type of Synchronous Tree Adjoining Grammar and associated probabilities from aligned tree/string training data and a method of converting these grammars to a weakly equivalent tree transducer for decoding is introduced.
Abstract: Tree Adjoining Grammars have well-known advantages, but are typically considered too difficult for practical systems. We demonstrate that, when done right, adjoining improves translation quality without becoming computationally intractable. Using adjoining to model optionality allows general translation patterns to be learned without the clutter of endless variations of optional material. The appropriate modifiers can later be spliced in as needed. In this paper, we describe a novel method for learning a type of Synchronous Tree Adjoining Grammar and associated probabilities from aligned tree/string training data. We introduce a method of converting these grammars to a weakly equivalent tree transducer for decoding. Finally, we show that adjoining results in an end-to-end improvement of +0.8 Bleu over a baseline statistical syntax-based MT model on a large-scale Arabic/English MT task.

69 citations


Journal ArticleDOI
TL;DR: The results imply undecidability of a number of decision problems of unary conjunctive grammars, as well as non-existence of a recursive function bounding the growth rate of the generated languages.
Abstract: It has recently been proved (Jez, DLT 2007) that conjunctive grammars (that is, context-free grammars augmented by conjunction) generate some non-regular languages over a one-letter alphabet. The present paper improves this result by constructing conjunctive grammars for a larger class of unary languages. The results imply undecidability of a number of decision problems of unary conjunctive grammars, as well as non-existence of a recursive function bounding the growth rate of the generated languages. An essential step of the argument is a simulation of a cellular automaton recognizing positional notation of numbers using language equations.

65 citations


Proceedings ArticleDOI
31 May 2009
TL;DR: This work presents a theoretically principled model which learns compact and simple grammars, uncovering latent linguistic structures (e.g., verb subcategorisation), and in doing so far out-performs a standard PCFG.
Abstract: Tree substitution grammars (TSGs) are a compelling alternative to context-free grammars for modelling syntax. However, many popular techniques for estimating weighted TSGs (under the moniker of Data Oriented Parsing) suffer from the problems of inconsistency and over-fitting. We present a theoretically principled model which solves these problems using a Bayesian non-parametric formulation. Our model learns compact and simple grammars, uncovering latent linguistic structures (e.g., verb subcategorisation), and in doing so far out-performs a standard PCFG.

63 citations


Book ChapterDOI
16 Feb 2009
TL;DR: The present paper introduces and illustrates the basics of grammar convergence, and to transform the grammars until they become syntactically identical.
Abstract: Grammar convergence is a lightweight verification method for establishing and maintaining the correspondence between grammar knowledge ingrained in all kinds of software artifacts, e.g., object models, XML schemas, parser descriptions, or language documents. The central idea is to extract grammars from diverse software artifacts, and to transform the grammars until they become syntactically identical. The present paper introduces and illustrates the basics of grammar convergence.

52 citations


Journal ArticleDOI
TL;DR: A new Stochastic Context Free Grammar based framework has been introduced allowing the production of binding site descriptors for analysis of protein sequences and suggests that this system may be particularly suited to deal with patterns shared by non-homologous proteins.
Abstract: Background In the last decade, there have been many applications of formal language theory in bioinformatics such as RNA structure prediction and detection of patterns in DNA. However, in the field of proteomics, the size of the protein alphabet and the complexity of relationship between amino acids have mainly limited the application of formal language theory to the production of grammars whose expressive power is not higher than stochastic regular grammars. However, these grammars, like other state of the art methods, cannot cover any higher-order dependencies such as nested and crossing relationships that are common in proteins. In order to overcome some of these limitations, we propose a Stochastic Context Free Grammar based framework for the analysis of protein sequences where grammars are induced using a genetic algorithm.

37 citations


Proceedings ArticleDOI
06 Aug 2009
TL;DR: A novel approach for syntax-based statistical MT, which builds on a variant of tree adjoining grammar (TAG) in combination with a discriminative model that can condition on rich features of the source-language string.
Abstract: We describe a novel approach for syntax-based statistical MT, which builds on a variant of tree adjoining grammar (TAG). Inspired by work in discriminative dependency parsing, the key idea in our approach is to allow highly flexible reordering operations during parsing, in combination with a discriminative model that can condition on rich features of the source-language string. Experiments on translation from German to English show improvements over phrase-based systems, both in terms of BLEU scores and in human evaluations.

36 citations


Proceedings ArticleDOI
30 Mar 2009
TL;DR: A new parsing algorithm is described that has the advantage to be incremental and to support PMCFG directly rather than the weaker MCFG formalism, which allows it to be used for grammar based word prediction.
Abstract: Parallel Multiple Context-Free Grammar (PMCFG) is an extension of context-free grammar for which the recognition problem is still solvable in polynomial time. We describe a new parsing algorithm that has the advantage to be incremental and to support PMCFG directly rather than the weaker MCFG formalism. The algorithm is also top-down which allows it to be used for grammar based word prediction.

36 citations


Journal ArticleDOI
TL;DR: It is demonstrated that every Boolean grammar can be transformed into an equivalent (under the new semantics) grammar in normal form, and an O(n^3) algorithm for parsing that applies to any such normalized Boolean grammar is proposed.
Abstract: Boolean grammars [A. Okhotin, Boolean grammars, Information and Computation 194 (1) (2004) 19-48] are a promising extension of context-free grammars that supports conjunction and negation in rule bodies. In this paper, we give a novel semantics for Boolean grammars which applies to all such grammars, independently of their syntax. The key idea of our proposal comes from the area of negation in logic programming, and in particular from the so-called well-founded semantics which is widely accepted in this area to be the ''correct'' approach to negation. We show that for every Boolean grammar there exists a distinguished (three-valued) interpretation of the non-terminal symbols, which satisfies all the rules of the grammar and at the same time is the least fixed-point of an operator associated with the grammar. Then, we demonstrate that every Boolean grammar can be transformed into an equivalent (under the new semantics) grammar in normal form. Based on this normal form, we propose an O(n^3) algorithm for parsing that applies to any such normalized Boolean grammar. In summary, the main contribution of this paper is to provide a semantics which applies to all Boolean grammars while at the same time retaining the complexity of parsing associated with this type of grammars.

35 citations


Journal ArticleDOI
TL;DR: A new syntactic model, called pure two-dimensional (2D) context-free grammar (P2DCFG) is introduced based on the notion of pure context- free string grammar, and the rectangular picture generative power of this 2D grammar model is investigated.

34 citations


Book ChapterDOI
01 Jan 2009

30 citations


Proceedings ArticleDOI
31 May 2009
TL;DR: This work presents a pair of grammar transformations that admit an efficient cubic-time CKY-style parsing algorithm despite leaving most of the grammar in n-ary form and describes a two-pass coarse-to-fine parsing approach that prunes the search space using predictions from a subset of the original grammar.
Abstract: The tree-transducer grammars that arise in current syntactic machine translation systems are large, flat, and highly lexicalized. We address the problem of parsing efficiently with such grammars in three ways. First, we present a pair of grammar transformations that admit an efficient cubic-time CKY-style parsing algorithm despite leaving most of the grammar in n-ary form. Second, we show how the number of intermediate symbols generated by this transformation can be substantially reduced through binarization choices. Finally, we describe a two-pass coarse-to-fine parsing approach that prunes the search space using predictions from a subset of the original grammar. In all, parsing time reduces by 81%. We also describe a coarse-to-fine pruning scheme for forest-based language model reranking that allows a 100-fold increase in beam size while reducing decoding time. The resulting translations improve by 1.3 BLEU.


Book ChapterDOI
07 Sep 2009
TL;DR: The unfolding semantics is generalized to the abstract setting of (single pushout) rewriting over adhesive categories, which applies to a wider class of systems, which is due to the use of a refined notion of grammar morphism.
Abstract: We generalize the unfolding semantics, previously developed for concrete formalisms such as Petri nets and graph grammars, to the abstract setting of (single pushout) rewriting over adhesive categories. The unfolding construction is characterized as a coreflection, i.e. the unfolding functor arises as the right adjoint to the embedding of the category of occurrence grammars into the category of grammars. As the unfolding represents potentially infinite computations, we need to work in adhesive categories with "well-behaved" colimits of ω-chains of monomorphisms. Compared to previous work on the unfolding of Petri nets and graph grammars, our results apply to a wider class of systems, which is due to the use of a refined notion of grammar morphism.

Journal ArticleDOI
TL;DR: This paper proposes an alternative definition of MCTAG that characterizes the trees in the tree language of an MCT AG via the properties of the derivation trees (in the underlying TAG) the M CTAG licences, and provides similar characterizations for various types of MctAG.
Abstract: Multicomponent Tree Adjoining Grammars (MCTAGs) are a formalism that has been shown to be useful for many natural language applications. The definition of non-local MCTAG however is problematic since it refers to the process of the derivation itself: a simultaneity constraint must be respected concerning the way the members of the elementary tree sets are added. Looking only at the result of a derivation (i.e., the derived tree and the derivation tree), this simultaneity is no longer visible and therefore cannot be checked. i.e., this way of characterizing MCTAG does not allow to abstract away from the concrete order of derivation. In this paper, we propose an alternative definition of MCTAG that characterizes the trees in the tree language of an MCTAG via the properties of the derivation trees (in the underlying TAG) the MCTAG licences. We provide similar characterizations for various types of MCTAG. These characterizations give a better understanding of the formalisms, they allow a more systematic comparison of different types of MCTAG, and, furthermore, they can be exploited for parsing.

Proceedings ArticleDOI
08 Jul 2009
TL;DR: The results show some benefit of using meta-grammars in GE and re-emphasize the grammar's impact on GE's performance, which can adapt to the environment by altering the grammar.
Abstract: This paper is concerned with the challenge of learning solutions to problems. The method employed here is a grammar based heuristic, where domain knowledge is encoded in a generative grammar, while evolution drives the update of the population of solutions. Furthermore the method can adapt to the environment by altering the grammar. The implementation consists of the grammar-based Genetic Programming approach of Grammatical Evolution (GE). A number of different constructions of grammars and operators for manipulating the grammars and the evolutionary algorithm are investigated, as well as a meta-grammar GE which allows a more flexible grammar. The results show some benefit of using meta-grammars in GE and re-emphasize the grammar's impact on GE's performance.

Proceedings ArticleDOI
30 Mar 2009
TL;DR: This work presents several algorithms for assigning heads in phrase structure trees, based on different linguistic intuitions on the role of heads in natural language syntax, and evaluates algorithms based on the match with gold standard head-annotations, and the comparative parsing accuracy of the lexicalized grammars they give rise to.
Abstract: We present several algorithms for assigning heads in phrase structure trees, based on different linguistic intuitions on the role of heads in natural language syntax. Starting point of our approach is the observation that a head-annotated treebank defines a unique lexicalized tree substitution grammar. This allows us to go back and forth between the two representations, and define objective functions for the unsupervised learning of head assignments in terms of features of the implicit lexicalized tree grammars. We evaluate algorithms based on the match with gold standard head-annotations, and the comparative parsing accuracy of the lexicalized grammars they give rise to. On the first task, we approach the accuracy of hand-designed heuristics for English and inter-annotation-standard agreement for German. On the second task, the implied lexicalized grammars score 4% points higher on parsing accuracy than lexicalized grammars derived by commonly used heuristics.

Book ChapterDOI
12 Aug 2009
TL;DR: The compressed membership problem for one-nonterminal conjunctive grammars over {a } is proved to be EXPTIME-complete, while the equivalence, finiteness and emptiness problems for these Grammars are shown to be undecidable.
Abstract: Conjunctive grammars over an alphabet Σ = {a } are studied, with the focus on the special case with a unique nonterminal symbol. Such a grammar is equivalent to an equation X = φ (X ) over sets of natural numbers, using union, intersection and addition. It is shown that every grammar with multiple nonterminals can be encoded into a grammar with a single nonterminal, with a slight modification of the language. Based on this construction, the compressed membership problem for one-nonterminal conjunctive grammars over {a } is proved to be EXPTIME-complete, while the equivalence, finiteness and emptiness problems for these grammars are shown to be undecidable.

Journal ArticleDOI
TL;DR: This work proposes a relational and logical approach to graph grammars that allows formal verification of systems using mathematical induction and allows proving properties of systems with infinite state-spaces.

Proceedings ArticleDOI
20 Sep 2009
TL;DR: A completed effort to recover the relationships between all the grammars that occur in the different versions of the Java Language Specification is described, suggesting a form of consistency management for the JLS in particular, and language specifications in general.
Abstract: We describe a completed effort to recover the relationships between all the grammars that occur in the different versions of the Java Language Specification (JLS). The relationships are represented as grammar transformations that capture all accidental or intended differences between the JLS grammars. This process is mechanized and it is driven by simple measures of nominal or structural differences between any pair of grammars involved. Our work suggests a form of consistency management for the JLS in particular, and language specifications in general.

Journal ArticleDOI
Fumiya Okubo1
TL;DR: In formal language theory, it is of interest to study semi-conditional grammars from the point of view of descriptional complexity regarding the number of nonterminals and theNumber of non-context-free productions.

Book ChapterDOI
27 Mar 2009
TL;DR: It is shown that if tree grammars are nondeterministic or non-linear, then reducing their numbers of parameters cannot be done without an exponential blow-up in grammar size.
Abstract: Trees can be conveniently compressed with linear straight-line context-free tree grammars. Such grammars generalize straight-line context-free string grammars which are widely used in the development of algorithms that execute directly on compressed structures (without prior decompression). It is shown that every linear straight-line context-free tree grammar can be transformed in polynomial time into a monadic (and linear) one. A tree grammar is monadic if each nonterminal uses at most one context parameter. Based on this result, a polynomial time algorithm is presented for testing whether a given nondeterministic tree automaton with sibling constraints accepts a tree given by a linear straight-line context-free tree grammar. It is shown that if tree grammars are nondeterministic or non-linear, then reducing their numbers of parameters cannot be done without an exponential blow-up in grammar size.

Book ChapterDOI
31 Mar 2009
TL;DR: Some results on the generative capacities of such grammars that Petri nets are restricted to some known structural subclasses of Petrinets are presented.
Abstract: A Petri net controlled grammar is a context-free grammar with a control by a Petri net whose transitions are labeled with rules of the grammar or the empty string and the associated language consists of all terminal strings which can be derived in the grammar and the sequence of rules in a derivation is in the image of a successful occurrence of transitions of the net. We present some results on the generative capacities of such grammars that Petri nets are restricted to some known structural subclasses of Petri nets.

Journal Article
TL;DR: In this paper, the authors studied cooperating distributed grammar systems working in the terminal derivation mode where the components are variants of permitting grammars, and they proved that the families of random contexts and languages generated by permitting components coincide.
Abstract: This paper studies cooperating distributed grammar systems working in the terminal derivation mode where the components are variants of permitting grammars. It proves that although the family of permitting languages is strictly included in the family of random context languages, the families of random context languages and languages generated by permitting cooperating distributed grammar systems in the above mentioned derivation mode coincide. Moreover, if the components are so-called left-permitting grammars, then cooperating distributed grammar systems in the terminal mode characterize the class of context-sensitive languages, or if erasing rules are allowed, the class of recursively enumerable languages. Descriptional complexity results are also presented. It is shown that the number of permitting components can be bounded, in the case of left-permitting components with erasing rules even together with the number of nonterminals.

01 Jan 2009
TL;DR: This thesis addresses the design of appropriate formalisms and algorithms to be used for natural language processing and focuses on the Tree-Adjoining Grammar formalism as a base and on the mechanism of grammar synchronization for managing relationships between the input and output of a natural languageprocessing system.
Abstract: This thesis addresses the design of appropriate formalisms and algorithms to be used for natural language processing. This entails a delicate balance between the ability of a formalism to capture the linguistic generalizations required by natural language processing applications and the ability of a natural language processing application based on the formalism to process the formalism efficiently enough to be useful. I focus on the Tree-Adjoining Grammar formalism as a base and on the mechanism of grammar synchronization for managing relationships between the input and output of a natural language processing system. Grammar synchronization is a formal concept by which the derivations of two distinct grammars occur in tandem so that a single isomorphic derivation produces distinct derived structures in each of the synchronized grammars. Using synchronization implies a strong assumption—one that I seek to justify in the second part of the thesis—namely that certain critical relationships in natural language applications, such as the relationship between the syntax and semantics of a language or the relationship between the syntax of two natural languages, are close enough to be expressed with grammars that share a derivational structure. The extent of the isomorphism between the derived structures of the related languages is determined only in part by the synchronization. The base formalism chosen can offer greater or lesser opportunity for divergence in the derived structures. My choice of a base formalism is motivated directly by research into applications of synchronous TAG-based grammars to two natural language applications: semantic interpretation and natural language translations. I first examine a range of TAG variants that have not previously been studied in this level of detail to determine their computational properties and to develop algorithms that can be used to process them. Original results on the complexity of these formalisms are presented as well as novel algorithms for factorizing grammars to reduce the time required to process them. In Part II, I develop applications of synchronous Limited Delay Tree-Local Multicomponent TAG to semantic interpretation and probabilistic synchronous Tree Insertion Grammar to statistical natural language translation.

Book ChapterDOI
20 Sep 2009
TL;DR: In this article, the global grammar constraint over restricted classes of context free grammars like deterministic and unambiguous context-free grammar was investigated, and it was shown that detecting disentailment for the GRAMMAR constraint in these cases is as hard as parsing an unrestricted context free grammar.
Abstract: We investigate the global GRAMMAR constraint over restricted classes of context free grammars like deterministic and unambiguous context-free grammars. We show that detecting disentailment for the GRAMMAR constraint in these cases is as hard as parsing an unrestricted context free grammar.We also consider the class of linear grammars and give a propagator that runs in quadratic time. Finally, to demonstrate the use of linear grammars, we show that a weighted linear GRAMMAR constraint can efficiently encode the EDITDISTANCE constraint, and a conjunction of the EDITDISTANCE constraint and the REGULAR constraint.

Book ChapterDOI
01 Jul 2009
TL;DR: This work gives an exposition of strongly regular grammars and a transformation by Mohri and Nederhof on sets of mutually recursive nonterminals and uses it as a subprocedure to obtain tighter regular approximations to a given context-free grammar.
Abstract: We consider algorithms for approximating context---free grammars by regular grammars, making use of Chomsky's characterization of non---self---embedding grammars as generating regular languages and a transformation by Mohri and Nederhof on sets of mutually recursive nonterminals. We give an exposition of strongly regular grammars and this transformation, and use it as a subprocedure to obtain tighter regular approximations to a given context-free grammar. In another direction, the generalization by a 1---lookahead extends Mohri and Nederhof's transformation by incorporating more context into the regular approximation at the expense of a larger grammar.

Book ChapterDOI
01 Jul 2009
TL;DR: It is shown that applying linear erasing to a Petri net language yields a language generated by a non-erasing matrix grammar, which yields a reformulation of the problem of whether erasing rules in matrix grammars can be eliminated.
Abstract: It is shown that applying linear erasing to a Petri net language yields a language generated by a non-erasing matrix grammar. The proof uses Petri net controlled grammars. These are context-free grammars, where the application of productions has to comply with a firing sequence in a Petri net. Petri net controlled grammars are equivalent to arbitrary matrix grammars (without appearance checking), but a certain restriction on them (linear Petri net controlled grammars) leads to the class of languages generated by non-erasing matrix grammars. It is also shown that in Petri net controlled grammars (with final markings and arbitrary labeling), erasing rules can be eliminated, which yields a reformulation of the problem of whether erasing rules in matrix grammars can be eliminated.


Proceedings ArticleDOI
07 Oct 2009
TL;DR: It is shown how parsing of trees can be formalized in terms of the intersection of two tree languages, with a focus on weighted regular tree grammars and weighted tree adjoining Grammars.
Abstract: We show how parsing of trees can be formalized in terms of the intersection of two tree languages. The focus is on weighted regular tree grammars and weighted tree adjoining grammars. Potential applications are discussed, such as parameter estimation across formalisms.

Book ChapterDOI
31 Mar 2009
TL;DR: This paper answers three open questions concerning the generative power of some simple variants of context-free grammars regulated by context conditions and presents some normal form results, an overview of known results, and unsolved problems.
Abstract: This paper answers three open questions concerning the generative power of some simple variants of context-free grammars regulated by context conditions. Specifically, it discusses the generative power of so-called context-free semi-conditional grammars (which are random context grammars where permitting and forbidding sets are replaced with permitting and forbidding strings) where permitting and forbidding strings of each production are of length no more than one, and of simple semi-conditional grammars where, in addition, no production has attached both a permitting and a forbidding string. Finally, this paper also presents some normal form results, an overview of known results, and unsolved problems.