scispace - formally typeset
Search or ask a question
Topic

Tree-adjoining grammar

About: Tree-adjoining grammar is a research topic. Over the lifetime, 2491 publications have been published within this topic receiving 57813 citations.


Papers
More filters
Journal ArticleDOI
TL;DR: In this article, it was shown that the membership problem for second order non-linear abstract categorical grammars is decidable, and that Montague-like semantics yield to a text generation problem.
Abstract: In this paper we show that the membership problem for second order non-linear Abstract Categorial Grammars is decidable. A consequence of that result is that Montague-like semantics yield to a decidable text generation problem. Furthermore the proof we propose is based on a new tool, Higher Order Intersection Signatures, which grasps statically dynamic properties of ?-terms and presents an interest in its own.

16 citations

01 Jan 2004
TL;DR: The model of unsupervised learning of linguistic structures, ADIOS, is compared to some recent work in computa- tional linguistics and in grammar theory, and how empirical and formal study of language can be best integrated is suggested.
Abstract: Bridging computational, formal and psycholinguistic approaches to language Zach Solan, David Horn, Eytan Ruppin Faculty of Exact Sciences Tel Aviv University Tel Aviv, Israel 69978 {zsolan,horn,ruppin}@post.tau.ac.il Shimon Edelman Department of Psychology Cornell University Ithaca, NY 14853, USA se37@cornell.edu quences (syntagms) is the basis for the classical distributional theory of language [10], as well as for some modern works [11]. Likewise, the pattern — the syntagm and the equiva- lence class of complementary-distribution symbols that may appear in its open slot — is the main representational build- ing block of our system, ADIOS (for Automatic DIstillation Of Structure). Our goal in the present paper is to help bridge statistical and formal approaches to language [12] by placing our work on the unsupervised learning of structure in the context of current research in grammar acquisition in computational lin- guistics, and at the same time to link it to certain formal theo- ries of grammar. Consequently, the following sections outline the main computational principles behind the ADIOS model, and compare these to select approaches from computational and formal linguistics. The algorithmic details of our ap- proach and accounts of its learning from CHILDES corpora and performance in various tests appear elsewhere [1, 2, 3]. In this paper, we chose to exert a tight control over the tar- get language by using a context-free grammar (Figure 1) to generate the learning and testing corpora. Abstract We compare our model of unsupervised learning of linguistic structures, ADIOS [1, 2, 3], to some recent work in computa- tional linguistics and in grammar theory. Our approach resem- bles the Construction Grammar in its general philosophy (e.g., in its reliance on structural generalizations rather than on syn- tax projected by the lexicon, as in the current generative the- ories), and the Tree Adjoining Grammar in its computational characteristics (e.g., in its apparent affinity with Mildly Con- text Sensitive Languages). The representations learned by our algorithm are truly emergent from the (unannotated) corpus data, whereas those found in published works on cognitive and construction grammars and on TAGs are hand-tailored. Thus, our results complement and extend both the computational and the more linguistically oriented research into language acqui- sition. We conclude by suggesting how empirical and formal study of language can be best integrated. The empirical problem of language acquisition The acquisition of language by children — a largely unsuper- vised, amazingly fast and almost invariably successful learn- ing stint — has long been the envy of natural language en- gineers [4, 5, 6] and a daunting enigma for cognitive scien- tists [7, 8]. Computational models of language acquisition or “grammar induction” are usually divided into two categories, depending on whether they subscribe to the classical gener- ative theory of syntax, or invoke “general-purpose” statisti- cal learning mechanisms. We believe that polarization be- tween classical and statistical approaches to syntax hampers the integration of the stronger aspects of each method into a common powerful framework. On the one hand, the statisti- cal approach is geared to take advantage of the considerable progress made to date in the areas of distributed represen- tation, probabilistic learning, and “connectionist” modeling, yet generic connectionist architectures are ill-suited to the ab- straction and processing of symbolic information. On the other hand, classical rule-based systems excel in just those tasks, yet are brittle and difficult to train. We are developing an approach to the acquisition of distri- butional information from raw input (e.g., transcribed speech corpora) that also supports the distillation of structural reg- ularities comparable to those captured by Context Sensitive Grammars out of the accrued statistical knowledge. In think- ing about such regularities, we adopt Langacker’s notion of grammar as “simply an inventory of linguistic units” ([9], p.63). To detect potentially useful units, we identify and pro- cess partially redundant sentences that share the same word sequences. We note that the detection of paradigmatic vari- ation within a slot in a set of otherwise identical aligned se- Figure 1: the context free grammar used to generate the cor- pora for the acquisition tests described here. The principles behind the ADIOS algorithm The representational power of ADIOS and its capacity for un- supervised learning rest on three principles: (1) probabilistic inference of pattern significance, (2) context-sensitive gener- alization, and (3) recursive construction of complex patterns. Each of these is described briefly below. Probabilistic inference of pattern significance. ADIOS rep- resents a corpus of sentences as an initially highly redundant directed graph, in which the vertices are the lexicon entries and the paths correspond, prior to running the algorithm, to corpus sentences. The graph can be informally visualized as a tangle of strands that are partially segregated into bundles.

16 citations

Journal ArticleDOI
TL;DR: It is shown that a graph grammar can be translated into an Event-B specification preserving its semantics, such that one can use several theorem provers available for Event- B to analyze the reachable states of the original graph grammar.
Abstract: Graph grammars may be used as specification technique for different kinds of systems, specially in situations in which states are complex structures that can be adequately modeled as graphs (possibly with an attribute data part) and in which the behavior involves a large amount of parallelism and can be described as reactions to stimuli that can be observed in the state of the system. The verification of properties of such systems is a difficult task due to many aspects: in many situations the systems have an infinite number of states; states themselves are complex and large; there are a number of different computation possibilities due to the fact that rule applications may occur in parallel. There are already some approaches to verification of graph grammars based on model checking, but in these cases only finite state systems can be analyzed. Other approaches propose over- and/or under-approximations of the state-space, but in this case it is not possible to check arbitrary properties. In this work, we propose to use the Event-B formal method and its theorem proving tools to analyze graph grammars. We show that a graph grammar can be translated into an Event-B specification preserving its semantics, such that one can use several theorem provers available for Event-B to analyze the reachable states of the original graph grammar. The translation is based on a relational definition of graph grammars, that was shown to be equivalent to the Single-Pushout approach to graph grammars.

16 citations

Book ChapterDOI
04 Jan 1996

16 citations

Journal ArticleDOI
TL;DR: Under very natural restrictions it can be shown that for two-level grammars pairs (G, G′) there exists a 1 — 1 correspondence between leftmost derivations in G and left most derivation in G′.
Abstract: Making use of the fact that two-level grammars (TLGs) may be thought of as finite specification of context-free grammars (CFGs) with "infinite" sets of productions, known techniques for parsing CFGs are applied to TLGs by first specifying a canonical CFG G? -- called skeleton grammar -- obtained from the "cross-reference" of the TLG G. Under very natural restrictions it can be shown that for these grammar pairs (G, G?) there exists a 1 -- 1 correspondence between leftmost derivations in G and leftmost derivations in G?. With these results a straightforward parsing algorithm for restricted TLGs is given.

16 citations


Network Information
Related Topics (5)
Graph (abstract data type)
69.9K papers, 1.2M citations
85% related
Parsing
21.5K papers, 545.4K citations
85% related
Time complexity
36K papers, 879.5K citations
84% related
Semantics
24.9K papers, 653K citations
82% related
Tree (data structure)
44.9K papers, 749.6K citations
81% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202315
202225
20217
20205
20196
201811