scispace - formally typeset
Search or ask a question
Topic

Context-sensitive grammar

About: Context-sensitive grammar is a research topic. Over the lifetime, 1938 publications have been published within this topic receiving 45911 citations. The topic is also known as: CSG.


Papers
More filters
Posted Content
Miles Osborne1
TL;DR: This thesis concentrates upon automatic grammar correction (or machine learning of grammar) as a solution to the problem of undergeneration by hypothesising that the combined use of data-driven and model-based learning would allowData-driven learning to compensate for model- based learning's incompleteness, whilst model-Based learning would compensate for data- driven learning's unsoundness.
Abstract: When parsing unrestricted language, wide-covering grammars often undergenerate. Undergeneration can be tackled either by sentence correction, or by grammar correction. This thesis concentrates upon automatic grammar correction (or machine learning of grammar) as a solution to the problem of undergeneration. Broadly speaking, grammar correction approaches can be classified as being either {\it data-driven}, or {\it model-based}. Data-driven learners use data-intensive methods to acquire grammar. They typically use grammar formalisms unsuited to the needs of practical text processing and cannot guarantee that the resulting grammar is adequate for subsequent semantic interpretation. That is, data-driven learners acquire grammars that generate strings that humans would judge to be grammatically ill-formed (they {\it overgenerate}) and fail to assign linguistically plausible parses. Model-based learners are knowledge-intensive and are reliant for success upon the completeness of a {\it model of grammaticality}. But in practice, the model will be incomplete. Given that in this thesis we deal with undergeneration by learning, we hypothesise that the combined use of data-driven and model-based learning would allow data-driven learning to compensate for model-based learning's incompleteness, whilst model-based learning would compensate for data-driven learning's unsoundness. We describe a system that we have used to test the hypothesis empirically. The system combines data-driven and model-based learning to acquire unification-based grammars that are more suitable for practical text parsing. Using the Spoken English Corpus as data, and by quantitatively measuring undergeneration, overgeneration and parse plausibility, we show that this hypothesis is correct.

8 citations

Journal Article
TL;DR: In this paper, the authors introduce stratified semantics for Boolean grammars and show how to check if a Boolean grammar generates a language according to this semantics, which covers a class of important and natural languages.
Abstract: We study Boolean grammars. We introduce stratified semantics for Boolean grammars. We show, how to check, if a Boolean grammar generates a language according to this semantics. We show, that stratified semantics covers a class of important and natural languages. We introduce a recognition algorithm for Boolean grammars compliant to this semantics.

8 citations

Journal ArticleDOI
TL;DR: Thedia improves upon related approaches not only in being fully automated and computationally tractable, but also with respect to the class of grammars it is able to invert and the performance of the new executable grammar produced.
Abstract: Reversibility of logic grammars in natural language processing is desirable for both theoretical and practical reasons. This paper addresses this topic in describing a new approach to automated inversion of logic grammars: the Direct Inversion Approach (dia). A logic grammar is inverted by automatically altering the order of literals in the grammar and reformulating certain recursive procedures at compile time. The inversion process results in a new executable grammar, which is evaluated top-down and left-to-right (using a standard Prolog interpreter), but not left-to-right with respect to the original grammar. Thedia improves upon related approaches not only in being fully automated and computationally tractable, but also with respect to the class of grammars it is able to invert and the performance of the new executable grammar produced.

8 citations

Proceedings Article
01 Nov 2013
TL;DR: A new method for machine learning-based optimization of linguist-written Constraint Grammars is presented and the effect of rule ordering/sorting, grammarsectioning and systematic rule changes is discussed and quantitatively evaluated.
Abstract: In this paper we present a new method for machine learning-based optimization of linguist-written Constraint Grammars. The effect of rule ordering/sorting, grammarsectioning and systematic rule changes is discussed and quantitatively evaluated. The F-score improvement was 0.41 percentage points for a mature (Danish) tagging grammar, and 1.36 percentage points for a half-size grammar, translating into a 7-15% error reduction relative to the performance of the untuned grammars.

8 citations

Posted Content
TL;DR: This article motivates a variant of Datalog grammars which allows for a meta-grammatical treatment of coordination, which improves in some respects over previous work on coordination in logic Grammars, although more research is needed for testing it in other respects.
Abstract: In previous work we studied a new type of DCGs, Datalog grammars, which are inspired on database theory. Their efficiency was shown to be better than that of their DCG counterparts under (terminating) OLDT-resolution. In this article we motivate a variant of Datalog grammars which allows us a meta-grammatical treatment of coordination. This treatment improves in some respects over previous work on coordination in logic grammars, although more research is needed for testing it in other respects.

8 citations


Network Information
Related Topics (5)
Graph (abstract data type)
69.9K papers, 1.2M citations
80% related
Time complexity
36K papers, 879.5K citations
79% related
Concurrency
13K papers, 347.1K citations
78% related
Model checking
16.9K papers, 451.6K citations
77% related
Directed graph
12.2K papers, 302.4K citations
77% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202311
202212
20211
20204
20191
20181