scispace - formally typeset
Open AccessJournal ArticleDOI

Fine-Grained Prediction of Syntactic Typology: Discovering Latent Structure with Supervised Learning

Reads0
Chats0
TLDR
This article used a large collection of realistic synthetic languages as training data to predict how often direct objects follow their verbs, how often adjectives follow their nouns, and in general the directionalities of all dependency relations.
Abstract
We show how to predict the basic word-order facts of a novel language given only a corpus of part-of-speech (POS) sequences. We predict how often direct objects follow their verbs, how often adjectives follow their nouns, and in general the directionalities of all dependency relations. Such typological properties could be helpful in grammar induction. While such a problem is usually regarded as unsupervised learning, our innovation is to treat it as supervised learning, using a large collection of realistic synthetic languages as training data. The supervised learner must identify surface features of a language’s POS sequence (hand-engineered or neural features) that correlate with the language’s deeper structure (latent trees). In the experiment, we show: 1) Given a small set of real languages, it helps to add many synthetic languages to the training data. 2) Our system is robust even when the POS sequences include noise. 3) Our system on this task outperforms a grammar induction baseline by a large margin.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Modeling Language Variation and Universals: A Survey on Typological Linguistics for Natural Language Processing

TL;DR: It is shown that to date, the use of information in existing typological databases has resulted in consistent but modest improvements in system performance, due to both intrinsic limitations of databases and under-employment of the typological features included in them.
Journal ArticleDOI

PhaseLink: A Deep Learning Approach to Seismic Phase Association

TL;DR: This work presents PhaseLink, a framework based on recent advances in deep learning for grid‐free earthquake phase association that is expected to improve the resolution of seismicity catalogs, add stability to real‐time seismic monitoring, and streamline automated processing of large seismic data sets.
Proceedings ArticleDOI

On Difficulties of Cross-Lingual Transfer with Order Differences: A Case Study on Dependency Parsing

TL;DR: The authors compare encoders and decoders based on Recurrent Neural Networks (RNNs) and modified self-attentive architectures for cross-lingual transfer and show that RNN-based architectures transfer well to languages that are close to English and perform especially well on distant languages.
Journal ArticleDOI

PhaseLink: A Deep Learning Approach to Seismic Phase Association

TL;DR: In this paper, the authors propose a grid-free approach to link phases together that share a common origin, which is trained on tens of millions of synthetic sequences of P- and S-wave arrival times generated using a simple 1D velocity model.
Posted Content

On Difficulties of Cross-Lingual Transfer with Order Differences: A Case Study on Dependency Parsing

TL;DR: Investigating crosslingual transfer and posit that an orderagnostic model will perform better when transferring to distant foreign languages shows that RNN-based architectures transfer well to languages that are close to English, while self-attentive models have better overall cross-lingualtransferability and perform especially well on distant languages.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings Article

Understanding the difficulty of training deep feedforward neural networks

TL;DR: The objective here is to understand better why standard gradient descent from random initialization is doing so poorly with deep neural networks, to better understand these recent relative successes and help design better algorithms in the future.
Book

Typology and Universals

TL;DR: The authors presents a comprehensive introduction to the method and theory used in studying typology and universals, and provides students and researchers with extensive examples of language universals in phonology, morphology, syntax and semantics.
Journal ArticleDOI

Applications of stochastic context-free grammars using the Inside-Outside algorithm

TL;DR: Two applications in speech recognition of the use of stochastic context-free grammars trained automatically via the Inside-Outside Algorithm, used to model VQ encoded speech for isolated word recognition and compared directly to HMMs used for the same task are described.
Related Papers (5)
Trending Questions (1)
What is syntax of discovery learning?

The provided paper does not discuss the syntax of discovery learning.