scispace - formally typeset
T

Timothy J. O'Donnell

Researcher at McGill University

Publications -  59
Citations -  1185

Timothy J. O'Donnell is an academic researcher from McGill University. The author has contributed to research in topics: Computer science & Parsing. The author has an hindex of 15, co-authored 54 publications receiving 901 citations. Previous affiliations of Timothy J. O'Donnell include Harvard University & Massachusetts Institute of Technology.

Papers
More filters
Journal ArticleDOI

Universality and diversity in human song

TL;DR: Analysis of a natural history of song shows that music appears in every society observed; that variation in song events is well characterized by three dimensions; that musical behavior varies more within societies than across them on these dimensions; and that music is regularly associated with behavioral contexts such as infant care, healing, dance, and love.
Journal ArticleDOI

Unsupervised Lexicon Discovery from Acoustic Input

TL;DR: It is shown that the model is competitive with state-of-the-art spoken term discovery systems, and analyses exploring the model’s behavior and the kinds of linguistic structures it learns are presented.
Journal ArticleDOI

Grammatical morphology as a source of early number word meanings

TL;DR: Although exposure to counting is important to learning number word meanings, hearing number words used outside of these routines—in the quantificational structures of language—may also be highly important in early acquisition.
Book

Productivity and Reuse in Language: A Theory of Linguistic Computation and Storage

TL;DR: This model treats productivity and reuse as the target of inference in a probabilistic framework, asking how an optimal agent can make use of the distribution of forms in the linguistic input to learn the distributionof productive word-formation processes and reusable units in a given language.
Journal Article

Productivity and reuse in language

TL;DR: This thesis presents a formal model of productivity and reuse which treats the problem as a structure-by-structure inference in a Bayesian framework and is built around two proposals: that anything that can be computed can be stored and that any stored item can include subparts which must be computed productively.