Results 1  10
of
132
A maximum entropy model of phonotactics and phonotactic learning
, 2006
"... The study of phonotactics (e.g., the ability of English speakers to distinguish possible words like blick from impossible words like *bnick) is a central topic in phonology. We propose a theory of phonotactic grammars and a learning algorithm that constructs such grammars from positive evidence. Our ..."
Abstract

Cited by 132 (15 self)
 Add to MetaCart
(Show Context)
The study of phonotactics (e.g., the ability of English speakers to distinguish possible words like blick from impossible words like *bnick) is a central topic in phonology. We propose a theory of phonotactic grammars and a learning algorithm that constructs such grammars from positive evidence. Our grammars consist of constraints that are assigned numerical weights according to the principle of maximum entropy. Possible words are assessed by these grammars based on the weighted sum of their constraint violations. The learning algorithm yields grammars that can capture both categorical and gradient phonotactic patterns. The algorithm is not provided with any constraints in advance, but uses its own resources to form constraints and weight them. A baseline model, in which Universal Grammar is reduced to a feature set and an SPEstyle constraint format, suffices to learn many phonotactic phenomena. In order to learn nonlocal phenomena such as stress and vowel harmony, it is necessary to augment the model with autosegmental tiers and metrical grids. Our results thus offer novel, learningtheoretic support for such representations. We apply the model to English syllable onsets, Shona vowel harmony, quantityinsensitive stress typology, and the full phonotactics of Wargamay, showing that the learned grammars capture the distributional generalizations of these languages and accurately predict the findings of a phonotactic experiment.
Learning Phonology With Substantive Bias: An Experimental and Computational Study of Velar Palatalization
, 2006
"... There is an active debate within the field of phonology concerning the cognitive status of substantive phonetic factors such as ease of articulation and perceptual distinctiveness. A new framework is proposed in which substance acts as a bias, or prior, on phonological learning. Two experiments test ..."
Abstract

Cited by 84 (2 self)
 Add to MetaCart
There is an active debate within the field of phonology concerning the cognitive status of substantive phonetic factors such as ease of articulation and perceptual distinctiveness. A new framework is proposed in which substance acts as a bias, or prior, on phonological learning. Two experiments tested this framework with a method in which participants are first provided highly impoverished evidence of a new phonological pattern, and then tested on how they extend this pattern to novel contexts and novel sounds. Participants were found to generalize velar palatalization (e.g., the change from [k]asinkeep to [t�ʃ]asincheap) in a way that accords with linguistic typology, and that is predicted by a cognitive bias in favor of changes that relate perceptually similar sounds. Velar palatalization was extended from the mid front vowel context (i.e., before [e]asincape) to the high front vowel context (i.e., before [i]asin keep), but not vice versa. The key explanatory notion of perceptual similarity is quantified with a psychological model of categorization, and the substantively biased framework is formalized as a conditional random field. Implications of these results for the debate on substance, theories of phonological generalization, and the formalization of similarity are discussed.
Maximum Entropy models and Stochastic Optimality Theory. Rutgers Optimality Archive 625
, 2004
"... In a series of recent publications (most notably Boersma (1998); see also Boersma and Hayes (2001)), Paul Boersma has developed a stochastic generalization of standard Optimality Theory in the sense of Prince and Smolensky (1993). While a classical OT grammar maps a set of candidates to its optimal ..."
Abstract

Cited by 52 (1 self)
 Add to MetaCart
In a series of recent publications (most notably Boersma (1998); see also Boersma and Hayes (2001)), Paul Boersma has developed a stochastic generalization of standard Optimality Theory in the sense of Prince and Smolensky (1993). While a classical OT grammar maps a set of candidates to its optimal element (or elements),
Harmonic grammar with linear programming: From linear . . .
, 2009
"... Harmonic Grammar (HG) is a model of linguistic constraint interaction in which wellformedness is calculated as the sum of weighted constraint violations. We show how linear programming algorithms can be used to determine whether there is a weighting for a set of constraints that fits a set of ling ..."
Abstract

Cited by 40 (9 self)
 Add to MetaCart
Harmonic Grammar (HG) is a model of linguistic constraint interaction in which wellformedness is calculated as the sum of weighted constraint violations. We show how linear programming algorithms can be used to determine whether there is a weighting for a set of constraints that fits a set of linguistic data. The associated software package OTHelp provides a practical tool for studying large and complex linguistic systems in the HG framework and comparing the results with those of OT. We first describe the translation from Harmonic Grammars to systems solvable by linear programming algorithms. We then develop an HG analysis of ATR harmony in Lango that is, we argue, superior to the existing OT and rulebased treatments. We further highlight the usefulness of OTHelp, and the analytic power of HG, with a set of studies of the predictions HG makes for phonological typology.
Convergence properties of a gradual learning algorithm for Harmonic Grammar. Rutgers Optimality Archive 970
, 2008
"... Abstract. This paper investigates a gradual online learning algorithm for Harmonic Grammar. By adapting existing convergence proofs for perceptrons, we show that for any nonvarying target language, HarmonicGrammar learners are guaranteed to converge to an appropriate grammar, if they receive compl ..."
Abstract

Cited by 39 (14 self)
 Add to MetaCart
Abstract. This paper investigates a gradual online learning algorithm for Harmonic Grammar. By adapting existing convergence proofs for perceptrons, we show that for any nonvarying target language, HarmonicGrammar learners are guaranteed to converge to an appropriate grammar, if they receive complete information about the structure of the learning data. We also prove convergence when the learner incorporates evaluation noise, as in Stochastic Optimality Theory. Computational tests of the algorithm show that it converges quickly. When learners receive incomplete information (e.g. some structure remains hidden), tests indicate that the algorithm is more likely to converge than two comparable OptimalityTheoretic learning algorithms.
Linear Optimality Theory as a Model of Gradience in Grammar
 In Gradience in Grammar: Generative Perspectives, ed. Gisbert Fanselow, Caroline Féry, Ralph Vogel, and Matthias Schlesewsky
, 2005
"... This paper provides an overview of Linear Optimality Theory (LOT), a variant of Optimality Theory (OT) designed for the modeling of gradient acceptability judgment data. We summarize the empirical properties of gradient data that have been reported in the experimental literature, and use them to mot ..."
Abstract

Cited by 33 (0 self)
 Add to MetaCart
(Show Context)
This paper provides an overview of Linear Optimality Theory (LOT), a variant of Optimality Theory (OT) designed for the modeling of gradient acceptability judgment data. We summarize the empirical properties of gradient data that have been reported in the experimental literature, and use them to motivate the design of LOT. We discuss LOT’s notions of constraint competition and optimality, as well as a new formulation of ranking argumentation, which makes it possible to apply standard parameter estimation techniques to LOT. Then the LOT model is compared to Standard OT, to Harmonic Grammar, and to recently proposed probabilisitic versions of OT. 1.
Gradient Grammar: An Effect of Animacy on the Syntax of give in New Zealand and American English
, 2007
"... ..."
Weighted Constraints in Generative Linguistics
 Cognitive Science
, 2009
"... Harmonic Grammar (HG) and Optimality Theory (OT) are closely related formal frameworks for the study of language. In both, the structure of a given language is determined by the relative strengths of a set of constraints. They differ in how these strengths are represented: as numerical weights (HG) ..."
Abstract

Cited by 21 (3 self)
 Add to MetaCart
(Show Context)
Harmonic Grammar (HG) and Optimality Theory (OT) are closely related formal frameworks for the study of language. In both, the structure of a given language is determined by the relative strengths of a set of constraints. They differ in how these strengths are represented: as numerical weights (HG) or as ranks (OT). Weighted constraints have advantages for the construction of accounts of language learning and other cognitive processes, partly because they allow for the adaptation of connectionist and statistical models. HG has been little studied in generative linguistics, however, largely due to influential claims that weighted constraints make incorrect predictions about the typology of natural languages, predictions that are not shared by the more popular OT. This paper makes the case that HG is in fact a promising framework for typological research, and reviews and extends the existing arguments for weighted over ranked constraints. 1