DMCA
Unsupervised induction of labeled parse trees by clustering with syntactic features. COLING ’08 (2008)
Cached
Download Links
Citations: | 10 - 4 self |
Citations
1258 | Variants of the Hungarian method for assignment problems
- Kuhn
- 1956
(Show Context)
Citation Context ...e number of times xi and yj share a constituent). To make the graph complete, we add zero weight edges between induced and target labels that do not share any constituent. The Kuhn-Munkres algorithm (=-=Kuhn, 1955-=-; Munkres, 1957) solves this problem, and we used it to perform the LL mapping (see also (Luo, 2005)). We assessed the overall quality of our algorithm, the quality of its labeling stage and the quali... |
540 | TnT – a statistical part-of-speech tagger
- Brants
- 2000
(Show Context)
Citation Context ...ally POS tagged corpus enables us to measure the performance of other induction stages in a controlled manner. Since supervised POS tagging is of very high quality and very efficient computationally (=-=Brants, 2000-=-), this requirement does not seriously limit the practical applicability of a grammar induction algorithm. Our labeled bracketings induction algorithm consists of three stages. We first induce unlabel... |
452 | Algorithms for the assignment and transportation problems
- Munkres
(Show Context)
Citation Context ...times xi and yj share a constituent). To make the graph complete, we add zero weight edges between induced and target labels that do not share any constituent. The Kuhn-Munkres algorithm (Kuhn, 1955; =-=Munkres, 1957-=-) solves this problem, and we used it to perform the LL mapping (see also (Luo, 2005)). We assessed the overall quality of our algorithm, the quality of its labeling stage and the quality of the synta... |
229 | Corpusbased induction of syntactic structure: Models of dependency and constituency.
- Klein, Manning
- 2004
(Show Context)
Citation Context ...Licensed under the Creative Commons Attribution-Noncommercial-Share Alike 3.0 Unported license (http://creativecommons.org/licenses/by-nc-sa/3.0/). Some rights reserved. mar. Most recent work (e.g., (=-=Klein and Manning, 2004-=-; Dennis, 2005; Bod, 2006a; Smith and Eisner, 2006; Seginer, 2007)) annotates text sentences using a hierarchical bracketing (constituents) or a dependency structure, and thus represents the induced g... |
156 | Inducing probabilistic grammars by Bayesian model merging.
- Stolcke, Omohundro
- 1994
(Show Context)
Citation Context ...ocus on the induction of labeled bracketing. Bayesian Model Merging 2 Using, as they did, a greedy mapping with an equal number of labels in the induced and target grammars. 722(BMM) (Stolcke, 1994; =-=Stolcke and Omohundro, 1994-=-) is a framework for inducing PCFG containing both a bracketing and a labeling. The characteristics of this framework (separating prior probability, data likelihood and heuristic search procedures) ca... |
141 | Bayesian learning of probabilistic language models.
- Stolcke
- 1994
(Show Context)
Citation Context ...this paper we focus on the induction of labeled bracketing. Bayesian Model Merging 2 Using, as they did, a greedy mapping with an equal number of labels in the induced and target grammars. 722(BMM) (=-=Stolcke, 1994-=-; Stolcke and Omohundro, 1994) is a framework for inducing PCFG containing both a bracketing and a labeling. The characteristics of this framework (separating prior probability, data likelihood and he... |
119 | A generative constituent-context model for improved grammar induction.
- Klein, Manning
- 2002
(Show Context)
Citation Context ...ch as the WSJ PennTreebank. Recently, works along this line have for the first time outperformed the right branching heuristic baseline for English. These include the constituent–context model (CCM) (=-=Klein and Manning, 2002-=-), its extension using a dependency model (Klein and Manning, 2004), (U)DOP based models (Bod, 2006a; Bod, 2006b; Bod, 2007), an exemplar– based approach (Dennis, 2005), guiding EM using contrastive e... |
105 | Radical construction grammar.
- Croft
- 2001
(Show Context)
Citation Context ...requires POS tags for its labeling stages. Parts-of-speech are widely considered to have a psychological reality (at least in English, including when they are viewed as low-level constructions as in (=-=Croft, 2001-=-)), so this kind of input is reasonable for theoretical research. Moreover, as POS induction is of medium quality (Clark, 2003), using a manually POS tagged corpus enables us to measure the performanc... |
102 | Unsupervised Language Acquisition. - Marcken - 1996 |
100 | Unsupervised learning of natural languages.
- Solan, Horn, et al.
- 2005
(Show Context)
Citation Context ...rk (separating prior probability, data likelihood and heuristic search procedures) can also be found in the grammar induction models of (Wolf, 1982; Langley and Stromsten, 2000; Petasis et al., 2004; =-=Solan et al., 2005-=-). The BMM model used here (Borensztajn and Zuidema, 2007) combines features of (Petasis et al., 2004) and Stolcke’s algorithm, applying the minimum description length (MDL) principle. We use it here ... |
76 | On coreference resolution performance metrics.
- Luo
- 2005
(Show Context)
Citation Context ...es between induced and target labels that do not share any constituent. The Kuhn-Munkres algorithm (Kuhn, 1955; Munkres, 1957) solves this problem, and we used it to perform the LL mapping (see also (=-=Luo, 2005-=-)). We assessed the overall quality of our algorithm, the quality of its labeling stage and the quality of the syntactic clustering (SC) stage. For the overall quality of the induced grammar (both bra... |
63 | Bayesian grammar induction for language modeling.
- Chen
- 1995
(Show Context)
Citation Context ...human language acquisition (in particular, an empirical study of the poverty of stimulus hypothesis), preprocessing for constructing large treebanks (Van Zaanen, 2001), and improving language models (=-=Chen, 1995-=-). In recent years efforts have been made to evaluate the algorithms on manually annotated corpora such as the WSJ PennTreebank. Recently, works along this line have for the first time outperformed th... |
61 | Language acquisition, data compression and generalization - Wolff - 1982 |
58 | Constructions at work.
- Goldberg
- 2006
(Show Context)
Citation Context ...nguistic and psycho-linguistic theories posit some kind of a hierarchical labeled constituent (or constructional) structure, arguing that it has a measurable psychological (cognitive) reality (e.g., (=-=Goldberg, 2006-=-)). The main practical arguments in favor of this approach are that it enables a detailed and large-scale evaluation using annotated corpora, as is done in this paper, and that the output format is su... |
55 | Combining distributional and morphological information for part of speech induction.
- Clark
- 2003
(Show Context)
Citation Context ...English, including when they are viewed as low-level constructions as in (Croft, 2001)), so this kind of input is reasonable for theoretical research. Moreover, as POS induction is of medium quality (=-=Clark, 2003-=-), using a manually POS tagged corpus enables us to measure the performance of other induction stages in a controlled manner. Since supervised POS tagging is of very high quality and very efficient co... |
55 | Memory-based lexical acquisition and processing
- Daelemans
(Show Context)
Citation Context ...s, except of Seginer’s algorithm, which uses plain text. All of these papers induce unlabeled bracketing or dependencies. There are other algorithmic approaches to the problem (e.g., (Adriaans, 1992; =-=Daelemans, 1995-=-; Van Zaanen, 2001)). None of these had evaluated labeled bracketing on annotated corpora. In this paper we focus on the induction of labeled bracketing. Bayesian Model Merging 2 Using, as they did, a... |
54 |
Fast unsupervised incremental parsing
- Seginer
- 2007
(Show Context)
Citation Context ....0 Unported license (http://creativecommons.org/licenses/by-nc-sa/3.0/). Some rights reserved. mar. Most recent work (e.g., (Klein and Manning, 2004; Dennis, 2005; Bod, 2006a; Smith and Eisner, 2006; =-=Seginer, 2007-=-)) annotates text sentences using a hierarchical bracketing (constituents) or a dependency structure, and thus represents the induced grammar through its behavior in a parsing task. Solan et al. (2005... |
51 |
The unsupervised learning of natural language structure, PhD thesis.
- Klein
- 2005
(Show Context)
Citation Context ... Section 3 we detail our algorithm. The experimental setup and results are presented in Sections 4 and 5. 2 Previous Work Unsupervised parsing has attracted researchers for decades (see (Clark, 2001; =-=Klein, 2005-=-) for recent reviews). Many types of input, syntax formalisms, search procedures, and success criteria were used. Among the theoretical and practical motivations to this problem are the study of human... |
46 | Unsupervised language acquisition: Theory and practice.
- Clark
- 2001
(Show Context)
Citation Context ...ious work. In Section 3 we detail our algorithm. The experimental setup and results are presented in Sections 4 and 5. 2 Previous Work Unsupervised parsing has attracted researchers for decades (see (=-=Clark, 2001-=-; Klein, 2005) for recent reviews). Many types of input, syntax formalisms, search procedures, and success criteria were used. Among the theoretical and practical motivations to this problem are the s... |
46 | Learning context-free grammars with a simplicity bias.
- Langley, Stromsten
- 2000
(Show Context)
Citation Context ...and a labeling. The characteristics of this framework (separating prior probability, data likelihood and heuristic search procedures) can also be found in the grammar induction models of (Wolf, 1982; =-=Langley and Stromsten, 2000-=-; Petasis et al., 2004; Solan et al., 2005). The BMM model used here (Borensztajn and Zuidema, 2007) combines features of (Petasis et al., 2004) and Stolcke’s algorithm, applying the minimum descripti... |
41 |
An All-subtrees Approach to Unsupervised Parsing.
- Bod
- 2006
(Show Context)
Citation Context ...ribution-Noncommercial-Share Alike 3.0 Unported license (http://creativecommons.org/licenses/by-nc-sa/3.0/). Some rights reserved. mar. Most recent work (e.g., (Klein and Manning, 2004; Dennis, 2005; =-=Bod, 2006a-=-; Smith and Eisner, 2006; Seginer, 2007)) annotates text sentences using a hierarchical bracketing (constituents) or a dependency structure, and thus represents the induced grammar through its behavio... |
39 | A minimum description length approach to grammar inference. In
- Grunwald
- 1994
(Show Context)
Citation Context ...s et al., 2004) and Stolcke’s algorithm, applying the minimum description length (MDL) principle. We use it here only for initial labeling of existing bracketings. The MDL principle was also used in (=-=Grunwald, 1994-=-; de Marcken, 1995; Clark, 2001). There are only two previous papers we are aware of that induce labeled bracketing and evaluate on corpora annotated with a similar representation (Haghighi and Klein,... |
34 | Bootstrapping Structure into Language: Alignment-Based Learning. - Zaanen - 2002 |
21 | Prototype-driven grammar induction
- Haghighi, Klein
- 2006
(Show Context)
Citation Context ...OS tagged) sentences. experimented with English (WSJ10, Brown10), German (NEGRA10) and Chinese (CTB10) corpora. When comparing to previous work that used manually annotated corpora in its evaluation (=-=Haghighi and Klein, 2006-=-) 2 , we obtained 59.5% labeled f-score on the WSJ10 setup vs. their 35.3% (Section 5). We also show substantial improvement over a random baseline, and that the clustering stage of our algorithm impr... |
18 | Unsupervised parsing with U-DOP
- Bod
- 2006
(Show Context)
Citation Context ...c baseline for English. These include the constituent–context model (CCM) (Klein and Manning, 2002), its extension using a dependency model (Klein and Manning, 2004), (U)DOP based models (Bod, 2006a; =-=Bod, 2006b-=-; Bod, 2007), an exemplar– based approach (Dennis, 2005), guiding EM using contrastive estimation (Smith and Eisner, 2006), and the incremental parser of (Seginer, 2007). All of these use as input POS... |
15 | An exemplar-based approach to unsupervised parsing.
- Dennis
- 2005
(Show Context)
Citation Context ...ve Commons Attribution-Noncommercial-Share Alike 3.0 Unported license (http://creativecommons.org/licenses/by-nc-sa/3.0/). Some rights reserved. mar. Most recent work (e.g., (Klein and Manning, 2004; =-=Dennis, 2005-=-; Bod, 2006a; Smith and Eisner, 2006; Seginer, 2007)) annotates text sentences using a hierarchical bracketing (constituents) or a dependency structure, and thus represents the induced grammar through... |
12 | Is the end of supervised parsing in sight?
- Bod
- 2007
(Show Context)
Citation Context ...or English. These include the constituent–context model (CCM) (Klein and Manning, 2002), its extension using a dependency model (Klein and Manning, 2004), (U)DOP based models (Bod, 2006a; Bod, 2006b; =-=Bod, 2007-=-), an exemplar– based approach (Dennis, 2005), guiding EM using contrastive estimation (Smith and Eisner, 2006), and the incremental parser of (Seginer, 2007). All of these use as input POS tag sequen... |
10 |
Incremental learning of context -free grammars
- Nakamura, Matsumoto
- 2002
(Show Context)
Citation Context ...erarchical bracketing (constituents) or a dependency structure, and thus represents the induced grammar through its behavior in a parsing task. Solan et al. (2005) uses a graph representation, while (=-=Nakamura, 2006-=-) simply uses a grammar formalism such as PCFG. When the bracketing approach is taken, some algorithms label the resulting constituents, while most do not. Each of these approaches can be justified or... |
6 | Yuka Teteisi, and Jun'ichi Tsujii. 2003. GENIA corpus—a semantically annotated corpus for bio-textmining - Kim, Ohta |
3 |
Learning Language from a Categorical Perspective
- ADRIAANS
- 1992
(Show Context)
Citation Context ...POS tag sequences, except of Seginer’s algorithm, which uses plain text. All of these papers induce unlabeled bracketing or dependencies. There are other algorithmic approaches to the problem (e.g., (=-=Adriaans, 1992-=-; Daelemans, 1995; Van Zaanen, 2001)). None of these had evaluated labeled bracketing on annotated corpora. In this paper we focus on the induction of labeled bracketing. Bayesian Model Merging 2 Usin... |
2 |
Entropy-based external cluster evaluation measure
- Rosenberg, Hirschberg
- 2007
(Show Context)
Citation Context ...; X × Y ) where edge (Xi, Yj) has weight wij, 8 Excluding punctuation and null elements, according to the scheme of (Klein, 2005). 9 There are many possible methods for evaluating clustering quality (=-=Rosenberg and Hirschberg, 2007-=-). For our task, overall f-score is a very natural one. We will address other methods in future papers. (3) find a (one-to-one) matching M from X to Y having a maximal weight. In our case, X is the se... |
1 | Georgios Paliouras and Vangelis Karkaletsis, 2004. E-grids: Computationally efficient grammatical inference from positive examples. Grammars - Petasis |
1 |
Annealing Structural Bias
- Smith, Eisner
- 2006
(Show Context)
Citation Context ...commercial-Share Alike 3.0 Unported license (http://creativecommons.org/licenses/by-nc-sa/3.0/). Some rights reserved. mar. Most recent work (e.g., (Klein and Manning, 2004; Dennis, 2005; Bod, 2006a; =-=Smith and Eisner, 2006-=-; Seginer, 2007)) annotates text sentences using a hierarchical bracketing (constituents) or a dependency structure, and thus represents the induced grammar through its behavior in a parsing task. Sol... |
1 | CoNLL X. Bod, 2007. Is the End of Supervised Parsing - of |