Results 1 - 10
of
4,251
A new characteristic property of rich words, Theoret
- Comput. Sci
"... ABSTRACT. Originally introduced and studied by the third and fourth authors together with J. Justin and S. Widmer (2008), rich words constitute a new class of finite and infinite words char-acterized by containing the maximal number of distinct palindromes. Several characterizations of rich words ha ..."
Abstract
-
Cited by 9 (3 self)
- Add to MetaCart
ABSTRACT. Originally introduced and studied by the third and fourth authors together with J. Justin and S. Widmer (2008), rich words constitute a new class of finite and infinite words char-acterized by containing the maximal number of distinct palindromes. Several characterizations of rich words
Feature-Rich Part-of-Speech Tagging with a Cyclic Dependency Network
- IN PROCEEDINGS OF HLT-NAACL
, 2003
"... We present a new part-of-speech tagger that demonstrates the following ideas: (i) explicit use of both preceding and following tag contexts via a dependency network representation, (ii) broad use of lexical features, including jointly conditioning on multiple consecutive words, (iii) effective ..."
Abstract
-
Cited by 693 (23 self)
- Add to MetaCart
We present a new part-of-speech tagger that demonstrates the following ideas: (i) explicit use of both preceding and following tag contexts via a dependency network representation, (ii) broad use of lexical features, including jointly conditioning on multiple consecutive words, (iii
Feature Selection for Extracting Semantically Rich Words
, 2004
"... The utility of semantic knowledge, in the form of ontologies, is widely acknowledged. In particular, semantic knowledge facilitates integration, visualization, and maintenance of information from various sources. However, the majority of previous work in this field has tried to learn ontologies for ..."
Abstract
-
Cited by 6 (0 self)
- Add to MetaCart
for relatively constrained domains. In other words, to date, there has been relatively little work on trying to construct ontologies for an open domain, where there are enormous needs for such ontologies. Moreover, there have been few studies that empirically examine the value of text learning techniques
Lambda Words: A Class of Rich Words Defined Over an Infinite Alphabet
, 2013
"... Lambda words are sequences obtained by encoding the differences between ordered elements of the form i + jθ, where i and j are non-negative integers and 1 < θ < 2. Lambda words are right-infinite words defined over an infinite alphabet that have connections with Sturmian words, Christoffel wor ..."
Abstract
- Add to MetaCart
words, and interspersion arrays. We show that Lambda words are infinite rich words. Furthermore, any Lambda word may be mapped onto a right-infinite word over a three-letter alphabet. Although the mapping preserves palindromes and non-palindromes of the Lambda word, the resulting Gamma word is not rich.
The Infinite Hidden Markov Model
- Machine Learning
, 2002
"... We show that it is possible to extend hidden Markov models to have a countably infinite number of hidden states. By using the theory of Dirichlet processes we can implicitly integrate out the infinitely many transition parameters, leaving only three hyperparameters which can be learned from data. Th ..."
Abstract
-
Cited by 637 (41 self)
- Add to MetaCart
. These three hyperparameters define a hierarchical Dirichlet process capable of capturing a rich set of transition dynamics. The three hyperparameters control the time scale of the dynamics, the sparsity of the underlying state-transition matrix, and the expected number of distinct hidden states in a finite
The information bottleneck method
, 1999
"... We define the relevant information in a signal x ∈ X as being the information that this signal provides about another signal y ∈ Y. Examples include the information that face images provide about the names of the people portrayed, or the information that speech sounds provide about the words spoken. ..."
Abstract
-
Cited by 540 (35 self)
- Add to MetaCart
We define the relevant information in a signal x ∈ X as being the information that this signal provides about another signal y ∈ Y. Examples include the information that face images provide about the names of the people portrayed, or the information that speech sounds provide about the words spoken
Finding structure in time
- COGNITIVE SCIENCE
, 1990
"... Time underlies many interesting human behaviors. Thus, the question of how to represent time in connectionist models is very important. One approach is to represent time implicitly by its effects on processing rather than explicitly (as in a spatial representation). The current report develops a pro ..."
Abstract
-
Cited by 2071 (23 self)
- Add to MetaCart
of prior internal states. A set of simulations is reported which range from relatively simple problems (temporal version of XOR) to discovering syntactic/semantic features for words. The networks are able to learn interesting internal representations which incorporate task demands with memory demands
Knowledge-rich Word Sense Disambiguation Rivaling Supervised Systems
"... One of the main obstacles to highperformance Word Sense Disambiguation (WSD) is the knowledge acquisition bottleneck. In this paper, we present a methodology to automatically extend WordNet with large amounts of semantic relations from an encyclopedic resource, namely Wikipedia. We show that, when p ..."
Abstract
-
Cited by 41 (7 self)
- Add to MetaCart
One of the main obstacles to highperformance Word Sense Disambiguation (WSD) is the knowledge acquisition bottleneck. In this paper, we present a methodology to automatically extend WordNet with large amounts of semantic relations from an encyclopedic resource, namely Wikipedia. We show that, when
Results 1 - 10
of
4,251