Results 1 - 10
of
227
WordNet: An on-line lexical database
- International Journal of Lexicography
, 1990
"... WordNet is an on-line lexical reference system whose design is inspired by current ..."
Abstract
-
Cited by 1976 (10 self)
- Add to MetaCart
WordNet is an on-line lexical reference system whose design is inspired by current
Interpretation as Abduction
, 1990
"... An approach to abductive inference developed in the TACITUS project has resulted in a dramatic simplification of how the problem of interpreting texts is conceptualized. Its use in solving the local pragmatics problems of reference, compound nominals, syntactic ambiguity, and metonymy is described ..."
Abstract
-
Cited by 687 (38 self)
- Add to MetaCart
An approach to abductive inference developed in the TACITUS project has resulted in a dramatic simplification of how the problem of interpreting texts is conceptualized. Its use in solving the local pragmatics problems of reference, compound nominals, syntactic ambiguity, and metonymy is described and illustrated. It also suggests an elegant and thorough integration of syntax, semantics, and pragmatics. 1
Using the web to obtain frequencies for unseen bigrams
- COMPUT. LINGUIST
, 2003
"... This paper shows that the web can be employed to obtain frequencies for bigrams that are unseen in a given corpus. We describe a method for retrieving counts for adjective-noun, noun-noun, and verb-object bigrams from the web by querying a search engine. We evaluate this method by demonstrating: (a) ..."
Abstract
-
Cited by 171 (2 self)
- Add to MetaCart
This paper shows that the web can be employed to obtain frequencies for bigrams that are unseen in a given corpus. We describe a method for retrieving counts for adjective-noun, noun-noun, and verb-object bigrams from the web by querying a search engine. We evaluate this method by demonstrating: (a) a high correlation between web frequencies and corpus frequencies; (b) a reliable correlation between web frequencies and plausibility judgments; (c) a reliable correlation between web frequencies and frequencies recreated using class-based smoothing; (d) a good performance of web frequencies in a pseudo-disambiguation task.
Revision-Based Generation of Natural Language Summaries Providing Historical Background -- Corpus-Based Analysis, Design, Implementation and Evaluation
, 1994
"... Automatically summarizing vast amounts of on-line quantitative data with a short natural language paragraph has a wide range of real-world applications. However, this specific task raises a number of difficult issues that are quite distinct from the generic task of language generation: conciseness, ..."
Abstract
-
Cited by 111 (6 self)
- Add to MetaCart
Automatically summarizing vast amounts of on-line quantitative data with a short natural language paragraph has a wide range of real-world applications. However, this specific task raises a number of difficult issues that are quite distinct from the generic task of language generation: conciseness, complex sentences, floating concepts, historical background, paraphrasing power and implicit content. In this thesis, I address these specific issues by proposing a new generation model in which a first pass builds a draft containing only the essential new facts to report and a second pass incrementally revises this draft to opportunistically add as many background facts as can fit within the space limit. This model requires a new type of linguistic knowledge: revision operations, which specifyies the various ways a draft can...
Designing Statistical Language Learners: Experiments on Noun Compounds
, 1995
"... Statistical language learning research takes the view that many traditional natural language processing tasks can be solved by training probabilistic models of language on a sufficient volume of training data. The design of statistical language learners therefore involves answering two questions: (i ..."
Abstract
-
Cited by 95 (0 self)
- Add to MetaCart
Statistical language learning research takes the view that many traditional natural language processing tasks can be solved by training probabilistic models of language on a sufficient volume of training data. The design of statistical language learners therefore involves answering two questions: (i) Which of the multitude of possible language models will most accurately reflect the properties necessary to a given task? (ii) What will constitute a sufficient volume of training data? Regarding the first question, though a variety of successful models have been discovered, the space of possible designs remains largely unexplored. Regarding the second, exploration of the design space has so far proceeded without an adequate answer. The goal of this thesis is to advance the exploration of the statistical language learning design space. In pursuit of that goal, the thesis makes two main theoretical contributions: it identifies a new class of designs by providing a novel theory of statistical natural language processing, and it presents the foundations for a predictive theory of data requirements to assist in future design explorations. The first of these contributions is called the meaning distributions theory. This theory
An overview of surge: a reusable comprehensive syntactic realization component
- In INLG ’96 Demonstrations and Posters
, 1996
"... j r~di. ufpe. br ..."
(Show Context)
Overextension of conjunctive concepts: Evidence for a unitary model of concept typicality and class inclusion
- Journal of Experimental Psychology: Learning, Memory, and Cognition
, 1988
"... Four experiments investigated how people judge both the typicality and membership of items in conjunctive concepts such as school furniture or sports which are games. Judgments of membership in conjunctions were overextended, and there was asymmetry between the constituent concepts in their influenc ..."
Abstract
-
Cited by 76 (12 self)
- Add to MetaCart
(Show Context)
Four experiments investigated how people judge both the typicality and membership of items in conjunctive concepts such as school furniture or sports which are games. Judgments of membership in conjunctions were overextended, and there was asymmetry between the constituent concepts in their influence on relative conjunctive concept membership. The results are discussed in the light of recent theoretical disputes about the modeling of concept representations and the process of forming
Automatic discovery of part-whole relations
- Computational Linguistics
, 2006
"... Language Computer Corporation An important problem in knowledge discovery from text is the automatic extraction of semantic relations. This paper presents a supervised, semantically intensive, domain independent approach for the automatic detection of part–whole relations in text. First an algorithm ..."
Abstract
-
Cited by 72 (6 self)
- Add to MetaCart
(Show Context)
Language Computer Corporation An important problem in knowledge discovery from text is the automatic extraction of semantic relations. This paper presents a supervised, semantically intensive, domain independent approach for the automatic detection of part–whole relations in text. First an algorithm is described that identifies lexico-syntactic patterns that encode part–whole relations. A difficulty is that these patterns also encode other semantic relations, and a learning method is necessary to discriminate whether or not a pattern contains a part–whole relation. A large set of training examples have been annotated and fed into a specialized learning system that learns classification rules. The rules are learned through an iterative semantic specialization (ISS) method applied to noun phrase constituents. Classification rules have been generated this way for different patterns such as genitives, noun compounds, and noun phrases containing prepositional phrases to extract part–whole relations from them. The applicability of these rules has been tested on a test corpus obtaining an overall average precision of 80.95 % and recall of 75.91%. The results demonstrate the importance of word sense disambiguation for this task. They also demonstrate that different lexico-syntactic patterns encode different semantic information and should be treated separately in the sense that different clarification rules apply to different patterns. 1.
Semi-Automatic Recognition of Noun Modifier Relationships
, 1998
"... Semantic relationships among words and phrases are often marked by explicit syntactic or lexical clues that help recognize such relationships in texts. Within complex nominals, however, few overt clues are available. Systems that analyze such nominals must compensate for the lack of surface cl ..."
Abstract
-
Cited by 64 (5 self)
- Add to MetaCart
Semantic relationships among words and phrases are often marked by explicit syntactic or lexical clues that help recognize such relationships in texts. Within complex nominals, however, few overt clues are available. Systems that analyze such nominals must compensate for the lack of surface clues with other information. One way is to load the system with lexical semantics for nouns or adjectives. This merely shifts the problem elsewhere: how do we define the lexical se- mantics and build large semantic lexicons? Another way is to find constructions similar to a given complex nominal, for which the relationships are already known. This is the way we chose, but it too has drawbacks.
Efficient Creativity: Constraint-Guided Conceptual Combination
- Cognitive Science
, 2000
"... This paper describes a theory that explains both the creativity and the efficiency of people's conceptual combination. In the constraint theory, conceptual combination is controlled by three constraints of diagnosticity, plausibility, and informativeness. The constraints derive from the prag ..."
Abstract
-
Cited by 57 (9 self)
- Add to MetaCart
(Show Context)
This paper describes a theory that explains both the creativity and the efficiency of people's conceptual combination. In the constraint theory, conceptual combination is controlled by three constraints of diagnosticity, plausibility, and informativeness. The constraints derive from the pragmatics of communication as applied to compound phrases. The creativity of combination arises because the constraints can be satisfied in many different ways. The constraint theory yields an algorithmic model of the efficiency of combination. The C model admits the full creativity of combination and yet efficiently settles on the best interpretation for a given phrase. The constraint theory explains many empirical regularities in conceptual combination, and makes various empirically verified predictions. In computer simulations of compound phrase interpretation, the C model has produced results in general agreement with people's responses to the same phrases