Results 1 - 10
of
3,419
AN OBJECT MODEL FOR NATURAL LANGUAGE DOCUMENT
"... Abstract. An object model of a text is de ned, and structural and manipulative aspects of the model are presented. The model proposed supports text in its general sence (as a natural language document), which comprises structural, syntactic, semantic, orthographic, stylistic properties of both the u ..."
Abstract
- Add to MetaCart
Abstract. An object model of a text is de ned, and structural and manipulative aspects of the model are presented. The model proposed supports text in its general sence (as a natural language document), which comprises structural, syntactic, semantic, orthographic, stylistic properties of both
Sentiment Mining for Natural Language Documents
"... The wide variety of possible applications for sentiment mining has made it the focus of considerable research in recent years. High accuracy classification has been achieved by using a variety of techniques, most of which are heavily reliant on machine learning. At its core, sentiment mining involve ..."
Abstract
- Add to MetaCart
Natural Language Processing (NLP) techniques to the problem of accurately classifying the sentiment a document contains relating to a certain subject. Common NLP methods that are implemented in pursuit this goal include part of speech tagging, grammatical structure parsing, and coreference resolution
Automatic generation of natural language documentation from statecharts
, 2000
"... The process of documenting designs is tedious and often error-prone. We discuss a system that automatically generates documentation for the single step transition behavior of Statecharts with particular focus on the correctness of the result in the sense that the document will present all and only t ..."
Abstract
-
Cited by 3 (1 self)
- Add to MetaCart
the facts corresponding to the design being documented. Our approach is to translate the Statechart into a propositional formula, then translate this formula into a natural language report. In the later translation, pragmatic e ects arise due to the way the information is presented. Whereas such e ects can
Hierarchical Meaning Representation and Analysis of Natural Language Documents
"... This paper attempts to systematize natural language analysis process by (1) use of a partitioned semantic network formalism as the meaning representation and (2) stepwise translation based on Montague Grammar. The meaning representation is obtained in two steps. The first step translates natural lan ..."
Abstract
- Add to MetaCart
This paper attempts to systematize natural language analysis process by (1) use of a partitioned semantic network formalism as the meaning representation and (2) stepwise translation based on Montague Grammar. The meaning representation is obtained in two steps. The first step translates natural
Automated extraction and validation of security policies from natural-language documents
, 2011
"... As one of the most fundamental security mechanisms of resources, Access Control Policies (ACP) specify which principals such as users or processes have access to which resources. Ensuring the correct specification and enforcement of ACPs is crucial to prevent security vulnerabilities. However, in pr ..."
Abstract
-
Cited by 1 (1 self)
- Add to MetaCart
, in practice, ACPs are commonly written in Natural Language (NL) and buried in large documents such as requirements documents, not directly checkable for correctness. It is very tedious and error-prone to manually identify and extract ACPs from these NL documents, and validate NL functional requirements
Comparing natural language documents: a DL based approach
"... We propose a method to compare semantically two natural language texts. The process is realized in two steps, the first translates the texts into description logics terminologies. The second computes the difference between the terminologies obtained. We show how the best covering problem can be used ..."
Abstract
- Add to MetaCart
We propose a method to compare semantically two natural language texts. The process is realized in two steps, the first translates the texts into description logics terminologies. The second computes the difference between the terminologies obtained. We show how the best covering problem can
Using Style Markers for Detecting Plagiarism in Natural Language Documents
, 2003
"... Most of the existing plagiarism detection systems compare a text to a database of other texts. These external approaches, however, are vulnerable because texts not contained in the database cannot be detected as source texts. This paper examines an internal plagiarism detection method that uses styl ..."
Abstract
- Add to MetaCart
Most of the existing plagiarism detection systems compare a text to a database of other texts. These external approaches, however, are vulnerable because texts not contained in the database cannot be detected as source texts. This paper examines an internal plagiarism detection method that uses style markers from authorship attribution studies in order to find stylistic changes in a text. These changes might pinpoint plagiarized passages. Additionally, a new style marker called "specific words" is introduced. A pre-study tests if the style markers can "fingerprint" an author's style and if they are constant with sample size. It is shown that vocabulary richness measures do not fulfil these prerequisites. The other style markers - simple ratio measures, readability scores, frequency lists, and entropy measures - have these characteristics and are, together with the new specific words measure, used in a main study with an unsupervised approach for detecting stylistic changes in plagiarized texts at sentence and paragraph levels. It is shown that at these small levels the style markers generally cannot detect plagiarized sections because of intra-authorial stylistic variations (i.e. noise), and that at bigger levels the results are strongly a#ected by the sliding window approach. The specific words measure, however, can pinpoint single sentences written by another author.
Sentiment Mining for Natural Language Documents Wikipedia-based Text Classification
, 2009
"... The Internet and even our personal computers are full of unlabeled text content and we could benefit from tools that automatically organize and tag this text content so that we can quickly access the appropriate content when needed. Supervised learning algorithms for text classification that learn f ..."
Abstract
- Add to MetaCart
The Internet and even our personal computers are full of unlabeled text content and we could benefit from tools that automatically organize and tag this text content so that we can quickly access the appropriate content when needed. Supervised learning algorithms for text classification that learn from labeled training examples are already available, however they rely on future data being classified to bear similarity to the training data for good performance and they often require a large amount of labeled training data. One major problem with supervised learning approaches is that they cannot easily generalize to terms that they have not seen before. So over time as new terms (like person names) enter the public lexicon, approaches trained on older data may have trouble generalizing to these new terms, which often bear a lot of informational content for purposes of text classification. However, there are rich knowledge sources like Wikipedia that are frequently updated and which carry rich category information about text terms like person names, organizations, and geographical locations, collectively termed named entities. This report investigates whether a knowledge-based approach to text classification that leverages Wikipedia's knowledge of named entities can outperform various supervised learning algorithms. The results show that the algorithm has some advantages to the current supervised learning algorithms, especially when little training data is available, but also a few weaknesses that need further improvement
From natural language documents to sharable product knowledge: A knowledge engineering approach
, 1997
"... A great part of the product knowledge in manufacturing enterprises is only available in the form of natural language documents. The know-how recorded in these documents is an essential resource for successful competition in the market. From the viewpoint of knowledge management, however, documents h ..."
Abstract
-
Cited by 10 (4 self)
- Add to MetaCart
A great part of the product knowledge in manufacturing enterprises is only available in the form of natural language documents. The know-how recorded in these documents is an essential resource for successful competition in the market. From the viewpoint of knowledge management, however, documents
Results 1 - 10
of
3,419