• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Improving lexical embeddings with semantic knowledge. (2014)

by M Yu, M Dredze
Add To MetaCart

Tools

Sorted by:
Results 1 - 7 of 7

Retrofitting word vectors to semantic lexicons

by Manaal Faruqui, Jesse Dodge, Sujay K. Jauhar, Chris Dyer, Eduard Hovy, Noah A. Smith - In Proceedings of the 2015 Conference of NAACL , 2015
"... Vector space word representations are typically learned using only co-occurrence statistics from text corpora. Although such statistics are informative, they disre-gard easily accessible (and often carefully curated) information archived in se-mantic lexicons such as WordNet, FrameNet, and the Parap ..."
Abstract - Cited by 6 (0 self) - Add to MetaCart
Vector space word representations are typically learned using only co-occurrence statistics from text corpora. Although such statistics are informative, they disre-gard easily accessible (and often carefully curated) information archived in se-mantic lexicons such as WordNet, FrameNet, and the Paraphrase Database. This paper proposes a technique to leverage both distributional and lexicon-derived ev-idence to obtain better representations. We run belief propagation on a word type graph constructed from word similarity information from lexicons to encourage connected (related) words to have similar representations, and also to be close to the unsupervised vectors. Evaluated on a battery of standard lexical semantic evaluation tasks in several languages, using several different underlying word vec-tor models, we obtain substantially improved vectors and consistently outperform existing approaches of incorporating semantic knowledge in word vectors. 1
(Show Context)

Citation Context

...5.2 44.8 54.7 73.3 40.8 74.1 Yu & Dredze (2014) 50.1 47.1 53.7 61.3 29.9 71.5 Retrofitting 60.5 57.7 58.4 81.3 52.5 75.7 Table 2: Semantic enrichment of word2vec CBOW vectors using Yu & Dredze (2014) =-=[35]-=- and retrofitting model using PPDB. Spearman’s correlation (3 left columns) and accuracy (3 right columns) on different tasks. Bold indicates best result across all vector types. Language, Task Origin...

PPDB 2.0: Better paraphrase ranking, fine-grained entailment relations, word embeddings, and style classification

by Ellie Pavlick, Pushpendre Rastogi, Juri Ganitkevitch, Benjamin Van Durme, Chris Callison-burch
"... We present a new release of the Para-phrase Database. PPDB 2.0 includes a discriminatively re-ranked set of para-phrases that achieve a higher correlation with human judgments than PPDB 1.0’s heuristic rankings. Each paraphrase pair in the database now also includes fine-grained entailment relations ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
We present a new release of the Para-phrase Database. PPDB 2.0 includes a discriminatively re-ranked set of para-phrases that achieve a higher correlation with human judgments than PPDB 1.0’s heuristic rankings. Each paraphrase pair in the database now also includes fine-grained entailment relations, word embed-ding similarities, and style annotations. 1

Named Entity Recognition for Chinese Social Media with Jointly Trained Embeddings

by Nanyun Peng, Mark Dredze
"... We consider the task of named entity recognition for Chinese social media. The long line of work in Chinese NER has fo-cused on formal domains, and NER for social media has been largely restricted to English. We present a new corpus of Weibo messages annotated for both name and nominal mentions. Add ..."
Abstract - Cited by 1 (1 self) - Add to MetaCart
We consider the task of named entity recognition for Chinese social media. The long line of work in Chinese NER has fo-cused on formal domains, and NER for social media has been largely restricted to English. We present a new corpus of Weibo messages annotated for both name and nominal mentions. Additionally, we evaluate three types of neural embeddings for representing Chinese text. Finally, we propose a joint training objective for the embeddings that makes use of both (NER) labeled and unlabeled raw text. Our meth-ods yield a 9 % improvement over a state-of-the-art baseline. 1

Convolutional Neural Network Based Semantic Tagging with Entity Embeddings

by Asli Celikyilmaz , Dilek Hakkani-Tur
"... Abstract Unsupervised word embeddings provide rich linguistic and conceptual information about words. However, they may provide weak information about domain specific semantic relations for certain tasks such as semantic parsing of natural language queries, where such information about words or phr ..."
Abstract - Add to MetaCart
Abstract Unsupervised word embeddings provide rich linguistic and conceptual information about words. However, they may provide weak information about domain specific semantic relations for certain tasks such as semantic parsing of natural language queries, where such information about words or phrases can be valuable. To encode the prior knowledge about the semantic word relations, we extended the neural network based lexical word embedding objective function by incorporating the information about relationship between entities that we extract from knowledge bases
(Show Context)

Citation Context

...log likelihood of each token given its context, namely, neighboring words within window size c: max 1 T T∑ t=1 log p(wt|wt+ct−c) (1) where wt+ct−c is the set of words in the window of size c centered at wt (wt included). Using the continuous bag of words model (CBOW) of word2vec, p(wt|wt+ct−c) predicts the current word based on the context as follows: p(wt|wt+ct−c) := exp ( eywt T . ∑ −c≤j≤c,j 6=0 ewt+j ) ∑ w exp ( eyw T . ∑ −c≤j≤c,j 6=0 ewt+j ) (2) In Eq. (2) ew and eyw represent input and output embeddings, the scalar vector representations of each word w. Relational Constrained Model (RTM) [13]. Learns embeddings to predict one word from another related word. Suppose we are given a list of synonymous or paraphrases of N words based on a knowledge source (e.g., Wordnet). RTM learns the word embeddings based on the paraphrase relation between the words. Thus, they introduce priors as paraphrases encoding synonymy relations such as ”analog”∼”analogue” or ”bolt”∼”screw”. They change the objective function of the word2vec by dropping the context and learn the embeddings on the paraphrase data. The objective to maximize is the sum over all the words in the vocabulary, which is similar to ...

Component-Enhanced Chinese Character Embeddings

by Yanran Li, Wenjie Li, Fei Sun, Sujian Li
"... Distributed word representations are very useful for capturing semantic information and have been successfully applied in a variety of NLP tasks, especially on En-glish. In this work, we innovatively de-velop two component-enhanced Chinese character embedding models and their bi-gram extensions. Dis ..."
Abstract - Add to MetaCart
Distributed word representations are very useful for capturing semantic information and have been successfully applied in a variety of NLP tasks, especially on En-glish. In this work, we innovatively de-velop two component-enhanced Chinese character embedding models and their bi-gram extensions. Distinguished from En-glish word embeddings, our models ex-plore the compositions of Chinese char-acters, which often serve as semantic in-dictors inherently. The evaluations on both word similarity and text classification demonstrate the effectiveness of our mod-els. 1
(Show Context)

Citation Context

..., a line of research deploys the order information of the words in the contexts by either deriving the contexts using dependency relations where the target word participates (Levy and Goldberg, 2014; =-=Yu and Dredze, 2014-=-; Bansal et al, 2014) or directly keeping the order features (Ling et al, 2015). As to another line, Luong et al (2013) captures morphological composition by using neural networks and Qiu et al (2014)...

AKNET: A General Framework for Learning Word Embedding using Morphological Knowledge

by Qing Cui, Bin Gao, Siyu Qiu
"... Neural network techniques are widely applied to obtain high-quality distributed representations of words, i.e., word embeddings, to address text mining, information retrieval, and natural language processing tasks. Recently, efficient methods have been proposed to learn word embeddings from context ..."
Abstract - Add to MetaCart
Neural network techniques are widely applied to obtain high-quality distributed representations of words, i.e., word embeddings, to address text mining, information retrieval, and natural language processing tasks. Recently, efficient methods have been proposed to learn word embeddings from context that captures both semantic and syntactic relationships between words. However, it is challenging to handle unseen words or rare words with insufficient context. In this paper, inspired by the study on word recognition process in cognitive psychology, we propose to take advantage of seemingly less obvious but essentially important morphological knowledge to address these challenges. In particular, we introduce a novel neural network architecture called KNET that leverages both contextual information and morphological word similarity built based on morphological knowledge to learn word embeddings. Meanwhile, the learning architecture is also able to refine the pre-defined morphological knowledge and obtain more accurate word similarity. Experiments on an analogical reasoning task and a word similarity task both demonstrate that the proposed KNET framework can greatly enhance the effectiveness of word embeddings. ACM Reference Format:
(Show Context)

Citation Context

..., Qiu et al. [Qiu et al. 2014] introduced a co-learning framework to produce both the word representation and the morpheme representation such that each of them can be mutually reinforced. Yu et al. [=-=Yu and Dredze 2014-=-] proposed a new learning objective that integrates both a neural language model objective and a semantic prior knowledge objective which can result in better word embedding for semantic tasks. Moreov...

Topics, Trends, and Resources in Natural Language Processing (NLP)

by Mohit Bansal
"... ..."
Abstract - Add to MetaCart
Abstract not found
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University