Results 11 - 20
of
191
Recursive Deep Models for Discourse Parsing
"... Text-level discourse parsing remains a challenge: most approaches employ fea-tures that fail to capture the intentional, se-mantic, and syntactic aspects that govern discourse coherence. In this paper, we pro-pose a recursive model for discourse pars-ing that jointly models distributed repre-sentati ..."
Abstract
-
Cited by 9 (2 self)
- Add to MetaCart
(Show Context)
Text-level discourse parsing remains a challenge: most approaches employ fea-tures that fail to capture the intentional, se-mantic, and syntactic aspects that govern discourse coherence. In this paper, we pro-pose a recursive model for discourse pars-ing that jointly models distributed repre-sentations for clauses, sentences, and en-tire discourses. The learned representa-tions can to some extent learn the seman-tic and intentional import of words and larger discourse units automatically,. The proposed framework obtains comparable performance regarding standard discours-ing parsing evaluations when compared against current state-of-art systems. 1
Existence of V (m, t) vectors
- J. Statist. Plann. Inference
"... Local resampling for patch-based texture synthesis in ..."
Abstract
-
Cited by 8 (3 self)
- Add to MetaCart
(Show Context)
Local resampling for patch-based texture synthesis in
A Multiplicative Model for Learning Distributed Text-Based Attribute Representations
"... In this paper we propose a general framework for learning distributed represen-tations of attributes: characteristics of text whose representations can be jointly learned with word embeddings. Attributes can correspond to a wide variety of concepts, such as document indicators (to learn sentence vec ..."
Abstract
-
Cited by 8 (1 self)
- Add to MetaCart
(Show Context)
In this paper we propose a general framework for learning distributed represen-tations of attributes: characteristics of text whose representations can be jointly learned with word embeddings. Attributes can correspond to a wide variety of concepts, such as document indicators (to learn sentence vectors), language in-dicators (to learn distributed language representations), meta-data and side infor-mation (such as the age, gender and industry of a blogger) or representations of authors. We describe a third-order model where word context and attribute vectors interact multiplicatively to predict the next word in a sequence. This leads to the notion of conditional word similarity: how meanings of words change when con-ditioned on different attributes. We perform several experimental tasks including sentiment classification, cross-lingual document classification, and blog author-ship attribution. We also qualitatively evaluate conditional word neighbours and attribute-conditioned text generation. 1
Bilingually-constrained phrase embeddings for machine translation.
- In ACL
, 2014
"... Abstract We propose Bilingually-constrained Recursive Auto-encoders (BRAE) to learn semantic phrase embeddings (compact vector representations for phrases), which can distinguish the phrases with different semantic meanings. The BRAE is trained in a way that minimizes the semantic distance of trans ..."
Abstract
-
Cited by 8 (0 self)
- Add to MetaCart
(Show Context)
Abstract We propose Bilingually-constrained Recursive Auto-encoders (BRAE) to learn semantic phrase embeddings (compact vector representations for phrases), which can distinguish the phrases with different semantic meanings. The BRAE is trained in a way that minimizes the semantic distance of translation equivalents and maximizes the semantic distance of nontranslation pairs simultaneously. After training, the model learns how to embed each phrase semantically in two languages and also learns how to transform semantic embedding space in one language to the other. We evaluate our proposed method on two end-to-end SMT tasks (phrase table pruning and decoding with phrasal semantic similarities) which need to measure semantic similarity between a source phrase and its translation candidates. Extensive experiments show that the BRAE is remarkably effective in these two tasks.
Modeling deep temporal dependencies with recurrent grammar cells
- In Advances in Neural Information Processing Systems 27
, 2014
"... We propose modeling time series by representing the transformations that take a frame at time t to a frame at time t+1. To this end we show how a bi-linear model of transformations, such as a gated autoencoder, can be turned into a recurrent net-work, by training it to predict future frames from the ..."
Abstract
-
Cited by 7 (0 self)
- Add to MetaCart
(Show Context)
We propose modeling time series by representing the transformations that take a frame at time t to a frame at time t+1. To this end we show how a bi-linear model of transformations, such as a gated autoencoder, can be turned into a recurrent net-work, by training it to predict future frames from the current one and the inferred transformation using backprop-through-time. We also show how stacking multi-ple layers of gating units in a recurrent pyramid makes it possible to represent the ”syntax ” of complicated time series, and that it can outperform standard recurrent neural networks in terms of prediction accuracy on a variety of tasks. 1
Factor-based compositional embedding models
- In NIPS Workshop on Learning Semantics
, 2014
"... Introduction Word embeddings, which are distributed word representations learned by neural language models [1, 2, 3], have been shown to be powerful word representations. They have been successfully applied to a range of NLP tasks, including syntax [2, 4, 5] and semantics [6, 7, 8]. Information abou ..."
Abstract
-
Cited by 7 (2 self)
- Add to MetaCart
Introduction Word embeddings, which are distributed word representations learned by neural language models [1, 2, 3], have been shown to be powerful word representations. They have been successfully applied to a range of NLP tasks, including syntax [2, 4, 5] and semantics [6, 7, 8]. Information about language structure is critical in many NLP tasks, where substructures of a sen-tence and its annotations inform downstream NLP task. Yet word representations alone do not
Adaptive Recursive Neural Network for Target-dependent Twitter Sentiment Classification
"... We propose Adaptive Recursive Neural Network (AdaRNN) for target-dependent Twitter sentiment classification. AdaRNN adaptively propagates the sentiments of words to target depending on the context and syntactic relationships between them. It consists of more than one composition functions, and we mo ..."
Abstract
-
Cited by 7 (0 self)
- Add to MetaCart
(Show Context)
We propose Adaptive Recursive Neural Network (AdaRNN) for target-dependent Twitter sentiment classification. AdaRNN adaptively propagates the sentiments of words to target depending on the context and syntactic relationships between them. It consists of more than one composition functions, and we model the adaptive sen-timent propagations as distributions over these composition functions. The experi-mental studies illustrate that AdaRNN im-proves the baseline methods. Further-more, we introduce a manually annotated dataset for target-dependent Twitter senti-ment analysis. 1
Improving Lexical Embeddings with Semantic Knowledge
"... Word embeddings learned on unlabeled data are a popular tool in semantics, but may not capture the desired semantics. We propose a new learning objective that in-corporates both a neural language model objective (Mikolov et al., 2013) and prior knowledge from semantic resources to learn improved lex ..."
Abstract
-
Cited by 7 (1 self)
- Add to MetaCart
(Show Context)
Word embeddings learned on unlabeled data are a popular tool in semantics, but may not capture the desired semantics. We propose a new learning objective that in-corporates both a neural language model objective (Mikolov et al., 2013) and prior knowledge from semantic resources to learn improved lexical semantic embed-dings. We demonstrate that our embed-dings improve over those learned solely on raw text in three settings: language mod-eling, measuring semantic similarity, and predicting human judgements.
Max-margin tensor neural network for chinese word segmentation
- In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers
, 2014
"... Abstract Recently, neural network models for natural language processing tasks have been increasingly focused on for their ability to alleviate the burden of manual feature engineering. In this paper, we propose a novel neural network model for Chinese word segmentation called Max-Margin Tensor Neu ..."
Abstract
-
Cited by 7 (0 self)
- Add to MetaCart
(Show Context)
Abstract Recently, neural network models for natural language processing tasks have been increasingly focused on for their ability to alleviate the burden of manual feature engineering. In this paper, we propose a novel neural network model for Chinese word segmentation called Max-Margin Tensor Neural Network (MMTNN). By exploiting tag embeddings and tensorbased transformation, MMTNN has the ability to model complicated interactions between tags and context characters. Furthermore, a new tensor factorization approach is proposed to speed up the model and avoid overfitting. Experiments on the benchmark dataset show that our model achieves better performances than previous neural network models and that our model can achieve a competitive performance with minimal feature engineering. Despite Chinese word segmentation being a specific case, MMTNN can be easily generalized and applied to other sequence labeling tasks.
Deep recursive neural networks for compositionality in language
- In Proceedings of NIPS
, 2014
"... Recursive neural networks comprise a class of architecture that can operate on structured input. They have been previously successfully applied to model com-positionality in natural language using parse-tree-based structural representations. Even though these architectures are deep in structure, the ..."
Abstract
-
Cited by 6 (0 self)
- Add to MetaCart
(Show Context)
Recursive neural networks comprise a class of architecture that can operate on structured input. They have been previously successfully applied to model com-positionality in natural language using parse-tree-based structural representations. Even though these architectures are deep in structure, they lack the capacity for hierarchical representation that exists in conventional deep feed-forward networks as well as in recently investigated deep recurrent neural networks. In this work we introduce a new architecture — a deep recursive neural network (deep RNN) — constructed by stacking multiple recursive layers. We evaluate the proposed model on the task of fine-grained sentiment classification. Our results show that deep RNNs outperform associated shallow counterparts that employ the same number of parameters. Furthermore, our approach outperforms previous baselines on the sentiment analysis task, including a multiplicative RNN variant as well as the re-cently introduced paragraph vectors, achieving new state-of-the-art results. We provide exploratory analyses of the effect of multiple layers and show that they capture different aspects of compositionality in language. 1