• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations | Disambiguate

DMCA

Learning to Represent Spatial Transformations with Factored Higher-Order Boltzmann Machines (2010)

Cached

  • Download as a PDF

Download Links

  • [www.cs.toronto.edu]
  • [www.cs.utoronto.ca]
  • [learning.cs.toronto.edu]
  • [www.cs.utoronto.ca]
  • [www.cs.toronto.edu]
  • [www.iro.umontreal.ca]
  • [www.cs.toronto.edu]
  • [learning.cs.toronto.edu]
  • [www.cs.utoronto.ca]
  • [www.iro.umontreal.ca]
  • [www.cs.toronto.edu]
  • [www.cs.utoronto.ca]
  • [www.iro.umontreal.ca]
  • [www.cs.utoronto.ca]
  • [www.cs.toronto.edu]
  • [www.iro.umontreal.ca]

  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Roland Memisevic , Geoffrey E. Hinton
Citations:71 - 18 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@MISC{Memisevic10learningto,
    author = {Roland Memisevic and Geoffrey E. Hinton},
    title = { Learning to Represent Spatial Transformations with Factored Higher-Order Boltzmann Machines},
    year = {2010}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

To allow the hidden units of a restricted Boltzmann machine to model the transformation between two successive images, Memisevic and Hinton (2007) introduced three-way multiplicative interactions that use the intensity of a pixel in the first image as a multiplicative gain on a learned, symmetric weight between a pixel in the second image and a hidden unit. This creates cubically many parameters, which form a three-dimensional interaction tensor. We describe a low-rank approximation to this interaction tensor that uses a sum of factors, each of which is a three-way outer product. This approximation allows efficient learning of transformations between larger image patches. Since each factor can be viewed as an image filter, the model as a whole learns optimal filter pairs for efficiently representing transformations. We demonstrate the learning of optimal filter pairs from various synthetic and real image sequences. We also show how learning about image transformations allows the model to perform a simple visual analogy task, and we show how a completely unsupervised network trained on transformations perceives multiple motions of transparent dot patterns in the same way as humans.

Keyphrases

factored higher-order boltzmann machine    dana ballard learning    represent spatial transformation    optimal filter pair    hidden unit    successive image    image filter    transparent dot pattern    simple visual analogy task    unsupervised network    three-way outer product    first image    second image    three-way multiplicative interaction    low-rank approximation    image patch    interaction tensor    many parameter    multiplicative gain    symmetric weight    real image sequence    restricted boltzmann machine    image transformation    three-dimensional interaction tensor    transformation perceives multiple motion    efficient learning   

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University