• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 3,989
Next 10 →

Algorithms for Fast Vector Quantization

by Sunil Arya, David M. Mount - Proc. of DCC '93: Data Compression Conference , 1993
"... Nearest neighbor searching is an important geometric subproblem in vector quantization. ..."
Abstract - Cited by 80 (11 self) - Add to MetaCart
Nearest neighbor searching is an important geometric subproblem in vector quantization.

Fast texture synthesis using tree-structured vector quantization

by Li-yi Wei, Marc Levoy , 2000
"... Figure 1: Our texture generation process takes an example texture patch (left) and a random noise (middle) as input, and modifies this random noise to make it look like the given example texture. The synthesized texture (right) can be of arbitrary size, and is perceived as very similar to the given ..."
Abstract - Cited by 561 (12 self) - Add to MetaCart
Field texture models and generates textures through a deterministic searching process. We accelerate this synthesis process using tree-structured vector quantization.

vector quantization

by L. Guillemot, S. Moussaoui, J. M. Moureaux
"... mixture model for the energy of blocks dedicated to ..."
Abstract - Add to MetaCart
mixture model for the energy of blocks dedicated to

Vector Quantization with Complexity Costs

by Joachim Buhmann, Hans Kühnel , 1993
"... Vector quantization is a data compression method where a set of data points is encoded by a reduced set of reference vectors, the codebook. We discuss a vector quantization strategy which jointly optimizes distortion errors and the codebook complexity, thereby, determining the size of the codebook. ..."
Abstract - Cited by 63 (20 self) - Add to MetaCart
Vector quantization is a data compression method where a set of data points is encoded by a reduced set of reference vectors, the codebook. We discuss a vector quantization strategy which jointly optimizes distortion errors and the codebook complexity, thereby, determining the size of the codebook

Multiresolution vector quantization

by Michelle Effros, Diego Dugatkin - IEEE TRANS. INF. THEORY , 2004
"... Multiresolution source codes are data compression algorithms yielding embedded source descriptions. The decoder of a multiresolution code can build a source reproduction by decoding the embedded bit stream in part or in whole. All decoding procedures start at the beginning of the binary source descr ..."
Abstract - Cited by 45 (4 self) - Add to MetaCart
description and decode some fraction of that string. Decoding a small portion of the binary string gives a low-resolution reproduction; decoding more yields a higher resolution reproduction; and so on. Multiresolution vector quantizers are block multiresolution source codes. This paper introduces algorithms

Soft Learning Vector Quantization

by Sambu Seo, Klaus Obermayer - NEURAL COMPUTATION , 2002
"... Learning Vector Quantization is a popular class of adaptive nearest prototype classifiers for multiclass classification, but learning algorithms from this family have so far been proposed on heuristic grounds. Here we take a more principled approach and derive two variants of Learning Vector Quantiz ..."
Abstract - Cited by 55 (0 self) - Add to MetaCart
Learning Vector Quantization is a popular class of adaptive nearest prototype classifiers for multiclass classification, but learning algorithms from this family have so far been proposed on heuristic grounds. Here we take a more principled approach and derive two variants of Learning Vector

Generalized learning vector quantization

by Atsushi Sato, Keiji Yamada - Hasselmo (Eds.), NIPS , 1995
"... We propose a new learning method, "Generalized Learning Vec-tor Quantization (GLVQ), " in which reference vectors are updated based on the steepest descent method in order to minimize the cost function. The cost function is determined so that the obtained learning rule satisfies the conver ..."
Abstract - Cited by 123 (0 self) - Add to MetaCart
We propose a new learning method, "Generalized Learning Vec-tor Quantization (GLVQ), " in which reference vectors are updated based on the steepest descent method in order to minimize the cost function. The cost function is determined so that the obtained learning rule satisfies

Distortion-Limited Vector Quantization

by Peter J. Hahn, V. John Mathews - in Proc.Data Compression Conf. - DCC ’96 , 1996
"... This paper presents a vector quantization system that limits the maximum distortion introduced to a pre-selected threshold value. This system uses a recently introduced variation of the L1 distortion measure that attempts to minimize the occurrences of quantization errors above a preselected thresho ..."
Abstract - Cited by 3 (1 self) - Add to MetaCart
This paper presents a vector quantization system that limits the maximum distortion introduced to a pre-selected threshold value. This system uses a recently introduced variation of the L1 distortion measure that attempts to minimize the occurrences of quantization errors above a preselected

Kernelizing Vector Quantization Algorithms

by Matthieu Geist, Olivier Pietquin, Gabriel Fricout
"... Abstract. The kernel trick is a well known approach allowing to implicitly cast a linear method into a nonlinear one by replacing any dot product by a kernel function. However few vector quantization algorithms have been kernelized. Indeed, they usually imply to compute linear transformations (e.g. ..."
Abstract - Add to MetaCart
Abstract. The kernel trick is a well known approach allowing to implicitly cast a linear method into a nonlinear one by replacing any dot product by a kernel function. However few vector quantization algorithms have been kernelized. Indeed, they usually imply to compute linear transformations (e

Fractal Dimension and Vector Quantization

by Krishna Kumaraswamy, Vasileios Megalooikonomou, Christos Faloutsos , 2004
"... We show that the performance of a vector quantizer for a self-similar data set is related to the intrinsic ("fractal") dimension of the data set. We derive a formula for predicting the error-rate, given the fractal dimension and discuss how we can use our result for evaluating the performa ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
We show that the performance of a vector quantizer for a self-similar data set is related to the intrinsic ("fractal") dimension of the data set. We derive a formula for predicting the error-rate, given the fractal dimension and discuss how we can use our result for evaluating
Next 10 →
Results 1 - 10 of 3,989
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University