Results 1  10
of
77
Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
, 2004
"... Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear m ..."
Abstract

Cited by 1513 (20 self)
 Add to MetaCart
Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear measurements do we need to recover objects from this class to within accuracy ɛ? This paper shows that if the objects of interest are sparse or compressible in the sense that the reordered entries of a signal f ∈ F decay like a powerlaw (or if the coefficient sequence of f in a fixed basis decays like a powerlaw), then it is possible to reconstruct f to within very high accuracy from a small number of random measurements. typical result is as follows: we rearrange the entries of f (or its coefficients in a fixed basis) in decreasing order of magnitude f  (1) ≥ f  (2) ≥... ≥ f  (N), and define the weakℓp ball as the class F of those elements whose entries obey the power decay law f  (n) ≤ C · n −1/p. We take measurements 〈f, Xk〉, k = 1,..., K, where the Xk are Ndimensional Gaussian
Regularization networks and support vector machines
 Advances in Computational Mathematics
, 2000
"... Regularization Networks and Support Vector Machines are techniques for solving certain problems of learning from examples – in particular the regression problem of approximating a multivariate function from sparse data. Radial Basis Functions, for example, are a special case of both regularization a ..."
Abstract

Cited by 366 (38 self)
 Add to MetaCart
(Show Context)
Regularization Networks and Support Vector Machines are techniques for solving certain problems of learning from examples – in particular the regression problem of approximating a multivariate function from sparse data. Radial Basis Functions, for example, are a special case of both regularization and Support Vector Machines. We review both formulations in the context of Vapnik’s theory of statistical learning which provides a general foundation for the learning problem, combining functional analysis and statistics. The emphasis is on regression: classification is treated as a special case.
Optimally Sparse Image Representation by the Easy Path Wavelet Transform
, 2009
"... Abstract The Easy Path Wavelet Transform (EPWT) ..."
(Show Context)
Nonlinear Methods of Approximation
, 2002
"... Our main interest in this paper is nonlinear approximation. The basic idea behind nonlinear approximation is that the elements used in the approximation do not come from a xed linear space but are allowed to depend on the function being approximated. While the scope of this paper is mostly theoretic ..."
Abstract

Cited by 71 (9 self)
 Add to MetaCart
Our main interest in this paper is nonlinear approximation. The basic idea behind nonlinear approximation is that the elements used in the approximation do not come from a xed linear space but are allowed to depend on the function being approximated. While the scope of this paper is mostly theoretical, we should note that this form of approximation appears in many numerical applications such as adaptive PDE solvers, compression of images and signals, statistical classication, and so on. The standard problem in this regard is the problem of mterm approximation where one xes a basis and looks to approximate a target function by a linear combination of m terms of the basis. When the basis is a wavelet basis or a basis of other waveforms, then this type of approximation is the starting point for compression algorithms. We are interested in the quantitative aspects of this type of approximation. Namely, we want to understand the properties (usually smoothness) of the function which govern its rate of approximation in some given norm (or metric). We are also interested in stable algorithms for nding good or near best approximations using m terms. Some of our earlier work has introduced and analyzed such algorithms. More recently, there has emerged another more complicated form of nonlinear approximation which we call highly nonlinear approximation. It takes many forms but has the basic ingredient that a basis is replaced by a larger system of functions that is usually redundant. Some types of approximation that fall into this general category are mathematical frames, adaptive pursuit (or greedy algorithms) and adaptive basis selection. Redundancy on the one hand oers much promise for greater eciency in terms of approximation rate, but on the other hand gives rise to h...
Optimal Representation of Piecewise Hölder Smooth Bivariate Functions by the Easy Path Wavelet Transform
"... The Easy Path Wavelet Transform (EPWT) [22] has recently been proposed by one of the authors as a tool for sparse representations of bivariate functions from discrete data, in particular from image data. The EPWT is a locally adaptive wavelet transform. It works along pathways through the array of f ..."
Abstract

Cited by 68 (3 self)
 Add to MetaCart
The Easy Path Wavelet Transform (EPWT) [22] has recently been proposed by one of the authors as a tool for sparse representations of bivariate functions from discrete data, in particular from image data. The EPWT is a locally adaptive wavelet transform. It works along pathways through the array of function values and it exploits the local correlations of the given data in a simple appropriate manner. Using polyharmonic spline interpolation, we show in this paper that the EPWT leads, for a suitable choice of the pathways, to optimal Nterm approximations for piecewise Hölder smooth functions with singularities along curves. Key words. sparse data representation, wavelet transform along pathways, Nterm approximation AMS Subject classifications. 41A25, 42C40, 68U10, 94A08 1
Vector Greedy Algorithms
"... Our objective is to study nonlinear approximation with regard to redundant systems. Redundancy on the one hand offers much promise for greater efficiency in terms of approximation rate, but on the other hand gives rise to highly nontrivial theoretical and practical problems. Greedy type approximati ..."
Abstract

Cited by 67 (11 self)
 Add to MetaCart
Our objective is to study nonlinear approximation with regard to redundant systems. Redundancy on the one hand offers much promise for greater efficiency in terms of approximation rate, but on the other hand gives rise to highly nontrivial theoretical and practical problems. Greedy type approximations proved to be convenient and efficient ways of constructing mterm approximants. We introduce and study vector greedy algorithms that are designed with aim of constructing mth greedy approximants simultaneously for a given finite number of elements. We prove convergence theorems and obtain some estimates for the rate of convergence of vector greedy algorithms when elements come from certain classes.
Fully adaptive multiresolution finite volume schemes for conservation laws
 Math. Comp
, 2003
"... Abstract. The use of multiresolution decompositions in the context of finite volume schemes for conservation laws was first proposed by A. Harten for the purpose of accelerating the evaluation of numerical fluxes through an adaptive computation. In this approach the solution is still represented at ..."
Abstract

Cited by 56 (15 self)
 Add to MetaCart
(Show Context)
Abstract. The use of multiresolution decompositions in the context of finite volume schemes for conservation laws was first proposed by A. Harten for the purpose of accelerating the evaluation of numerical fluxes through an adaptive computation. In this approach the solution is still represented at each time step on the finest grid, resulting in an inherent limitation of the potential gain in memory space and computational time. The present paper is concerned with the development and the numerical analysis of fully adaptive multiresolution schemes, in which the solution is represented and computed in a dynamically evolved adaptive grid. A crucial problem is then the accurate computation of the flux without the full knowledge of fine grid cell averages. Several solutions to this problem are proposed, analyzed, and compared in terms of accuracy and complexity. 1.
Joint SourceChannel Communication for Distributed Estimation in Sensor Networks
"... Power and bandwidth are scarce resources in dense wireless sensor networks and it is widely recognized that joint optimization of the operations of sensing, processing and communication can result in significant savings in the use of network resources. In this paper, a distributed joint sourcechan ..."
Abstract

Cited by 53 (3 self)
 Add to MetaCart
Power and bandwidth are scarce resources in dense wireless sensor networks and it is widely recognized that joint optimization of the operations of sensing, processing and communication can result in significant savings in the use of network resources. In this paper, a distributed joint sourcechannel communication architecture is proposed for energyefficient estimation of sensor field data at a distant destination and the corresponding relationships between power, distortion, and latency are analyzed as a function of number of sensor nodes. The approach is applicable to a broad class of sensed signal fields and is based on distributed computation of appropriately chosen projections of sensor data at the destination – phasecoherent transmissions from the sensor nodes enable exploitation of the distributed beamforming gain for energy efficiency. Random projections are used when little or no prior knowledge is available about the signal field. Distinct features of the proposed scheme include: 1) processing and communication are combined into one distributed projection operation; 2) it virtually eliminates the need for innetwork processing and communication; 3) given sufficient prior knowledge about the sensed data, consistent estimation is possible with increasing sensor density even with vanishing total network power; and 4) consistent signal estimation is possible with power and latency requirements growing at most sublinearly with the number of sensor nodes even when little or no prior knowledge about the sensed data is assumed at the sensor nodes.
Compressive Sensing
, 2010
"... Compressive sensing is a new type of sampling theory, which predicts that sparse signals and images can be reconstructed from what was previously believed to be incomplete information. As a main feature, efficient algorithms such as ℓ1minimization can be used for recovery. The theory has many poten ..."
Abstract

Cited by 50 (12 self)
 Add to MetaCart
(Show Context)
Compressive sensing is a new type of sampling theory, which predicts that sparse signals and images can be reconstructed from what was previously believed to be incomplete information. As a main feature, efficient algorithms such as ℓ1minimization can be used for recovery. The theory has many potential applications in signal processing and imaging. This chapter gives an introduction and overview on both theoretical and numerical aspects of compressive sensing.
Tree Approximation and Optimal Encoding
 J. Appl. Comp. Harm. Anal
, 2000
"... Tree approximation is a new form of nonlinear approximation which appears naturally in some applications such as image processing and adaptive numerical methods. It is somewhat more restrictive than the usual nterm approximation. We show that the restrictions of tree approximation cost little in ..."
Abstract

Cited by 49 (9 self)
 Add to MetaCart
(Show Context)
Tree approximation is a new form of nonlinear approximation which appears naturally in some applications such as image processing and adaptive numerical methods. It is somewhat more restrictive than the usual nterm approximation. We show that the restrictions of tree approximation cost little in terms of rates of approximation. We then use that result to design encoders for compression. These encoders are universal (they apply to general functions) and progressive (increasing accuracy is obtained by sending bit stream increments). We show optimality of the encoders in the sense that they provide upper estimates for the Kolmogorov entropy of Besov balls. AMS subject classication: 41A25, 41A46, 65F99, 65N12, 65N55. Key Words: compression, nterm approximation, encoding, Kolmogorov entropy . 1 Introduction Wavelets are utilized in many applications including image/signal processing and numerical methods for PDEs. Their usefulness stems in part from the fact that they provide ...