Results 1  10
of
75,930
Localized Jacobian ILU preconditioners for hydraulic fractures
"... We discuss the properties of a class of sparse localized approximations to the Jacobian operator that arises in modelling the evolution of a hydraulically driven fracture in a multilayered elastic medium. The governing equations involve a highly nonlinear coupled system of integropartial differen ..."
Abstract

Cited by 3 (0 self)
 Add to MetaCart
We discuss the properties of a class of sparse localized approximations to the Jacobian operator that arises in modelling the evolution of a hydraulically driven fracture in a multilayered elastic medium. The governing equations involve a highly nonlinear coupled system of integro
Globally Consistent Range Scan Alignment for Environment Mapping
 AUTONOMOUS ROBOTS
, 1997
"... A robot exploring an unknown environmentmay need to build a world model from sensor measurements. In order to integrate all the frames of sensor data, it is essential to align the data properly. An incremental approach has been typically used in the past, in which each local frame of data is alig ..."
Abstract

Cited by 536 (8 self)
 Add to MetaCart
A robot exploring an unknown environmentmay need to build a world model from sensor measurements. In order to integrate all the frames of sensor data, it is essential to align the data properly. An incremental approach has been typically used in the past, in which each local frame of data
A Tutorial on Visual Servo Control
 IEEE Transactions on Robotics and Automation
, 1996
"... This paper provides a tutorial introduction to visual servo control of robotic manipulators. Since the topic spans many disciplines our goal is limited to providing a basic conceptual framework. We begin by reviewing the prerequisite topics from robotics and computer vision, including a brief review ..."
Abstract

Cited by 822 (25 self)
 Add to MetaCart
This paper provides a tutorial introduction to visual servo control of robotic manipulators. Since the topic spans many disciplines our goal is limited to providing a basic conceptual framework. We begin by reviewing the prerequisite topics from robotics and computer vision, including a brief review of coordinate transformations, velocity representation, and a description of the geometric aspects of the image formation process. We then present a taxonomy of visual servo control systems. The two major classes of systems, positionbased and imagebased systems, are then discussed. Since any visual servo system must be capable of tracking image features in a sequence of images, we include an overview of featurebased and correlationbased methods for tracking. We conclude the tutorial with a number of observations on the current directions of the research field of visual servo control. 1 Introduction Today there are over 800,000 robots in the world, mostly working in factory environment...
A New Extension of the Kalman Filter to Nonlinear Systems
, 1997
"... The Kalman filter(KF) is one of the most widely used methods for tracking and estimation due to its simplicity, optimality, tractability and robustness. However, the application of the KF to nonlinear systems can be difficult. The most common approach is to use the Extended Kalman Filter (EKF) which ..."
Abstract

Cited by 747 (6 self)
 Add to MetaCart
The Kalman filter(KF) is one of the most widely used methods for tracking and estimation due to its simplicity, optimality, tractability and robustness. However, the application of the KF to nonlinear systems can be difficult. The most common approach is to use the Extended Kalman Filter (EKF) which simply linearises all nonlinear models so that the traditional linear Kalman filter can be applied. Although the EKF (in its many forms) is a widely used filtering strategy, over thirty years of experience with it has led to a general consensus within the tracking and control community that it is difficult to implement, difficult to tune, and only reliable for systems which are almost linear on the time scale of the update intervals. In this paper a new linear estimator is developed and demonstrated. Using the principle that a set of discretely sampled points can be used to parameterise mean and covariance, the estimator yields performance equivalent to the KF for linear systems yet general...
The irreducibility of the space of curves of given genus
 Publ. Math. IHES
, 1969
"... Fix an algebraically closed field k. Let Mg be the moduli space of curves of genus g over k. The main result of this note is that Mg is irreducible for every k. Of course, whether or not M s is irreducible depends only on the characteristic of k. When the characteristic s o, we can assume that k ~ ..."
Abstract

Cited by 512 (2 self)
 Add to MetaCart
Fix an algebraically closed field k. Let Mg be the moduli space of curves of genus g over k. The main result of this note is that Mg is irreducible for every k. Of course, whether or not M s is irreducible depends only on the characteristic of k. When the characteristic s o, we can assume that k ~ (1, and then the result is classical. A simple proof appears in EnriquesChisini [E, vol. 3, chap. 3], based on analyzing the totality of coverings of p1 of degree n, with a fixed number d of ordinary branch points. This method has been extended to char. p by William Fulton [F], using specializations from char. o to char. p provided that p> 2g qi. Unfortunately, attempts to extend this method to all p seem to get stuck on difficult questions of wild ramification. Nowadays, the Teichmtiller theory gives a thoroughly analytic but very profound insight into this irreducibility when kC. Our approach however is closest to Severi's incomplete proof ([Se], Anhang F; the error is on pp. 344345 and seems to be quite basic) and follows a suggestion of Grothendieck for using the result in char. o to deduce the result in char. p. The basis of both Severi's and Grothendieck's ideas is to construct families of curves X, some singular, with pa(X)=g, over nonsingular parameter spaces, which in some sense contain enough singular curves to link together any two components that Mg might have. The essential thing that makes this method work now is a recent " stable reduction theorem " for abelian varieties. This result was first proved independently in char. o by Grothendieck, using methods of etale cohomology (private correspondence with J. Tate), and by Mumford, applying the easy half of Theorem (2.5), to go from curves to abelian varieties (cf. [M2]). Grothendieck has recently strengthened his method so that it applies in all characteristics (SGA 7, ~968) 9 Mumford has also given a proof using theta functions in char. ~2. The result is this: Stable Reduction Theorem. Let R be a discrete valuation ring with quotient field K. Let A be an abelian variety over K. Then there exists a finite algebraic extension L of K such
Bundle Adjustment  A Modern Synthesis
 VISION ALGORITHMS: THEORY AND PRACTICE, LNCS
, 2000
"... This paper is a survey of the theory and methods of photogrammetric bundle adjustment, aimed at potential implementors in the computer vision community. Bundle adjustment is the problem of refining a visual reconstruction to produce jointly optimal structure and viewing parameter estimates. Topics c ..."
Abstract

Cited by 555 (12 self)
 Add to MetaCart
This paper is a survey of the theory and methods of photogrammetric bundle adjustment, aimed at potential implementors in the computer vision community. Bundle adjustment is the problem of refining a visual reconstruction to produce jointly optimal structure and viewing parameter estimates. Topics covered include: the choice of cost function and robustness; numerical optimization including sparse Newton methods, linearly convergent approximations, updating and recursive methods; gauge (datum) invariance; and quality control. The theory is developed for general robust cost functions rather than restricting attention to traditional nonlinear least squares.
Mean shift, mode seeking, and clustering
 IEEE Transactions on Pattern Analysis and Machine Intelligence
, 1995
"... AbstractMean shift, a simple iterative procedure that shifts each data point to the average of data points in its neighborhood, is generalized and analyzed in this paper. This generalization makes some kmeans like clustering algorithms its special cases. It is shown that mean shift is a modeseeki ..."
Abstract

Cited by 620 (0 self)
 Add to MetaCart
AbstractMean shift, a simple iterative procedure that shifts each data point to the average of data points in its neighborhood, is generalized and analyzed in this paper. This generalization makes some kmeans like clustering algorithms its special cases. It is shown that mean shift is a modeseeking process on a surface constructed with a “shadow ” kernel. For Gaussian kernels, mean shift is a gradient mapping. Convergence is studied for mean shift iterations. Cluster analysis is treated as a deterministic problem of finding a fixed point of mean shift that characterizes the data. Applications in clustering and Hough transform are demonstrated. Mean shift is also considered as an evolutionary strategy that performs multistart global optimization. Index TermsMean shift, gradient descent, global optimization, Hough transform, cluster analysis, kmeans clustering. I.
SNOPT: An SQP Algorithm For LargeScale Constrained Optimization
, 2002
"... Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first deriv ..."
Abstract

Cited by 582 (23 self)
 Add to MetaCart
Sequential quadratic programming (SQP) methods have proved highly effective for solving constrained optimization problems with smooth nonlinear functions in the objective and constraints. Here we consider problems with general inequality constraints (linear and nonlinear). We assume that first derivatives are available, and that the constraint gradients are sparse. We discuss
Exploiting Generative Models in Discriminative Classifiers
 In Advances in Neural Information Processing Systems 11
, 1998
"... Generative probability models such as hidden Markov models provide a principled way of treating missing information and dealing with variable length sequences. On the other hand, discriminative methods such as support vector machines enable us to construct flexible decision boundaries and often resu ..."
Abstract

Cited by 538 (11 self)
 Add to MetaCart
Generative probability models such as hidden Markov models provide a principled way of treating missing information and dealing with variable length sequences. On the other hand, discriminative methods such as support vector machines enable us to construct flexible decision boundaries and often result in classification performance superior to that of the model based approaches. An ideal classifier should combine these two complementary approaches. In this paper, we develop a natural way of achieving this combination by deriving kernel functions for use in discriminative methods such as support vector machines from generative probability models. We provide a theoretical justification for this combination as well as demonstrate a substantial improvement in the classification performance in the context of DNA and protein sequence analysis.
Results 1  10
of
75,930