• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Sparsekit: a basic tool kit for sparse matrix computations (1990)

by Y Saad
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 314
Next 10 →

QMR: a Quasi-Minimal Residual Method for Non-Hermitian Linear Systems

by Roland W. Freund, Noël M. Nachtigal , 1991
"... ... In this paper, we present a novel BCG-like approach, the quasi-minimal residual (QMR) method, which overcomes the problems of BCG. An implementation of QMR based on a look-ahead version of the nonsymmetric Lanczos algorithm is proposed. It is shown how BCG iterates can be recovered stably from t ..."
Abstract - Cited by 395 (26 self) - Add to MetaCart
... In this paper, we present a novel BCG-like approach, the quasi-minimal residual (QMR) method, which overcomes the problems of BCG. An implementation of QMR based on a look-ahead version of the nonsymmetric Lanczos algorithm is proposed. It is shown how BCG iterates can be recovered stably from the QMR process. Some further properties of the QMR approach are given and an error bound is presented. Finally, numerical experiments are reported.

A sparse approximate inverse preconditioner for nonsymmetric linear systems

by Michele Benzi, Miroslav Tůma - SIAM J. SCI. COMPUT , 1998
"... This paper is concerned with a new approach to preconditioning for large, sparse linear systems. A procedure for computing an incomplete factorization of the inverse of a nonsymmetric matrix is developed, and the resulting factorized sparse approximate inverse is used as an explicit preconditioner f ..."
Abstract - Cited by 197 (22 self) - Add to MetaCart
This paper is concerned with a new approach to preconditioning for large, sparse linear systems. A procedure for computing an incomplete factorization of the inverse of a nonsymmetric matrix is developed, and the resulting factorized sparse approximate inverse is used as an explicit preconditioner for conjugate gradient–type methods. Some theoretical properties of the preconditioner are discussed, and numerical experiments on test matrices from the Harwell–Boeing collection and from Tim Davis’s collection are presented. Our results indicate that the new preconditioner is cheaper to construct than other approximate inverse preconditioners. Furthermore, the new technique insures convergence rates of the preconditioned iteration which are comparable with those obtained with standard implicit preconditioners.

OSKI: A library of automatically tuned sparse matrix kernels

by Richard Vuduc, James W Demmel, Katherine A Yelick - Institute of Physics Publishing , 2005
"... kernels ..."
Abstract - Cited by 157 (20 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

... hierarchy-friendly kernels: The kernels defined in our interface are a superset of those available in similar library interfaces, including the Sparse BLAS standard [13, 14] and the SPARSKIT library =-=[15]-=-, among others [16]. Our “higherlevel” kernels are designed for cache-based machines and can execute much faster than their equivalent implementations using “standard” kernels. For example, in additio...

SuperLU DIST: A scalable distributed-memory sparse direct solver for unsymmetric linear systems

by Xiaoyes. Li, James W. Demmel - ACM Trans. Mathematical Software , 2003
"... We present the main algorithmic features in the software package SuperLU DIST, a distributedmemory sparse direct solver for large sets of linear equations. We give in detail our parallelization strategies, with a focus on scalability issues, and demonstrate the software’s parallel performance and sc ..."
Abstract - Cited by 145 (18 self) - Add to MetaCart
We present the main algorithmic features in the software package SuperLU DIST, a distributedmemory sparse direct solver for large sets of linear equations. We give in detail our parallelization strategies, with a focus on scalability issues, and demonstrate the software’s parallel performance and scalability on current machines. The solver is based on sparse Gaussian elimination, with an innovative static pivoting strategy proposed earlier by the authors. The main advantage of static pivoting over classical partial pivoting is that it permits a priori determination of data structures and communication patterns, which lets us exploit techniques used in parallel sparse Cholesky algorithms to better parallelize both LU decomposition and triangular solution on large-scale distributed machines.
(Show Context)

Citation Context

...), (2), (6), and the diagonal perturbation in step (4)). Now we turn to the second conguration of our algorithm, in which restarted GMRES [55] was used in step (6) (we used the version from SPARSKIT [=-=5-=-6]). The restart value is 50. Here, our LU factorization is used in preconditioning for GMRES. The convergence test is based on residual norm: jjr i jj 2 rtol jjr 0 jj 2 + atol, where the relative t...

A Combined Unifrontal/Multifrontal Method for Unsymmetric Sparse Matrices

by T A Davis, I S Duff
"... ..."
Abstract - Cited by 114 (15 self) - Add to MetaCart
Abstract not found

Matrix Market: A Web Resource for Test Matrix Collections

by Ronald F. Boisvert, Roldan Pozo, Karin Remington, Richard F. Barrett, Jack J. Dongarra - The Quality of Numerical Software: Assessment and Enhancement , 1997
"... We describe a repository of data for the testing of numerical algorithms and mathematical software for matrix computations. The repository is designed to accommodate both dense and sparse matrices, as well as software to generate matrices. It has been seeded with the well-known Harwell-Boeing sparse ..."
Abstract - Cited by 85 (6 self) - Add to MetaCart
We describe a repository of data for the testing of numerical algorithms and mathematical software for matrix computations. The repository is designed to accommodate both dense and sparse matrices, as well as software to generate matrices. It has been seeded with the well-known Harwell-Boeing sparse matrix collection. The raw data files have been augmented with an integrated World Wide Web interface which describes the matrices in the collection quantitatively and visually. For example, each matrix has a Web page which details its attributes, graphically depicts its sparsity pattern, and provides access to the matrix itself in several formats. In addition, a search mechanism is included which allows retrieval of matrices based on a variety of attributes, such as type and size, as well as through free-text search in abstracts. The URL is http://math.nist.gov/MatrixMarket/ .

Exponential integrators

by Marlis Hochbruck, Alexander Ostermann , 2010
"... In this paper we consider the construction, analysis, implementation and application of exponential integrators. The focus will be on two types of stiff problems. The first one is characterized by a Jacobian that possesses eigenvalues with large negative real parts. Parabolic partial differential eq ..."
Abstract - Cited by 68 (5 self) - Add to MetaCart
In this paper we consider the construction, analysis, implementation and application of exponential integrators. The focus will be on two types of stiff problems. The first one is characterized by a Jacobian that possesses eigenvalues with large negative real parts. Parabolic partial differential equations and their spatial discretization are typical examples. The second class consists of highly oscillatory problems with purely imaginary eigenvalues of large modulus. Apart from motivating the construction of exponential integrators for various classes of problems, our main intention in this article is to present the mathematics behind these methods. We will derive error bounds that are independent of stiffness or highest frequencies in the system. Since the implementation of exponential integrators requires the evaluation of the product of a matrix function with a vector, we will briefly discuss some possible approaches as well. The paper concludes with some applications, in

Model-driven autotuning of sparse matrix-vector multiply on GPUs

by Jee W. Choi, Amik Singh, Richard W. Vuduc - In PPoPP , 2010
"... We present a performance model-driven framework for automated performance tuning (autotuning) of sparse matrix-vector multiply (SpMV) on systems accelerated by graphics processing units (GPU). Our study consists of two parts. First, we describe several carefully hand-tuned SpMV implementations for G ..."
Abstract - Cited by 65 (4 self) - Add to MetaCart
We present a performance model-driven framework for automated performance tuning (autotuning) of sparse matrix-vector multiply (SpMV) on systems accelerated by graphics processing units (GPU). Our study consists of two parts. First, we describe several carefully hand-tuned SpMV implementations for GPUs, identifying key GPU-specific performance limitations, enhancements, and tuning opportunities. These implementations, which include variants on classical blocked compressed sparse row (BCSR) and blocked ELLPACK (BELLPACK) storage formats, match or exceed state-of-the-art implementations. For instance, our best BELLPACK implementation achieves up to 29.0 Gflop/s in single-precision and 15.7 Gflop/s in doubleprecision on the NVIDIA T10P multiprocessor (C1060), enhancing prior state-of-the-art unblocked implementations (Bell and Garland, 2009) by up to 1.8 × and 1.5 × for single- and doubleprecision respectively. However, achieving this level of performance requires input matrix-dependent parameter tuning. Thus, in the second part of this study, we develop a performance model that can guide tuning. Like prior autotuning models for CPUs (e.g., Im, Yelick, and Vuduc, 2004), this model requires offline measurements and run-time estimation, but more directly models the structure of multithreaded vector processors like GPUs. We show that our model can identify the implementations that achieve within 15 % of those found through exhaustive search.
(Show Context)

Citation Context

...re), the warp executes a parallel reduction to produce the final value. In the blocked variant of CSR, BCSR, we store r × c dense subblocks of non-zeros rather than storing each non-zero individually =-=[15]-=-, as illustrated in Figure 1. Depending on the matrix, we can in principle reduce the column index storage by up to roughly 1 r·c , since we need only store 1 column index per block.Name Dimensions N...

Optimizing the performance of sparse matrix-vector multiplication

by Eun-jin Im , 2000
"... Copyright 2000 by Eun-Jin Im ..."
Abstract - Cited by 65 (2 self) - Add to MetaCart
Copyright 2000 by Eun-Jin Im
(Show Context)

Citation Context

...mory bandwidth reduction, and found some benefits on symmetric multiprocessors [38], but very little effect on uniprocessors.s100 Sparse Matrix Packages for Multiprocessors Yousef Saad built SPARSKIT =-=[61]-=- and, with Andrei Malevsky, PSPARSLIB [62]. SPARSKIT is a collection of FORTRAN subroutines for sparse matrix computations. It includes format conversion routines among various sparse matrix formats, ...

Orderings for incomplete factorization preconditioning of nonsymmetric problems

by Michele Benzi, Daniel B. Szyld, Arno Van Duin - SIAM J. SCI. COMPUT , 1999
"... Numerical experiments are presented whereby the effect of reorderings on the convergence of preconditioned Krylov subspace methods for the solution of nonsymmetric linear systems is shown. The preconditioners used in this study are different variants of incomplete factorizations. It is shown that c ..."
Abstract - Cited by 60 (11 self) - Add to MetaCart
Numerical experiments are presented whereby the effect of reorderings on the convergence of preconditioned Krylov subspace methods for the solution of nonsymmetric linear systems is shown. The preconditioners used in this study are different variants of incomplete factorizations. It is shown that certain reorderings for direct methods, such as reverse Cuthill–McKee, can be very beneficial. The benefit can be seen in the reduction of the number of iterations and also in measuring the deviation of the preconditioned operator from the identity.
(Show Context)

Citation Context

... problems. For this reason, we discuss additional experiments performed on a selection of nonsymmetric matrices from various sources, including the Harwell--Boeing collection [19] and Saad's SPARSKIT =-=[40]-=-. These matrices arise from di#erent application areas: oil reservoir modeling, plasma physics, neutron di#usion, metal forming simulation, etc. Some of these matrices arise from finite element modeli...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University