Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
5,013
result(s) for
"Numerical linear algebra"
Sort by:
Robust optimisation algorithm for the measurement matrix in compressed sensing
by
Liu, Jixin
,
Zhou, Ying
,
Sun, Quansen
in
(B0260) Optimisation techniques
,
(B0290H) Linear algebra (numerical analysis)
,
(B6135) Optical, image and video signal processing
2018
The measurement matrix which plays an important role in compressed sensing has got a lot of attention. However, the existing measurement matrices ignore the energy concentration characteristic of the natural images in the sparse domain, which can help to improve the sensing efficiency and the construction efficiency. Here, the authors propose a simple but efficient measurement matrix based on the Hadamard matrix, named Hadamard-diagonal matrix (HDM). In HDM, the energy conservation in the sparse domain is maximised. In addition, considering the reconstruction performance can be further improved by decreasing the mutual coherence of the measurement matrix, an effective optimisation strategy is adopted in order to reducing the mutual coherence for better reconstruction quality. The authors conduct several experiments to evaluate the performance of HDM and the effectiveness of optimisation algorithm. The experimental results show that HDM performs better than other popular measurement matrices, and the optimisation algorithm can improve the performance of not only the HDM but also the other popular measurement matrices.
Journal Article
Optimization algorithms on matrix manifolds
2008
Many problems in the sciences and engineering can be rephrased as optimization problems on matrix search spaces endowed with a so-called manifold structure. This book shows how to exploit the special structure of such problems to develop efficient numerical algorithms. It places careful emphasis on both the numerical formulation of the algorithm and its differential geometric abstraction--illustrating how good algorithms draw equally from the insights of differential geometry, optimization, and numerical analysis. Two more theoretical chapters provide readers with the background in differential geometry necessary to algorithmic development. In the other chapters, several well-known optimization methods such as steepest descent and conjugate gradients are generalized to abstract manifolds. The book provides a generic development of each of these methods, building upon the material of the geometric chapters. It then guides readers through the calculations that turn these geometrically formulated methods into concrete numerical algorithms. The state-of-the-art algorithms given as examples are competitive with the best existing algorithms for a selection of eigenspace problems in numerical linear algebra.
Tensor Decompositions and Applications
2009
This survey provides an overview of higher-order tensor decompositions, their applications, and available software. A tensor is a multidimensional or N-way array. Decompositions of higher-order tensors (i.e., N-way arrays with N ≥ 3) have applications in psychometrics, chemometrics, signal processing, numerical linear algebra, computer vision, numerical analysis, data mining, neuroscience, graph analysis, and elsewhere. Two particular tensor decompositions can be considered to be higher-order extensions of the matrix singular value decomposition: CANDECOMP/PARAFAC (CP) decomposes a tensor as a sum of rank-one tensors, and the Tucker decomposition is a higher-order form of principal component analysis. There are many other tensor decompositions, including INDSCAL, PARAFAC2, CANDELINC, DEDICOM, and PARATUCK2 as well as nonnegative variants of all of the above. The N-way Toolbox, Tensor Toolbox, and Multilinear Engine are examples of software packages for working with tensors.
Journal Article
THE SOLUTION PATH OF THE GENERALIZED LASSO
2011
We present a path algorithm for the generalized lasso problem. This problem penalizes the ℓ 1 norm of a matrix D times the coefficient vector, and has a wide range of applications, dictated by the choice of D. Our algorithm is based on solving the dual of the generalized lasso, which greatly facilitates computation of the path. For D = I (the usual lasso), we draw a connection between our approach and the well-known LARS algorithm. For an arbitrary D, we derive an unbiased estimate of the degrees of freedom of the generalized lasso fit. This estimate turns out to be quite intuitive in many applications.
Journal Article
Regression shrinkage and selection via the lasso: a retrospective
2011
In the paper I give a brief review of the basic idea and some history and then discuss some developments since the original paper on regression shrinkage and selection via the lasso. [PUBLICATION ABSTRACT]
Journal Article
Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions
by
Halko, N.
,
Martinsson, P. G.
,
Tropp, J. A.
in
Algorithmics. Computability. Computer arithmetics
,
Algorithms
,
Applied sciences
2011
Low-rank matrix approximations, such as the truncated singular value decomposition and the rank-revealing QR decomposition, play a central role in data analysis and scientific computing. This work surveys and extends recent research which demonstrates that randomization offers a powerful tool for performing low-rank matrix approximation. These techniques exploit modern computational architectures more fully than classical methods and open the possibility of dealing with truly massive data sets. This paper presents a modular framework for constructing randomized algorithms that compute partial matrix decompositions. These methods use random sampling to identify a subspace that captures most of the action of a matrix. The input matrix is then compressed—either explicitly or implicitly—to this subspace, and the reduced matrix is manipulated deterministically to obtain the desired low-rank factorization. In many cases, this approach beats its classical competitors in terms of accuracy, robustness, and/or speed. These claims are supported by extensive numerical experiments and a detailed error analysis. The specific benefits of randomized techniques depend on the computational environment. Consider the model problem of finding the k dominant components of the singular value decomposition of an m × n matrix. (i) For a dense input matrix, randomized algorithms require O(mn log(k)) floating-point operations (flops) in contrast to O(mnk) for classical algorithms. (ii) For a sparse input matrix, the flop count matches classical Krylov subspace methods, but the randomized approach is more robust and can easily be reorganized to exploit multi-processor architectures. (iii) For a matrix that is too large to fit in fast memory, the randomized techniques require only a constant number of passes over the data, as opposed to O(k) passes for classical algorithms. In fact, it is sometimes possible to perform matrix approximation with a single pass over the data.
Journal Article
A Constrained ℓ1 Minimization Approach to Sparse Precision Matrix Estimation
by
Cai, Tony
,
Liu, Weidong
,
Luo, Xi
in
Acceleration of convergence
,
Analytical estimating
,
Applications
2011
This article proposes a constrained ℓ
1
minimization method for estimating a sparse inverse covariance matrix based on a sample of n iid p-variate random variables. The resulting estimator is shown to have a number of desirable properties. In particular, the rate of convergence between the estimator and the true s-sparse precision matrix under the spectral norm is
when the population distribution has either exponential-type tails or polynomial-type tails. We present convergence rates under the elementwise ℓ
∞
norm and Frobenius norm. In addition, we consider graphical model selection. The procedure is easily implemented by linear programming. Numerical performance of the estimator is investigated using both simulated and real data. In particular, the procedure is applied to analyze a breast cancer dataset and is found to perform favorably compared with existing methods.
Journal Article