Matching Items (5)
151692-Thumbnail Image.png
Description
A new method for generating artificial fingerprints is presented. Due to their uniqueness and durability, fingerprints are invaluable tools for identification for law enforcement and other purposes. Large databases of varied, realistic artificial fingerprints are needed to aid in the development and evaluation of automated systems for criminal or biometric

A new method for generating artificial fingerprints is presented. Due to their uniqueness and durability, fingerprints are invaluable tools for identification for law enforcement and other purposes. Large databases of varied, realistic artificial fingerprints are needed to aid in the development and evaluation of automated systems for criminal or biometric identification. Further, an effective method for simulating fingerprints may provide insight into the biological processes underlying print formation. However, previous attempts at simulating prints have been unsatisfactory. We approach the problem of creating artificial prints through a pattern formation model. We demonstrate how it is possible to generate distinctive patterns that strongly resemble real fingerprints via a system of partial differential equations with a suitable domain and initial conditions.
ContributorsColtin, Kevin (Author) / Armbruster, Hans D (Thesis advisor) / Platte, Rodrigo B (Committee member) / Welfert, Bruno D (Committee member) / Arizona State University (Publisher)
Created2013
149953-Thumbnail Image.png
Description
The theme for this work is the development of fast numerical algorithms for sparse optimization as well as their applications in medical imaging and source localization using sensor array processing. Due to the recently proposed theory of Compressive Sensing (CS), the $\ell_1$ minimization problem attracts more attention for its ability

The theme for this work is the development of fast numerical algorithms for sparse optimization as well as their applications in medical imaging and source localization using sensor array processing. Due to the recently proposed theory of Compressive Sensing (CS), the $\ell_1$ minimization problem attracts more attention for its ability to exploit sparsity. Traditional interior point methods encounter difficulties in computation for solving the CS applications. In the first part of this work, a fast algorithm based on the augmented Lagrangian method for solving the large-scale TV-$\ell_1$ regularized inverse problem is proposed. Specifically, by taking advantage of the separable structure, the original problem can be approximated via the sum of a series of simple functions with closed form solutions. A preconditioner for solving the block Toeplitz with Toeplitz block (BTTB) linear system is proposed to accelerate the computation. An in-depth discussion on the rate of convergence and the optimal parameter selection criteria is given. Numerical experiments are used to test the performance and the robustness of the proposed algorithm to a wide range of parameter values. Applications of the algorithm in magnetic resonance (MR) imaging and a comparison with other existing methods are included. The second part of this work is the application of the TV-$\ell_1$ model in source localization using sensor arrays. The array output is reformulated into a sparse waveform via an over-complete basis and study the $\ell_p$-norm properties in detecting the sparsity. An algorithm is proposed for minimizing a non-convex problem. According to the results of numerical experiments, the proposed algorithm with the aid of the $\ell_p$-norm can resolve closely distributed sources with higher accuracy than other existing methods.
ContributorsShen, Wei (Author) / Mittlemann, Hans D (Thesis advisor) / Renaut, Rosemary A. (Committee member) / Jackiewicz, Zdzislaw (Committee member) / Gelb, Anne (Committee member) / Ringhofer, Christian (Committee member) / Arizona State University (Publisher)
Created2011
157649-Thumbnail Image.png
Description
I focus on algorithms that generate good sampling points for function approximation. In 1D, it is well known that polynomial interpolation using equispaced points is unstable. On the other hand, using Chebyshev nodes provides both stable and highly accurate points for polynomial interpolation. In higher dimensional complex regions, optimal sampling

I focus on algorithms that generate good sampling points for function approximation. In 1D, it is well known that polynomial interpolation using equispaced points is unstable. On the other hand, using Chebyshev nodes provides both stable and highly accurate points for polynomial interpolation. In higher dimensional complex regions, optimal sampling points are not known explicitly. This work presents robust algorithms that find good sampling points in complex regions for polynomial interpolation, least-squares, and radial basis function (RBF) methods. The quality of these nodes is measured using the Lebesgue constant. I will also consider optimal sampling for constrained optimization, used to solve PDEs, where boundary conditions must be imposed. Furthermore, I extend the scope of the problem to include finding near-optimal sampling points for high-order finite difference methods. These high-order finite difference methods can be implemented using either piecewise polynomials or RBFs.
ContributorsLiu, Tony (Author) / Platte, Rodrigo B (Thesis advisor) / Renaut, Rosemary (Committee member) / Kaspar, David (Committee member) / Moustaoui, Mohamed (Committee member) / Motsch, Sebastien (Committee member) / Arizona State University (Publisher)
Created2019
158635-Thumbnail Image.png
Description
Dimensionality reduction methods are examined for large-scale discrete problems, specifically for the solution of three-dimensional geophysics problems: the inversion of gravity and magnetic data. The matrices for the associated forward problems have beneficial structure for each depth layer of the volume domain, under mild assumptions, which facilitates the use of

Dimensionality reduction methods are examined for large-scale discrete problems, specifically for the solution of three-dimensional geophysics problems: the inversion of gravity and magnetic data. The matrices for the associated forward problems have beneficial structure for each depth layer of the volume domain, under mild assumptions, which facilitates the use of the two dimensional fast Fourier transform for evaluating forward and transpose matrix operations, providing considerable savings in both computational costs and storage requirements. Application of this approach for the magnetic problem is new in the geophysics literature. Further, the approach is extended for padded volume domains.

Stabilized inversion is obtained efficiently by applying novel randomization techniques within each update of the iteratively reweighted scheme. For a general rectangular linear system, a randomization technique combined with preconditioning is introduced and investigated. This is shown to provide well-conditioned inversion, stabilized through truncation. Applying this approach, while implementing matrix operations using the two dimensional fast Fourier transform, yields computationally effective inversion, in memory and cost. Validation is provided via synthetic data sets, and the approach is contrasted with the well-known LSRN algorithm when applied to these data sets. The results demonstrate a significant reduction in computational cost with the new algorithm. Further, this new algorithm produces results for inversion of real magnetic data consistent with those provided in literature.

Typically, the iteratively reweighted least squares algorithm depends on a standard Tikhonov formulation. Here, this is solved using both a randomized singular value de- composition and the iterative LSQR Krylov algorithm. The results demonstrate that the new algorithm is competitive with these approaches and offers the advantage that no regularization parameter needs to be found at each outer iteration.

Given its efficiency, investigating the new algorithm for the joint inversion of these data sets may be fruitful. Initial research on joint inversion using the two dimensional fast Fourier transform has recently been submitted and provides the basis for future work. Several alternative directions for dimensionality reduction are also discussed, including iteratively applying an approximate pseudo-inverse and obtaining an approximate Kronecker product decomposition via randomization for a general matrix. These are also topics for future consideration.
ContributorsHogue, Jarom David (Author) / Renaut, Rosemary A. (Thesis advisor) / Jackiewicz, Zdzislaw (Committee member) / Platte, Rodrigo B (Committee member) / Ringhofer, Christian (Committee member) / Wlefert, Bruno (Committee member) / Arizona State University (Publisher)
Created2020
158258-Thumbnail Image.png
Description
This work is concerned with the study and numerical solution of large reaction diffusion systems with applications to the simulation of degradation effects in solar cells. A discussion of the basics of solar cells including the function of solar cells, the degradation of energy efficiency that happens over time, defects

This work is concerned with the study and numerical solution of large reaction diffusion systems with applications to the simulation of degradation effects in solar cells. A discussion of the basics of solar cells including the function of solar cells, the degradation of energy efficiency that happens over time, defects that affect solar cell efficiency and specific defects that can be modeled with a reaction diffusion system are included. Also included is a simple model equation of a solar cell. The basics of stoichiometry theory, how it applies to kinetic reaction systems, and some conservation properties are introduced. A model that considers the migration of defects in addition to the reaction processes is considered. A discussion of asymptotics and how it relates to the numerical simulation of the lifetime of solar cells is included. A reduced solution is considered and a presentation of a numerical comparison of the reduced solution with the full solution on a simple test problem is included. Operator splitting techniques are introduced and discussed. Asymptotically preserving schemes combine asymptotics and operator splitting to use reasonable time steps. A presentation of a realistic example of this study applied to solar cells follows.
ContributorsShapiro, Bruce G. (Author) / Ringhofer, Christian (Thesis advisor) / Gardner, Carl L (Committee member) / Jackiewicz, Zdzislaw (Committee member) / Platte, Rodrigo B (Committee member) / Vasileska, Dragica (Committee member) / Arizona State University (Publisher)
Created2020