Matching Items (152)
Filtering by

Clear all filters

150930-Thumbnail Image.png
Description
In this thesis, an integrated waveform-agile multi-modal tracking-beforedetect sensing system is investigated and the performance is evaluated using an experimental platform. The sensing system of adapting asymmetric multi-modal sensing operation platforms using radio frequency (RF) radar and electro-optical (EO) sensors allows for integration of complementary information from different sensors. However,

In this thesis, an integrated waveform-agile multi-modal tracking-beforedetect sensing system is investigated and the performance is evaluated using an experimental platform. The sensing system of adapting asymmetric multi-modal sensing operation platforms using radio frequency (RF) radar and electro-optical (EO) sensors allows for integration of complementary information from different sensors. However, there are many challenges to overcome, including tracking low signal-to-noise ratio (SNR) targets, waveform configurations that can optimize tracking performance and statistically dependent measurements. Address some of these challenges, a particle filter (PF) based recursive waveformagile track-before-detect (TBD) algorithm is developed to avoid information loss caused by conventional detection under low SNR environments. Furthermore, a waveform-agile selection technique is integrated into the PF-TBD to allow for adaptive waveform configurations. The embedded exponential family (EEF) approach is used to approximate distributions of parameters of dependent RF and EO measurements and to further improve target detection rate and tracking performance. The performance of the integrated algorithm is evaluated using real data from three experimental scenarios.
ContributorsLiu, Shubo (Author) / Papandreou-Suppappola, Antonia (Thesis advisor) / Duman, Tolga (Committee member) / Kovvali, Narayan (Committee member) / Arizona State University (Publisher)
Created2012
150896-Thumbnail Image.png
Description
Human operators have difficulty driving cranes quickly, accurately, and safely because of the slow response of heavy crane structures, non-intuitive control interfaces, and payload oscillations. Recently, a novel hand-motion crane control system has been proposed to improve performance by coupling an intuitive control interface with an element that reduces the

Human operators have difficulty driving cranes quickly, accurately, and safely because of the slow response of heavy crane structures, non-intuitive control interfaces, and payload oscillations. Recently, a novel hand-motion crane control system has been proposed to improve performance by coupling an intuitive control interface with an element that reduces the complex oscillatory behavior of the payload. Hand-motion control allows operators to drive a crane by simply moving a hand-held radio-frequency tag through the desired path. Real-time location sensors are used to track the movements of the tag and the tag position is used in a feedback control loop to drive the crane. An input shaper is added to eliminate dangerous payload oscillations. However, tag position measurements are corrupted by noise. It is important to understand the noise properties so that appropriate filters can be designed to mitigate the effects of noise and improve tracking accuracy. This work discusses implementing filtering techniques to address the issue of noise in the operating environment. Five different filters are used on experimentally-acquired tag trajectories to reduce noise. The filtered trajectories are then used to drive crane simulations. Filter performance is evaluated with respect to the energy usage of the crane trolley, the settling time of the crane payload oscillations, and the safety corridor of the crane trajectory. The effects of filter window lengths on these parameters are also investigated. An adaptive filtering technique, namely the Kalman filter, adapts to the noise characteristics of the workspace to minimize the tag tracking error and performs better than the other filtering techniques examined.
ContributorsRagunathan, Sudarshan (Author) / Frakes, David (Thesis advisor) / Singhose, William (Committee member) / Tillery, Stephen Helms (Committee member) / Arizona State University (Publisher)
Created2012
150830-Thumbnail Image.png
Description
Research on developing new algorithms to improve information on brain functionality and structure is ongoing. Studying neural activity through dipole source localization with electroencephalography (EEG) and magnetoencephalography (MEG) sensor measurements can lead to diagnosis and treatment of a brain disorder and can also identify the area of the brain from

Research on developing new algorithms to improve information on brain functionality and structure is ongoing. Studying neural activity through dipole source localization with electroencephalography (EEG) and magnetoencephalography (MEG) sensor measurements can lead to diagnosis and treatment of a brain disorder and can also identify the area of the brain from where the disorder has originated. Designing advanced localization algorithms that can adapt to environmental changes is considered a significant shift from manual diagnosis which is based on the knowledge and observation of the doctor, to an adaptive and improved brain disorder diagnosis as these algorithms can track activities that might not be noticed by the human eye. An important consideration of these localization algorithms, however, is to try and minimize the overall power consumption in order to improve the study and treatment of brain disorders. This thesis considers the problem of estimating dynamic parameters of neural dipole sources while minimizing the system's overall power consumption; this is achieved by minimizing the number of EEG/MEG measurements sensors without a loss in estimation performance accuracy. As the EEG/MEG measurements models are related non-linearity to the dipole source locations and moments, these dynamic parameters can be estimated using sequential Monte Carlo methods such as particle filtering. Due to the large number of sensors required to record EEG/MEG Measurements for use in the particle filter, over long period recordings, a large amounts of power is required for storage and transmission. In order to reduce the overall power consumption, two methods are proposed. The first method used the predicted mean square estimation error as the performance metric under the constraint of a maximum power consumption. The performance metric of the second method uses the distance between the location of the sensors and the location estimate of the dipole source at the previous time step; this sensor scheduling scheme results in maximizing the overall signal-to-noise ratio. The performance of both methods is demonstrated using simulated data, and both methods show that they can provide good estimation results with significant reduction in the number of activated sensors at each time step.
ContributorsMichael, Stefanos (Author) / Papandreou-Suppappola, Antonia (Thesis advisor) / Chakrabarti, Chaitali (Committee member) / Kovvali, Narayan (Committee member) / Arizona State University (Publisher)
Created2012
150437-Thumbnail Image.png
Description
Magnetic Resonance Imaging (MRI) is limited in speed and resolution by the inherently low Signal to Noise Ratio (SNR) of the underlying signal. Advances in sampling efficiency are required to support future improvements in scan time and resolution. SNR efficiency is improved by sampling data for a larger proportion of

Magnetic Resonance Imaging (MRI) is limited in speed and resolution by the inherently low Signal to Noise Ratio (SNR) of the underlying signal. Advances in sampling efficiency are required to support future improvements in scan time and resolution. SNR efficiency is improved by sampling data for a larger proportion of total imaging time. This is challenging as these acquisitions are typically subject to artifacts such as blurring and distortions. The current work proposes a set of tools to help with the creation of different types of SNR efficient scans. An SNR efficient pulse sequence providing diffusion imaging data with full brain coverage and minimal distortion is first introduced. The proposed method acquires single-shot, low resolution image slabs which are then combined to reconstruct the full volume. An iterative deblurring algorithm allowing the lengthening of spiral SPoiled GRadient echo (SPGR) acquisition windows in the presence of rapidly varying off-resonance fields is then presented. Finally, an efficient and practical way of collecting 3D reformatted data is proposed. This method constitutes a good tradeoff between 2D and 3D neuroimaging in terms of scan time and data presentation. These schemes increased the SNR efficiency of currently existing methods and constitute key enablers for the development of SNR efficient MRI.
ContributorsAboussouan, Eric (Author) / Frakes, David (Thesis advisor) / Pipe, James (Thesis advisor) / Debbins, Joseph (Committee member) / Towe, Bruce (Committee member) / Arizona State University (Publisher)
Created2011
150423-Thumbnail Image.png
Description
In this thesis, an adaptive waveform selection technique for dynamic target tracking under low signal-to-noise ratio (SNR) conditions is investigated. The approach is integrated with a track-before-detect (TBD) algorithm and uses delay-Doppler matched filter (MF) outputs as raw measurements without setting any threshold for extracting delay-Doppler estimates. The particle filter

In this thesis, an adaptive waveform selection technique for dynamic target tracking under low signal-to-noise ratio (SNR) conditions is investigated. The approach is integrated with a track-before-detect (TBD) algorithm and uses delay-Doppler matched filter (MF) outputs as raw measurements without setting any threshold for extracting delay-Doppler estimates. The particle filter (PF) Bayesian sequential estimation approach is used with the TBD algorithm (PF-TBD) to estimate the dynamic target state. A waveform-agile TBD technique is proposed that integrates the PF-TBD with a waveform selection technique. The new approach predicts the waveform to transmit at the next time step by minimizing the predicted mean-squared error (MSE). As a result, the radar parameters are adaptively and optimally selected for superior performance. Based on previous work, this thesis highlights the applicability of the predicted covariance matrix to the lower SNR waveform-agile tracking problem. The adaptive waveform selection algorithm's MSE performance was compared against fixed waveforms using Monte Carlo simulations. It was found that the adaptive approach performed at least as well as the best fixed waveform when focusing on estimating only position or only velocity. When these estimates were weighted by different amounts, then the adaptive performance exceeded all fixed waveforms. This improvement in performance demonstrates the utility of the predicted covariance in waveform design, at low SNR conditions that are poorly handled with more traditional tracking algorithms.
ContributorsPiwowarski, Ryan (Author) / Papandreou-Suppappola, Antonia (Thesis advisor) / Chakrabarti, Chaitali (Committee member) / Kovvali, Narayan (Committee member) / Arizona State University (Publisher)
Created2011
150476-Thumbnail Image.png
Description
Multidimensional (MD) discrete Fourier transform (DFT) is a key kernel algorithm in many signal processing applications, such as radar imaging and medical imaging. Traditionally, a two-dimensional (2-D) DFT is computed using Row-Column (RC) decomposition, where one-dimensional (1-D) DFTs are computed along the rows followed by 1-D DFTs along the columns.

Multidimensional (MD) discrete Fourier transform (DFT) is a key kernel algorithm in many signal processing applications, such as radar imaging and medical imaging. Traditionally, a two-dimensional (2-D) DFT is computed using Row-Column (RC) decomposition, where one-dimensional (1-D) DFTs are computed along the rows followed by 1-D DFTs along the columns. However, architectures based on RC decomposition are not efficient for large input size data which have to be stored in external memories based Synchronous Dynamic RAM (SDRAM). In this dissertation, first an efficient architecture to implement 2-D DFT for large-sized input data is proposed. This architecture achieves very high throughput by exploiting the inherent parallelism due to a novel 2-D decomposition and by utilizing the row-wise burst access pattern of the SDRAM external memory. In addition, an automatic IP generator is provided for mapping this architecture onto a reconfigurable platform of Xilinx Virtex-5 devices. For a 2048x2048 input size, the proposed architecture is 1.96 times faster than RC decomposition based implementation under the same memory constraints, and also outperforms other existing implementations. While the proposed 2-D DFT IP can achieve high performance, its output is bit-reversed. For systems where the output is required to be in natural order, use of this DFT IP would result in timing overhead. To solve this problem, a new bandwidth-efficient MD DFT IP that is transpose-free and produces outputs in natural order is proposed. It is based on a novel decomposition algorithm that takes into account the output order, FPGA resources, and the characteristics of off-chip memory access. An IP generator is designed and integrated into an in-house FPGA development platform, AlgoFLEX, for easy verification and fast integration. The corresponding 2-D and 3-D DFT architectures are ported onto the BEE3 board and their performance measured and analyzed. The results shows that the architecture can maintain the maximum memory bandwidth throughout the whole procedure while avoiding matrix transpose operations used in most other MD DFT implementations. The proposed architecture has also been ported onto the Xilinx ML605 board. When clocked at 100 MHz, 2048x2048 images with complex single-precision can be processed in less than 27 ms. Finally, transpose-free imaging flows for range-Doppler algorithm (RDA) and chirp-scaling algorithm (CSA) in SAR imaging are proposed. The corresponding implementations take advantage of the memory access patterns designed for the MD DFT IP and have superior timing performance. The RDA and CSA flows are mapped onto a unified architecture which is implemented on an FPGA platform. When clocked at 100MHz, the RDA and CSA computations with data size 4096x4096 can be completed in 323ms and 162ms, respectively. This implementation outperforms existing SAR image accelerators based on FPGA and GPU.
ContributorsYu, Chi-Li (Author) / Chakrabarti, Chaitali (Thesis advisor) / Papandreou-Suppappola, Antonia (Committee member) / Karam, Lina (Committee member) / Cao, Yu (Committee member) / Arizona State University (Publisher)
Created2012
150720-Thumbnail Image.png
Description
Current treatment methods for cerebral aneurysms are providing life-saving measures for patients suffering from these blood vessel wall protrusions; however, the drawbacks present unfortunate circumstances in the invasive procedure or with efficient occlusion of the aneurysms. With the advancement of medical devices, liquid-to-solid gelling materials that could be delivered endovascularly

Current treatment methods for cerebral aneurysms are providing life-saving measures for patients suffering from these blood vessel wall protrusions; however, the drawbacks present unfortunate circumstances in the invasive procedure or with efficient occlusion of the aneurysms. With the advancement of medical devices, liquid-to-solid gelling materials that could be delivered endovascularly have gained interest. The development of these systems stems from the need to circumvent surgical methods and the requirement for improved occlusion of aneurysms to prevent recanalization and potential complications. The work presented herein reports on a liquid-to-solid gelling material, which undergoes gelation via dual mechanisms. Using a temperature-responsive polymer, poly(N-isopropylacrylamide) (poly(NIPAAm), the gelling system can transition from a solution at low temperatures to a gel at body temperature (physical gelation). Additionally, by conjugating reactive functional groups onto the polymers, covalent cross-links can be formed via chemical reaction between the two moieties (chemical gelation). The advantage of this gelling system comprises of its water-based properties as well as the ability of the physical and chemical gelation to occur within physiological conditions. By developing the polymer gelling system in a ground-up approach via synthesis, its added benefit is the capability of modifying the properties of the system as needed for particular applications, in this case for embolization of cerebral aneurysms. The studies provided in this doctoral work highlight the synthesis, characterization and testing of these polymer gelling systems for occlusion of aneurysms. Conducted experiments include thermal, mechanical, structural and chemical characterization, as well as analysis of swelling, degradation, kinetics, cytotoxicity, in vitro glass models and in vivo swine study. Data on thermoresponsive poly(NIPAAm) indicated that the phase transition it undertakes comes as a result of the polymer chains associating as temperature is increased. Poly(NIPAAm) was functionalized with thiols and vinyls to provide for added chemical cross-linking. By combining both modes of gelation, physical and chemical, a gel with reduced creep flow and increased strength was developed. Being waterborne, the gels demonstrated excellent biocompatibility and were easily delivered via catheters and injected within aneurysms, without undergoing degradation. The dual gelling polymer systems demonstrated potential in use as embolic agents for cerebral aneurysm embolization.
ContributorsBearat, Hanin H (Author) / Vernon, Brent L (Thesis advisor) / Frakes, David (Committee member) / Massia, Stephen (Committee member) / Pauken, Christine (Committee member) / Preul, Mark (Committee member) / Solis, Francisco (Committee member) / Arizona State University (Publisher)
Created2012
151165-Thumbnail Image.png
Description
Camera calibration has applications in the fields of robotic motion, geographic mapping, semiconductor defect characterization, and many more. This thesis considers camera calibration for the purpose of high accuracy three-dimensional reconstruction when characterizing ball grid arrays within the semiconductor industry. Bouguet's calibration method is used following a set of criteria

Camera calibration has applications in the fields of robotic motion, geographic mapping, semiconductor defect characterization, and many more. This thesis considers camera calibration for the purpose of high accuracy three-dimensional reconstruction when characterizing ball grid arrays within the semiconductor industry. Bouguet's calibration method is used following a set of criteria with the purpose of studying the method's performance according to newly proposed standards. The performance of the camera calibration method is currently measured using standards such as pixel error and computational time. This thesis proposes the use of standard deviation of the intrinsic parameter estimation within a Monte Carlo simulation as a new standard of performance measure. It specifically shows that the standard deviation decreases based on the increased number of images input into the calibration routine. It is also shown that the default thresholds of the non-linear maximum likelihood estimation problem of the calibration method require change in order to improve computational time performance; however, the accuracy lost is negligable even for high accuracy requirements such as ball grid array characterization.
ContributorsStenger, Nickolas Arthur (Author) / Papandreou-Suppappola, Antonia (Thesis advisor) / Kovvali, Narayan (Committee member) / Tepedelenlioğlu, Cihan (Committee member) / Arizona State University (Publisher)
Created2012
151215-Thumbnail Image.png
Description
Today's mobile devices have to support computation-intensive multimedia applications with a limited energy budget. In this dissertation, we present architecture level and algorithm-level techniques that reduce energy consumption of these devices with minimal impact on system quality. First, we present novel techniques to mitigate the effects of SRAM memory failures

Today's mobile devices have to support computation-intensive multimedia applications with a limited energy budget. In this dissertation, we present architecture level and algorithm-level techniques that reduce energy consumption of these devices with minimal impact on system quality. First, we present novel techniques to mitigate the effects of SRAM memory failures in JPEG2000 implementations operating in scaled voltages. We investigate error control coding schemes and propose an unequal error protection scheme tailored for JPEG2000 that reduces overhead without affecting the performance. Furthermore, we propose algorithm-specific techniques for error compensation that exploit the fact that in JPEG2000 the discrete wavelet transform outputs have larger values for low frequency subband coefficients and smaller values for high frequency subband coefficients. Next, we present use of voltage overscaling to reduce the data-path power consumption of JPEG codecs. We propose an algorithm-specific technique which exploits the characteristics of the quantized coefficients after zig-zag scan to mitigate errors introduced by aggressive voltage scaling. Third, we investigate the effect of reducing dynamic range for datapath energy reduction. We analyze the effect of truncation error and propose a scheme that estimates the mean value of the truncation error during the pre-computation stage and compensates for this error. Such a scheme is very effective for reducing the noise power in applications that are dominated by additions and multiplications such as FIR filter and transform computation. We also present a novel sum of absolute difference (SAD) scheme that is based on most significant bit truncation. The proposed scheme exploits the fact that most of the absolute difference (AD) calculations result in small values, and most of the large AD values do not contribute to the SAD values of the blocks that are selected. Such a scheme is highly effective in reducing the energy consumption of motion estimation and intra-prediction kernels in video codecs. Finally, we present several hybrid energy-saving techniques based on combination of voltage scaling, computation reduction and dynamic range reduction that further reduce the energy consumption while keeping the performance degradation very low. For instance, a combination of computation reduction and dynamic range reduction for Discrete Cosine Transform shows on average, 33% to 46% reduction in energy consumption while incurring only 0.5dB to 1.5dB loss in PSNR.
ContributorsEmre, Yunus (Author) / Chakrabarti, Chaitali (Thesis advisor) / Bakkaloglu, Bertan (Committee member) / Cao, Yu (Committee member) / Papandreou-Suppappola, Antonia (Committee member) / Arizona State University (Publisher)
Created2012
151200-Thumbnail Image.png
Description
In recent years, we have observed the prevalence of stream applications in many embedded domains. Stream programs distinguish themselves from traditional sequential programming languages through well defined independent actors, explicit data communication, and stable code/data access patterns. In order to achieve high performance and low power, scratch pad memory (SPM)

In recent years, we have observed the prevalence of stream applications in many embedded domains. Stream programs distinguish themselves from traditional sequential programming languages through well defined independent actors, explicit data communication, and stable code/data access patterns. In order to achieve high performance and low power, scratch pad memory (SPM) has been introduced in today's embedded multicore processors. Current design frameworks for developing stream applications on SPM enhanced embedded architectures typically do not include a compiler that can perform automatic partitioning, mapping and scheduling under limited on-chip SPM capacities and memory access delays. Consequently, many designs are implemented manually, which leads to lengthy tasks and inferior designs. In this work, optimization techniques that automatically compile stream programs onto embedded multi-core architectures are proposed. As an initial case study, we implemented an automatic target recognition (ATR) algorithm on the IBM Cell Broadband Engine (BE). Then integer linear programming (ILP) and heuristic approaches were proposed to schedule stream programs on a single core embedded processor that has an SPM with code overlay. Later, ILP and heuristic approaches for Compiling Stream programs on SPM enhanced Multicore Processors (CSMP) were studied. The proposed CSMP ILP and heuristic approaches do not optimize for cycles in stream applications. Further, the number of software pipeline stages in the implementation is dependent on actor to processing engine (PE) mapping and is uncontrollable. We next presented a Retiming technique for Throughput optimization on Embedded Multi-core processors (RTEM). RTEM approach inherently handles cycles and can accept an upper bound on the number of software pipeline stages to be generated. We further enhanced RTEM by incorporating unrolling (URSTEM) that preserves all the beneficial properties of RTEM heuristic and also scales with the number of PEs through unrolling.
ContributorsChe, Weijia (Author) / Chatha, Karam Singh (Thesis advisor) / Vrudhula, Sarma (Committee member) / Chakrabarti, Chaitali (Committee member) / Shrivastava, Aviral (Committee member) / Arizona State University (Publisher)
Created2012