Matching Items (167)
149992-Thumbnail Image.png
Description
Process variations have become increasingly important for scaled technologies starting at 45nm. The increased variations are primarily due to random dopant fluctuations, line-edge roughness and oxide thickness fluctuation. These variations greatly impact all aspects of circuit performance and pose a grand challenge to future robust IC design. To improve robustness,

Process variations have become increasingly important for scaled technologies starting at 45nm. The increased variations are primarily due to random dopant fluctuations, line-edge roughness and oxide thickness fluctuation. These variations greatly impact all aspects of circuit performance and pose a grand challenge to future robust IC design. To improve robustness, efficient methodology is required that considers effect of variations in the design flow. Analyzing timing variability of complex circuits with HSPICE simulations is very time consuming. This thesis proposes an analytical model to predict variability in CMOS circuits that is quick and accurate. There are several analytical models to estimate nominal delay performance but very little work has been done to accurately model delay variability. The proposed model is comprehensive and estimates nominal delay and variability as a function of transistor width, load capacitance and transition time. First, models are developed for library gates and the accuracy of the models is verified with HSPICE simulations for 45nm and 32nm technology nodes. The difference between predicted and simulated σ/μ for the library gates is less than 1%. Next, the accuracy of the model for nominal delay is verified for larger circuits including ISCAS'85 benchmark circuits. The model predicted results are within 4% error of HSPICE simulated results and take a small fraction of the time, for 45nm technology. Delay variability is analyzed for various paths and it is observed that non-critical paths can become critical because of Vth variation. Variability on shortest paths show that rate of hold violations increase enormously with increasing Vth variation.
ContributorsGummalla, Samatha (Author) / Chakrabarti, Chaitali (Thesis advisor) / Cao, Yu (Thesis advisor) / Bakkaloglu, Bertan (Committee member) / Arizona State University (Publisher)
Created2011
150046-Thumbnail Image.png
Description
This thesis describes a synthetic task environment, CyberCog, created for the purposes of 1) understanding and measuring individual and team situation awareness in the context of a cyber security defense task and 2) providing a context for evaluating algorithms, visualizations, and other interventions that are intended to improve cyber situation

This thesis describes a synthetic task environment, CyberCog, created for the purposes of 1) understanding and measuring individual and team situation awareness in the context of a cyber security defense task and 2) providing a context for evaluating algorithms, visualizations, and other interventions that are intended to improve cyber situation awareness. CyberCog provides an interactive environment for conducting human-in-loop experiments in which the participants of the experiment perform the tasks of a cyber security defense analyst in response to a cyber-attack scenario. CyberCog generates the necessary performance measures and interaction logs needed for measuring individual and team cyber situation awareness. Moreover, the CyberCog environment provides good experimental control for conducting effective situation awareness studies while retaining realism in the scenario and in the tasks performed.
ContributorsRajivan, Prashanth (Author) / Femiani, John (Thesis advisor) / Cooke, Nancy J. (Thesis advisor) / Lindquist, Timothy (Committee member) / Gary, Kevin (Committee member) / Arizona State University (Publisher)
Created2011
149644-Thumbnail Image.png
Description
Intuitive decision making refers to decision making based on situational pattern recognition, which happens without deliberation. It is a fast and effortless process that occurs without complete awareness. Moreover, it is believed that implicit learning is one means by which a foundation for intuitive decision making is developed. Accordingly, the

Intuitive decision making refers to decision making based on situational pattern recognition, which happens without deliberation. It is a fast and effortless process that occurs without complete awareness. Moreover, it is believed that implicit learning is one means by which a foundation for intuitive decision making is developed. Accordingly, the present study investigated several factors that affect implicit learning and the development of intuitive decision making in a simulated real-world environment: (1) simple versus complex situational patterns; (2) the diversity of the patterns to which an individual is exposed; (3) the underlying mechanisms. The results showed that simple patterns led to higher levels of implicit learning and intuitive decision-making accuracy than complex patterns; increased diversity enhanced implicit learning and intuitive decision-making accuracy; and an embodied mechanism, labeling, contributes to the development of intuitive decision making in a simulated real-world environment. The results suggest that simulated real-world environments can provide the basis for training intuitive decision making, that diversity is influential in the process of training intuitive decision making, and that labeling contributes to the development of intuitive decision making. These results are interpreted in the context of applied situations such as military applications involving remotely piloted aircraft.
ContributorsCovas-Smith, Christine Marie (Author) / Cooke, Nancy J. (Thesis advisor) / Patterson, Robert (Committee member) / Glenberg, Arthur (Committee member) / Homa, Donald (Committee member) / Arizona State University (Publisher)
Created2011
150067-Thumbnail Image.png
Description
The objective of this project was to evaluate human factors based cognitive aids on endoscope reprocessing. The project stems from recent failures in reprocessing (cleaning) endoscopes, contributing to the spread of harmful bacterial and viral agents between patients. Three themes were found to represent a majority of problems:

The objective of this project was to evaluate human factors based cognitive aids on endoscope reprocessing. The project stems from recent failures in reprocessing (cleaning) endoscopes, contributing to the spread of harmful bacterial and viral agents between patients. Three themes were found to represent a majority of problems: 1) lack of visibility (parts and tools were difficult to identify), 2) high memory demands, and 3) insufficient user feedback. In an effort to improve completion rate and eliminate error, cognitive aids were designed utilizing human factors principles that would replace existing manufacturer visual aids. Then, a usability test was conducted, which compared the endoscope reprocessing performance of novices using the standard manufacturer-provided visual aids and the new cognitive aids. Participants successfully completed 87.1% of the reprocessing procedure in the experimental condition with the use of the cognitive aids, compared to 46.3% in the control condition using only existing support materials. Twenty-five of sixty subtasks showed significant improvement in completion rates. When given a cognitive aid designed with human factors principles, participants were able to more successfully complete the reprocessing task. This resulted in an endoscope that was more likely to be safe for patient use.
ContributorsJolly, Jonathan D (Author) / Branaghan, Russell J (Thesis advisor) / Cooke, Nancy J. (Committee member) / Sanchez, Christopher (Committee member) / Arizona State University (Publisher)
Created2011
150167-Thumbnail Image.png
Description
Redundant Binary (RBR) number representations have been extensively used in the past for high-throughput Digital Signal Processing (DSP) systems. Data-path components based on this number system have smaller critical path delay but larger area compared to conventional two's complement systems. This work explores the use of RBR number representation for

Redundant Binary (RBR) number representations have been extensively used in the past for high-throughput Digital Signal Processing (DSP) systems. Data-path components based on this number system have smaller critical path delay but larger area compared to conventional two's complement systems. This work explores the use of RBR number representation for implementing high-throughput DSP systems that are also energy-efficient. Data-path components such as adders and multipliers are evaluated with respect to critical path delay, energy and Energy-Delay Product (EDP). A new design for a RBR adder with very good EDP performance has been proposed. The corresponding RBR parallel adder has a much lower critical path delay and EDP compared to two's complement carry select and carry look-ahead adder implementations. Next, several RBR multiplier architectures are investigated and their performance compared to two's complement systems. These include two new multiplier architectures: a purely RBR multiplier where both the operands are in RBR form, and a hybrid multiplier where the multiplicand is in RBR form and the other operand is represented in conventional two's complement form. Both the RBR and hybrid designs are demonstrated to have better EDP performance compared to conventional two's complement multipliers. The hybrid multiplier is also shown to have a superior EDP performance compared to the RBR multiplier, with much lower implementation area. Analysis on the effect of bit-precision is also performed, and it is shown that the performance gain of RBR systems improves for higher bit precision. Next, in order to demonstrate the efficacy of the RBR representation at the system-level, the performance of RBR and hybrid implementations of some common DSP kernels such as Discrete Cosine Transform, edge detection using Sobel operator, complex multiplication, Lifting-based Discrete Wavelet Transform (9, 7) filter, and FIR filter, is compared with two's complement systems. It is shown that for relatively large computation modules, the RBR to two's complement conversion overhead gets amortized. In case of systems with high complexity, for iso-throughput, both the hybrid and RBR implementations are demonstrated to be superior with lower average energy consumption. For low complexity systems, the conversion overhead is significant, and overpowers the EDP performance gain obtained from the RBR computation operation.
ContributorsMahadevan, Rupa (Author) / Chakrabarti, Chaitali (Thesis advisor) / Kiaei, Sayfe (Committee member) / Cao, Yu (Committee member) / Arizona State University (Publisher)
Created2011
150187-Thumbnail Image.png
Description
Genomic and proteomic sequences, which are in the form of deoxyribonucleic acid (DNA) and amino acids respectively, play a vital role in the structure, function and diversity of every living cell. As a result, various genomic and proteomic sequence processing methods have been proposed from diverse disciplines, including biology, chemistry,

Genomic and proteomic sequences, which are in the form of deoxyribonucleic acid (DNA) and amino acids respectively, play a vital role in the structure, function and diversity of every living cell. As a result, various genomic and proteomic sequence processing methods have been proposed from diverse disciplines, including biology, chemistry, physics, computer science and electrical engineering. In particular, signal processing techniques were applied to the problems of sequence querying and alignment, that compare and classify regions of similarity in the sequences based on their composition. However, although current approaches obtain results that can be attributed to key biological properties, they require pre-processing and lack robustness to sequence repetitions. In addition, these approaches do not provide much support for efficiently querying sub-sequences, a process that is essential for tracking localized database matches. In this work, a query-based alignment method for biological sequences that maps sequences to time-domain waveforms before processing the waveforms for alignment in the time-frequency plane is first proposed. The mapping uses waveforms, such as time-domain Gaussian functions, with unique sequence representations in the time-frequency plane. The proposed alignment method employs a robust querying algorithm that utilizes a time-frequency signal expansion whose basis function is matched to the basic waveform in the mapped sequences. The resulting WAVEQuery approach is demonstrated for both DNA and protein sequences using the matching pursuit decomposition as the signal basis expansion. The alignment localization of WAVEQuery is specifically evaluated over repetitive database segments, and operable in real-time without pre-processing. It is demonstrated that WAVEQuery significantly outperforms the biological sequence alignment method BLAST for queries with repetitive segments for DNA sequences. A generalized version of the WAVEQuery approach with the metaplectic transform is also described for protein sequence structure prediction. For protein alignment, it is often necessary to not only compare the one-dimensional (1-D) primary sequence structure but also the secondary and tertiary three-dimensional (3-D) space structures. This is done after considering the conformations in the 3-D space due to the degrees of freedom of these structures. As a result, a novel directionality based 3-D waveform mapping for the 3-D protein structures is also proposed and it is used to compare protein structures using a matched filter approach. By incorporating a 3-D time axis, a highly-localized Gaussian-windowed chirp waveform is defined, and the amino acid information is mapped to the chirp parameters that are then directly used to obtain directionality in the 3-D space. This mapping is unique in that additional characteristic protein information such as hydrophobicity, that relates the sequence with the structure, can be added as another representation parameter. The additional parameter helps tracking similarities over local segments of the structure, this enabling classification of distantly related proteins which have partial structural similarities. This approach is successfully tested for pairwise alignments over full length structures, alignments over multiple structures to form a phylogenetic trees, and also alignments over local segments. Also, basic classification over protein structural classes using directional descriptors for the protein structure is performed.
ContributorsRavichandran, Lakshminarayan (Author) / Papandreou-Suppappola, Antonia (Thesis advisor) / Spanias, Andreas S (Thesis advisor) / Chakrabarti, Chaitali (Committee member) / Tepedelenlioğlu, Cihan (Committee member) / Lacroix, Zoé (Committee member) / Arizona State University (Publisher)
Created2011
150085-Thumbnail Image.png
Description
The wood-framing trade has not sufficiently been investigated to understand the work task sequencing and coordination among crew members. A new mental framework for a performing crew was developed and tested through four case studies. This framework ensured similar team performance as the one provided by task micro-scheduling in planning

The wood-framing trade has not sufficiently been investigated to understand the work task sequencing and coordination among crew members. A new mental framework for a performing crew was developed and tested through four case studies. This framework ensured similar team performance as the one provided by task micro-scheduling in planning software. It also allowed evaluation of the effect of individual coordination within the crew on the crew's productivity. Using design information, a list of micro-activities/tasks and their predecessors was automatically generated for each piece of lumber in the four wood frames. The task precedence was generated by applying elementary geometrical and technological reasoning to each frame. Then, the duration of each task was determined based on observations from videotaped activities. Primavera's (P6) resource leveling rules were used to calculate the sequencing of tasks and the minimum duration of the whole activity for various crew sizes. The results showed quick convergence towards the minimum production time and allowed to use information from Building Information Models (BIM) to automatically establish the optimal crew sizes for frames. Late Start (LS) leveling priority rule gave the shortest duration in every case. However, the logic of LS tasks rule is too complex to be conveyed to the framing crew. Therefore, the new mental framework of a well performing framer was developed and tested to ensure high coordination. This mental framework, based on five simple rules, can be easily taught to the crew and ensures a crew productivity congruent with the one provided by the LS logic. The case studies indicate that once the worst framer in the crew surpasses the limit of 11% deviation from applying the said five rules, every additional percent of deviation reduces the productivity of the whole crew by about 4%.
ContributorsMaghiar, Marcel M (Author) / Wiezel, Avi (Thesis advisor) / Mitropoulos, Panagiotis (Committee member) / Cooke, Nancy J. (Committee member) / Arizona State University (Publisher)
Created2011
150139-Thumbnail Image.png
Description
Although there are many forms of organization on the Web, one of the most prominent ways to organize web content and websites are tags. Tags are keywords or terms that are assigned to a specific piece of content in order to help users understand the common relationships between pieces of

Although there are many forms of organization on the Web, one of the most prominent ways to organize web content and websites are tags. Tags are keywords or terms that are assigned to a specific piece of content in order to help users understand the common relationships between pieces of content. Tags can either be assigned by an algorithm, the author, or the community. These tags can also be organized into tag clouds, which are visual representations of the structure and organization contained implicitly within these tags. Importantly, little is known on how we use these different tagging structures to understand the content and structure of a given site. This project examines 2 different characteristics of tagging structures: font size and spatial orientation. In order to examine how these different characteristics might interact with individual differences in attentional control, a measure of working memory capacity (WMC) was included. The results showed that spatial relationships affect how well users understand the structure of a website. WMC was not shown to have any significant effect; neither was varying the font size. These results should better inform how tags and tag clouds are used on the Web, and also provide an estimation of what properties to include when designing and implementing a tag cloud on a website.
ContributorsBanas, Steven (Author) / Sanchez, Christopher A (Thesis advisor) / Branaghan, Russell (Committee member) / Cooke, Nancy J. (Committee member) / Arizona State University (Publisher)
Created2011
152360-Thumbnail Image.png
Description
In this work, we present approximate adders and multipliers to reduce data-path complexity of specialized hardware for various image processing systems. These approximate circuits have a lower area, latency and power consumption compared to their accurate counterparts and produce fairly accurate results. We build upon the work on approximate adders

In this work, we present approximate adders and multipliers to reduce data-path complexity of specialized hardware for various image processing systems. These approximate circuits have a lower area, latency and power consumption compared to their accurate counterparts and produce fairly accurate results. We build upon the work on approximate adders and multipliers presented in [23] and [24]. First, we show how choice of algorithm and parallel adder design can be used to implement 2D Discrete Cosine Transform (DCT) algorithm with good performance but low area. Our implementation of the 2D DCT has comparable PSNR performance with respect to the algorithm presented in [23] with ~35-50% reduction in area. Next, we use the approximate 2x2 multiplier presented in [24] to implement parallel approximate multipliers. We demonstrate that if some of the 2x2 multipliers in the design of the parallel multiplier are accurate, the accuracy of the multiplier improves significantly, especially when two large numbers are multiplied. We choose Gaussian FIR Filter and Fast Fourier Transform (FFT) algorithms to illustrate the efficacy of our proposed approximate multiplier. We show that application of the proposed approximate multiplier improves the PSNR performance of 32x32 FFT implementation by 4.7 dB compared to the implementation using the approximate multiplier described in [24]. We also implement a state-of-the-art image enlargement algorithm, namely Segment Adaptive Gradient Angle (SAGA) [29], in hardware. The algorithm is mapped to pipelined hardware blocks and we synthesized the design using 90 nm technology. We show that a 64x64 image can be processed in 496.48 µs when clocked at 100 MHz. The average PSNR performance of our implementation using accurate parallel adders and multipliers is 31.33 dB and that using approximate parallel adders and multipliers is 30.86 dB, when evaluated against the original image. The PSNR performance of both designs is comparable to the performance of the double precision floating point MATLAB implementation of the algorithm.
ContributorsVasudevan, Madhu (Author) / Chakrabarti, Chaitali (Thesis advisor) / Frakes, David (Committee member) / Gupta, Sandeep (Committee member) / Arizona State University (Publisher)
Created2013
152344-Thumbnail Image.png
Description
Structural integrity is an important characteristic of performance for critical components used in applications such as aeronautics, materials, construction and transportation. When appraising the structural integrity of these components, evaluation methods must be accurate. In addition to possessing capability to perform damage detection, the ability to monitor the level of

Structural integrity is an important characteristic of performance for critical components used in applications such as aeronautics, materials, construction and transportation. When appraising the structural integrity of these components, evaluation methods must be accurate. In addition to possessing capability to perform damage detection, the ability to monitor the level of damage over time can provide extremely useful information in assessing the operational worthiness of a structure and in determining whether the structure should be repaired or removed from service. In this work, a sequential Bayesian approach with active sensing is employed for monitoring crack growth within fatigue-loaded materials. The monitoring approach is based on predicting crack damage state dynamics and modeling crack length observations. Since fatigue loading of a structural component can change while in service, an interacting multiple model technique is employed to estimate probabilities of different loading modes and incorporate this information in the crack length estimation problem. For the observation model, features are obtained from regions of high signal energy in the time-frequency plane and modeled for each crack length damage condition. Although this observation model approach exhibits high classification accuracy, the resolution characteristics can change depending upon the extent of the damage. Therefore, several different transmission waveforms and receiver sensors are considered to create multiple modes for making observations of crack damage. Resolution characteristics of the different observation modes are assessed using a predicted mean squared error criterion and observations are obtained using the predicted, optimal observation modes based on these characteristics. Calculation of the predicted mean square error metric can be computationally intensive, especially if performed in real time, and an approximation method is proposed. With this approach, the real time computational burden is decreased significantly and the number of possible observation modes can be increased. Using sensor measurements from real experiments, the overall sequential Bayesian estimation approach, with the adaptive capability of varying the state dynamics and observation modes, is demonstrated for tracking crack damage.
ContributorsHuff, Daniel W (Author) / Papandreou-Suppappola, Antonia (Thesis advisor) / Kovvali, Narayan (Committee member) / Chakrabarti, Chaitali (Committee member) / Chattopadhyay, Aditi (Committee member) / Arizona State University (Publisher)
Created2013