This collection includes both ASU Theses and Dissertations, submitted by graduate students, and the Barrett, Honors College theses submitted by undergraduate students. 

Displaying 1 - 10 of 145
151718-Thumbnail Image.png
Description
The increasing popularity of Twitter renders improved trustworthiness and relevance assessment of tweets much more important for search. However, given the limitations on the size of tweets, it is hard to extract measures for ranking from the tweet's content alone. I propose a method of ranking tweets by generating a

The increasing popularity of Twitter renders improved trustworthiness and relevance assessment of tweets much more important for search. However, given the limitations on the size of tweets, it is hard to extract measures for ranking from the tweet's content alone. I propose a method of ranking tweets by generating a reputation score for each tweet that is based not just on content, but also additional information from the Twitter ecosystem that consists of users, tweets, and the web pages that tweets link to. This information is obtained by modeling the Twitter ecosystem as a three-layer graph. The reputation score is used to power two novel methods of ranking tweets by propagating the reputation over an agreement graph based on tweets' content similarity. Additionally, I show how the agreement graph helps counter tweet spam. An evaluation of my method on 16~million tweets from the TREC 2011 Microblog Dataset shows that it doubles the precision over baseline Twitter Search and achieves higher precision than current state of the art method. I present a detailed internal empirical evaluation of RAProp in comparison to several alternative approaches proposed by me, as well as external evaluation in comparison to the current state of the art method.
ContributorsRavikumar, Srijith (Author) / Kambhampati, Subbarao (Thesis advisor) / Davulcu, Hasan (Committee member) / Liu, Huan (Committee member) / Arizona State University (Publisher)
Created2013
152236-Thumbnail Image.png
Description
Continuous Delivery, as one of the youngest and most popular member of agile model family, has become a popular concept and method in software development industry recently. Instead of the traditional software development method, which requirements and solutions must be fixed before starting software developing, it promotes adaptive planning, evolutionary

Continuous Delivery, as one of the youngest and most popular member of agile model family, has become a popular concept and method in software development industry recently. Instead of the traditional software development method, which requirements and solutions must be fixed before starting software developing, it promotes adaptive planning, evolutionary development and delivery, and encourages rapid and flexible response to change. However, several problems prevent Continuous Delivery to be introduced into education world. Taking into the consideration of the barriers, we propose a new Cloud based Continuous Delivery Software Developing System. This system is designed to fully utilize the whole life circle of software developing according to Continuous Delivery concepts in a virtualized environment in Vlab platform.
ContributorsDeng, Yuli (Author) / Huang, Dijiang (Thesis advisor) / Davulcu, Hasan (Committee member) / Chen, Yinong (Committee member) / Arizona State University (Publisher)
Created2013
152136-Thumbnail Image.png
Description
Reductive dechlorination by members of the bacterial genus Dehalococcoides is a common and cost-effective avenue for in situ bioremediation of sites contaminated with the chlorinated solvents, trichloroethene (TCE) and perchloroethene (PCE). The overarching goal of my research was to address some of the challenges associated with bioremediation timeframes by improving

Reductive dechlorination by members of the bacterial genus Dehalococcoides is a common and cost-effective avenue for in situ bioremediation of sites contaminated with the chlorinated solvents, trichloroethene (TCE) and perchloroethene (PCE). The overarching goal of my research was to address some of the challenges associated with bioremediation timeframes by improving the rates of reductive dechlorination and the growth of Dehalococcoides in mixed communities. Biostimulation of contaminated sites or microcosms with electron donor fails to consistently promote dechlorination of PCE/TCE beyond cis-dichloroethene (cis-DCE), even when the presence of Dehalococcoides is confirmed. Supported by data from microcosm experiments, I showed that the stalling at cis-DCE is due a H2 competition in which components of the soil or sediment serve as electron acceptors for competing microorganisms. However, once competition was minimized by providing selective enrichment techniques, I illustrated how to obtain both fast rates and high-density Dehalococcoides using three distinct enrichment cultures. Having achieved a heightened awareness of the fierce competition for electron donor, I then identified bicarbonate (HCO3-) as a potential H2 sink for reductive dechlorination. HCO3- is the natural buffer in groundwater but also the electron acceptor for hydrogenotrophic methanogens and homoacetogens, two microbial groups commonly encountered with Dehalococcoides. By testing a range of concentrations in batch experiments, I showed that methanogens are favored at low HCO3 and homoacetogens at high HCO3-. The high HCO3- concentrations increased the H2 demand which negatively affected the rates and extent of dechlorination. By applying the gained knowledge on microbial community management, I ran the first successful continuous stirred-tank reactor (CSTR) at a 3-d hydraulic retention time for cultivation of dechlorinating cultures. I demonstrated that using carefully selected conditions in a CSTR, cultivation of Dehalococcoides at short retention times is feasible, resulting in robust cultures capable of fast dechlorination. Lastly, I provide a systematic insight into the effect of high ammonia on communities involved in dechlorination of chloroethenes. This work documents the potential use of landfill leachate as a substrate for dechlorination and an increased tolerance of Dehalococcoides to high ammonia concentrations (2 g L-1 NH4+-N) without loss of the ability to dechlorinate TCE to ethene.
ContributorsDelgado, Anca Georgiana (Author) / Krajmalnik-Brown, Rosa (Thesis advisor) / Cadillo-Quiroz, Hinsby (Committee member) / Halden, Rolf U. (Committee member) / Rittmann, Bruce E. (Committee member) / Stout, Valerie (Committee member) / Arizona State University (Publisher)
Created2013
151371-Thumbnail Image.png
Description
This dissertation presents the Temporal Event Query Language (TEQL), a new language for querying event streams. Event Stream Processing enables online querying of streams of events to extract relevant data in a timely manner. TEQL enables querying of interval-based event streams using temporal database operators. Temporal databases and temporal query

This dissertation presents the Temporal Event Query Language (TEQL), a new language for querying event streams. Event Stream Processing enables online querying of streams of events to extract relevant data in a timely manner. TEQL enables querying of interval-based event streams using temporal database operators. Temporal databases and temporal query languages have been a subject of research for more than 30 years and are a natural fit for expressing queries that involve a temporal dimension. However, operators developed in this context cannot be directly applied to event streams. The research extends a preexisting relational framework for event stream processing to support temporal queries. The language features and formal semantic extensions to extend the relational framework are identified. The extended framework supports continuous, step-wise evaluation of temporal queries. The incremental evaluation of TEQL operators is formalized to avoid re-computation of previous results. The research includes the development of a prototype that supports the integrated event and temporal query processing framework, with support for incremental evaluation and materialization of intermediate results. TEQL enables reporting temporal data in the output, direct specification of conditions over timestamps, and specification of temporal relational operators. Through the integration of temporal database operators with event languages, a new class of temporal queries is made possible for querying event streams. New features include semantic aggregation, extraction of temporal patterns using set operators, and a more accurate specification of event co-occurrence.
ContributorsShiva, Foruhar Ali (Author) / Urban, Susan D (Thesis advisor) / Chen, Yi (Thesis advisor) / Davulcu, Hasan (Committee member) / Sarjoughian, Hessam S. (Committee member) / Arizona State University (Publisher)
Created2012
151380-Thumbnail Image.png
Description
Microbial mat communities that inhabit hot springs in Yellowstone National Park have been studied for their biodiversity, energetics and evolutionary history, yet little is know about how these communities cope with nutrient limitation. In the present study the changes in assimilatory gene expression levels for nitrogen (nrgA), phosphorus (phoA), and

Microbial mat communities that inhabit hot springs in Yellowstone National Park have been studied for their biodiversity, energetics and evolutionary history, yet little is know about how these communities cope with nutrient limitation. In the present study the changes in assimilatory gene expression levels for nitrogen (nrgA), phosphorus (phoA), and iron (yusV) were measured under various nutrient enrichment experiments. While results for nrgA and phoA were inconclusive, results for yusV showed an increase in expression with the addition of N and Fe. This is the first data that shows the impact of nutrients on siderophore uptake regulation in hot spring microbes.
ContributorsThorne, Michele (Author) / Elser, James J (Thesis advisor) / Touchman, Jeffrey (Committee member) / Stout, Valerie (Committee member) / Arizona State University (Publisher)
Created2012
151275-Thumbnail Image.png
Description
The pay-as-you-go economic model of cloud computing increases the visibility, traceability, and verifiability of software costs. Application developers must understand how their software uses resources when running in the cloud in order to stay within budgeted costs and/or produce expected profits. Cloud computing's unique economic model also leads naturally to

The pay-as-you-go economic model of cloud computing increases the visibility, traceability, and verifiability of software costs. Application developers must understand how their software uses resources when running in the cloud in order to stay within budgeted costs and/or produce expected profits. Cloud computing's unique economic model also leads naturally to an earn-as-you-go profit model for many cloud based applications. These applications can benefit from low level analyses for cost optimization and verification. Testing cloud applications to ensure they meet monetary cost objectives has not been well explored in the current literature. When considering revenues and costs for cloud applications, the resource economic model can be scaled down to the transaction level in order to associate source code with costs incurred while running in the cloud. Both static and dynamic analysis techniques can be developed and applied to understand how and where cloud applications incur costs. Such analyses can help optimize (i.e. minimize) costs and verify that they stay within expected tolerances. An adaptation of Worst Case Execution Time (WCET) analysis is presented here to statically determine worst case monetary costs of cloud applications. This analysis is used to produce an algorithm for determining control flow paths within an application that can exceed a given cost threshold. The corresponding results are used to identify path sections that contribute most to cost excess. A hybrid approach for determining cost excesses is also presented that is comprised mostly of dynamic measurements but that also incorporates calculations that are based on the static analysis approach. This approach uses operational profiles to increase the precision and usefulness of the calculations.
ContributorsBuell, Kevin, Ph.D (Author) / Collofello, James (Thesis advisor) / Davulcu, Hasan (Committee member) / Lindquist, Timothy (Committee member) / Sen, Arunabha (Committee member) / Arizona State University (Publisher)
Created2012
151467-Thumbnail Image.png
Description
A semiconductor supply chain modeling and simulation platform using Linear Program (LP) optimization and parallel Discrete Event System Specification (DEVS) process models has been developed in a joint effort by ASU and Intel Corporation. A Knowledge Interchange Broker (KIBDEVS/LP) was developed to broker information synchronously between the DEVS and LP

A semiconductor supply chain modeling and simulation platform using Linear Program (LP) optimization and parallel Discrete Event System Specification (DEVS) process models has been developed in a joint effort by ASU and Intel Corporation. A Knowledge Interchange Broker (KIBDEVS/LP) was developed to broker information synchronously between the DEVS and LP models. Recently a single-echelon heuristic Inventory Strategy Module (ISM) was added to correct for forecast bias in customer demand data using different smoothing techniques. The optimization model could then use information provided by the forecast model to make better decisions for the process model. The composition of ISM with LP and DEVS models resulted in the first realization of what is now called the Optimization Simulation Forecast (OSF) platform. It could handle a single echelon supply chain system consisting of single hubs and single products In this thesis, this single-echelon simulation platform is extended to handle multiple echelons with multiple inventory elements handling multiple products. The main aspect for the multi-echelon OSF platform was to extend the KIBDEVS/LP such that ISM interactions with the LP and DEVS models could also be supported. To achieve this, a new, scalable XML schema for the KIB has been developed. The XML schema has also resulted in strengthening the KIB execution engine design. A sequential scheme controls the executions of the DEVS-Suite simulator, CPLEX optimizer, and ISM engine. To use the ISM for multiple echelons, it is extended to compute forecast customer demands and safety stocks over multiple hubs and products. Basic examples for semiconductor manufacturing spanning single and two echelon supply chain systems have been developed and analyzed. Experiments using perfect data were conducted to show the correctness of the OSF platform design and implementation. Simple, but realistic experiments have also been conducted. They highlight the kinds of supply chain dynamics that can be evaluated using discrete event process simulation, linear programming optimization, and heuristics forecasting models.
ContributorsSmith, James Melkon (Author) / Sarjoughian, Hessam S. (Thesis advisor) / Davulcu, Hasan (Committee member) / Fainekos, Georgios (Committee member) / Arizona State University (Publisher)
Created2012
151524-Thumbnail Image.png
Description
Process migration is a heavily studied research area and has a number of applications in distributed systems. Process migration means transferring a process running on one machine to another such that it resumes execution from the point at which it was suspended. The conventional approach to implement process migration is

Process migration is a heavily studied research area and has a number of applications in distributed systems. Process migration means transferring a process running on one machine to another such that it resumes execution from the point at which it was suspended. The conventional approach to implement process migration is to move the entire state information of the process (including hardware context, virtual memory, files etc.) from one machine to another. Copying all the state information is costly. This thesis proposes and demonstrates a new approach of migrating a process between two cores of Intel Single Chip Cloud (SCC), an experimental 48-core processor by Intel, with each core running a separate instance of the operating system. In this method the amount of process state to be transferred from one core's memory to another is reduced by making use of special registers called Lookup tables (LUTs) present on each core of SCC. Thus this new approach is faster than the conventional method.
ContributorsJain, Vaibhav (Author) / Dasgupta, Partha (Thesis advisor) / Shriavstava, Aviral (Committee member) / Davulcu, Hasan (Committee member) / Arizona State University (Publisher)
Created2013
152310-Thumbnail Image.png
Description
The wide adoption and continued advancement of information and communications technologies (ICT) have made it easier than ever for individuals and groups to stay connected over long distances. These advances have greatly contributed in dramatically changing the dynamics of the modern day workplace to the point where it is now

The wide adoption and continued advancement of information and communications technologies (ICT) have made it easier than ever for individuals and groups to stay connected over long distances. These advances have greatly contributed in dramatically changing the dynamics of the modern day workplace to the point where it is now commonplace to see large, distributed multidisciplinary teams working together on a daily basis. However, in this environment, motivating, understanding, and valuing the diverse contributions of individual workers in collaborative enterprises becomes challenging. To address these issues, this thesis presents the goals, design, and implementation of Taskville, a distributed workplace game played by teams on large, public displays. Taskville uses a city building metaphor to represent the completion of individual and group tasks within an organization. Promising results from two usability studies and two longitudinal studies at a multidisciplinary school demonstrate that Taskville supports personal reflection and improves team awareness through an engaging workplace activity.
ContributorsNikkila, Shawn (Author) / Sundaram, Hari (Thesis advisor) / Byrne, Daragh (Committee member) / Davulcu, Hasan (Committee member) / Olson, Loren (Committee member) / Arizona State University (Publisher)
Created2013
152583-Thumbnail Image.png
Description
The prevalence of antibiotic resistant bacterial pathogens has increased since the introduction of penicillin in the 1940s. Insufficient development of novel antibacterial agents is leaving us with a failing arsenal of therapies to combat these pathogenic organisms. We have identified a clay mineral mixture (designated CB) that exhibits in vitro

The prevalence of antibiotic resistant bacterial pathogens has increased since the introduction of penicillin in the 1940s. Insufficient development of novel antibacterial agents is leaving us with a failing arsenal of therapies to combat these pathogenic organisms. We have identified a clay mineral mixture (designated CB) that exhibits in vitro antibacterial activity against a broad spectrum of bacterial pathogens, yet the antibacterial mechanism of action remains unknown. Antibacterial susceptibility testing of four different clay samples collected from the same source revealed that these natural clays had markedly different antibacterial activity. X-ray diffraction analyses of these minerals revealed minor mineralogical differences across the samples; however, ICP analyses demonstrated that the concentrations of many elements, Fe, Co, Cu, Ni, and Zn in particular, vary greatly across the four clay mixture leachates. Supplementation of a non-antibacterial leachate containing lower concentrations of Fe, Co, Ni, Cu, and Zn to final ion concentrations and a pH equivalent to that of the antibacterial leachate resulted in antibacterial activity against E. coli and MRSA, confirming the role of these ions in the in vitro antibacterial clay mixture leachates. The prevailing hypothesis is that metal ions participate in redox cycling and produce ROS, leading to oxidative damage to macromolecules and resulting in cellular death. However, E. coli cells showed no increase in DNA or protein oxidative lesions and a slight increase in lipid peroxidation following exposure to CB-L. Supplementation of CB-L with ROS scavengers eliminated oxidative damage in E. coli, but did not rescue the cells from killing, indicating that in vitro killing is due to direct metal toxicity and not to indirect oxidative damage. Finally, we ion-exchanged non-antibacterial clays with Fe, Co, Cu, and Zn and established antibacterial activity in these samples. Treatment of MRSA skin infections with both natural and ion-exchanged clays significantly decreased the bacterial load after 7 days of treatment. We conclude that 1) in vitro clay-mediated killing is due to toxicity associated directly with released metal ions and not to indirect oxidative damage and 2) that in vivo killing is due to the physical properties of the clays rather than metal ion toxicity.
ContributorsOtto, Caitin Carol (Author) / Haydel, Shelley (Thesis advisor) / Stout, Valerie (Committee member) / Roberson, Robby (Committee member) / Sandrin, Todd (Committee member) / Rege, Kaushal (Committee member) / Arizona State University (Publisher)
Created2014