This collection includes most of the ASU Theses and Dissertations from 2011 to present. ASU Theses and Dissertations are available in downloadable PDF format; however, a small percentage of items are under embargo. Information about the dissertations/theses includes degree information, committee members, an abstract, supporting data or media.

In addition to the electronic theses found in the ASU Digital Repository, ASU Theses and Dissertations can be found in the ASU Library Catalog.

Dissertations and Theses granted by Arizona State University are archived and made available through a joint effort of the ASU Graduate College and the ASU Libraries. For more information or questions about this collection contact or visit the Digital Repository ETD Library Guide or contact the ASU Graduate College at gradformat@asu.edu.

Displaying 1 - 10 of 136
Filtering by

Clear all filters

152255-Thumbnail Image.png
Description
Many manmade chemicals used in consumer products are ultimately washed down the drain and are collected in municipal sewers. Efficient chemical monitoring at wastewater treatment (WWT) plants thus may provide up-to-date information on chemical usage rates for epidemiological assessments. The objective of the present study was to extrapolate this concept,

Many manmade chemicals used in consumer products are ultimately washed down the drain and are collected in municipal sewers. Efficient chemical monitoring at wastewater treatment (WWT) plants thus may provide up-to-date information on chemical usage rates for epidemiological assessments. The objective of the present study was to extrapolate this concept, termed 'sewage epidemiology', to include municipal sewage sludge (MSS) in identifying and prioritizing contaminants of emerging concern (CECs). To test this the following specific aims were defined: i) to screen and identify CECs in nationally representative samples of MSS and to provide nationwide inventories of CECs in U.S. MSS; ii) to investigate the fate and persistence in MSS-amended soils, of sludge-borne hydrophobic CECs; and iii) to develop an analytical tool relying on contaminant levels in MSS as an indicator for identifying and prioritizing hydrophobic CECs. Chemicals that are primarily discharged to the sewage systems (alkylphenol surfactants) and widespread persistent organohalogen pollutants (perfluorochemicals and brominated flame retardants) were analyzed in nationally representative MSS samples. A meta-analysis showed that CECs contribute about 0.04-0.15% to the total dry mass of MSS, a mass equivalent of 2,700-7,900 metric tonnes of chemicals annually. An analysis of archived mesocoms from a sludge weathering study showed that 64 CECs persisted in MSS/soil mixtures over the course of the experiment, with half-lives ranging between 224 and >990 days; these results suggest an inherent persistence of CECs that accumulate in MSS. A comparison of the spectrum of chemicals (n=52) analyzed in nationally representative biological specimens from humans and MSS revealed 70% overlap. This observed co-occurrence of contaminants in both matrices suggests that MSS may serve as an indicator for ongoing human exposures and body burdens of pollutants in humans. In conclusion, I posit that this novel approach in sewage epidemiology may serve to pre-screen and prioritize the several thousands of known or suspected CECs to identify those that are most prone to pose a risk to human health and the environment.
ContributorsVenkatesan, Arjunkrishna (Author) / Halden, Rolf U. (Thesis advisor) / Westerhoff, Paul (Committee member) / Fox, Peter (Committee member) / Arizona State University (Publisher)
Created2013
151867-Thumbnail Image.png
Description
Automating aspects of biocuration through biomedical information extraction could significantly impact biomedical research by enabling greater biocuration throughput and improving the feasibility of a wider scope. An important step in biomedical information extraction systems is named entity recognition (NER), where mentions of entities such as proteins and diseases are located

Automating aspects of biocuration through biomedical information extraction could significantly impact biomedical research by enabling greater biocuration throughput and improving the feasibility of a wider scope. An important step in biomedical information extraction systems is named entity recognition (NER), where mentions of entities such as proteins and diseases are located within natural-language text and their semantic type is determined. This step is critical for later tasks in an information extraction pipeline, including normalization and relationship extraction. BANNER is a benchmark biomedical NER system using linear-chain conditional random fields and the rich feature set approach. A case study with BANNER locating genes and proteins in biomedical literature is described. The first corpus for disease NER adequate for use as training data is introduced, and employed in a case study of disease NER. The first corpus locating adverse drug reactions (ADRs) in user posts to a health-related social website is also described, and a system to locate and identify ADRs in social media text is created and evaluated. The rich feature set approach to creating NER feature sets is argued to be subject to diminishing returns, implying that additional improvements may require more sophisticated methods for creating the feature set. This motivates the first application of multivariate feature selection with filters and false discovery rate analysis to biomedical NER, resulting in a feature set at least 3 orders of magnitude smaller than the set created by the rich feature set approach. Finally, two novel approaches to NER by modeling the semantics of token sequences are introduced. The first method focuses on the sequence content by using language models to determine whether a sequence resembles entries in a lexicon of entity names or text from an unlabeled corpus more closely. The second method models the distributional semantics of token sequences, determining the similarity between a potential mention and the token sequences from the training data by analyzing the contexts where each sequence appears in a large unlabeled corpus. The second method is shown to improve the performance of BANNER on multiple data sets.
ContributorsLeaman, James Robert (Author) / Gonzalez, Graciela (Thesis advisor) / Baral, Chitta (Thesis advisor) / Cohen, Kevin B (Committee member) / Liu, Huan (Committee member) / Ye, Jieping (Committee member) / Arizona State University (Publisher)
Created2013
151940-Thumbnail Image.png
Description
Biological systems are complex in many dimensions as endless transportation and communication networks all function simultaneously. Our ability to intervene within both healthy and diseased systems is tied directly to our ability to understand and model core functionality. The progress in increasingly accurate and thorough high-throughput measurement technologies has provided

Biological systems are complex in many dimensions as endless transportation and communication networks all function simultaneously. Our ability to intervene within both healthy and diseased systems is tied directly to our ability to understand and model core functionality. The progress in increasingly accurate and thorough high-throughput measurement technologies has provided a deluge of data from which we may attempt to infer a representation of the true genetic regulatory system. A gene regulatory network model, if accurate enough, may allow us to perform hypothesis testing in the form of computational experiments. Of great importance to modeling accuracy is the acknowledgment of biological contexts within the models -- i.e. recognizing the heterogeneous nature of the true biological system and the data it generates. This marriage of engineering, mathematics and computer science with systems biology creates a cycle of progress between computer simulation and lab experimentation, rapidly translating interventions and treatments for patients from the bench to the bedside. This dissertation will first discuss the landscape for modeling the biological system, explore the identification of targets for intervention in Boolean network models of biological interactions, and explore context specificity both in new graphical depictions of models embodying context-specific genomic regulation and in novel analysis approaches designed to reveal embedded contextual information. Overall, the dissertation will explore a spectrum of biological modeling with a goal towards therapeutic intervention, with both formal and informal notions of biological context, in such a way that will enable future work to have an even greater impact in terms of direct patient benefit on an individualized level.
ContributorsVerdicchio, Michael (Author) / Kim, Seungchan (Thesis advisor) / Baral, Chitta (Committee member) / Stolovitzky, Gustavo (Committee member) / Collofello, James (Committee member) / Arizona State University (Publisher)
Created2013
151951-Thumbnail Image.png
Description
The consumption of feedstocks from agriculture and forestry by current biofuel production has raised concerns about food security and land availability. In the meantime, intensive human activities have created a large amount of marginal lands that require management. This study investigated the viability of aligning land management with biofuel production

The consumption of feedstocks from agriculture and forestry by current biofuel production has raised concerns about food security and land availability. In the meantime, intensive human activities have created a large amount of marginal lands that require management. This study investigated the viability of aligning land management with biofuel production on marginal lands. Biofuel crop production on two types of marginal lands, namely urban vacant lots and abandoned mine lands (AMLs), were assessed. The investigation of biofuel production on urban marginal land was carried out in Pittsburgh between 2008 and 2011, using the sunflower gardens developed by a Pittsburgh non-profit as an example. Results showed that the crops from urban marginal lands were safe for biofuel. The crop yield was 20% of that on agricultural land while the low input agriculture was used in crop cultivation. The energy balance analysis demonstrated that the sunflower gardens could produce a net energy return even at the current low yield. Biofuel production on AML was assessed from experiments conducted in a greenhouse for sunflower, soybean, corn, canola and camelina. The research successfully created an industrial symbiosis by using bauxite as soil amendment to enable plant growth on very acidic mine refuse. Phytoremediation and soil amendments were found to be able to effectively reduce contamination in the AML and its runoff. Results from this research supported that biofuel production on marginal lands could be a unique and feasible option for cultivating biofuel feedstocks.
ContributorsZhao, Xi (Author) / Landis, Amy (Thesis advisor) / Fox, Peter (Committee member) / Chester, Mikhail (Committee member) / Arizona State University (Publisher)
Created2013
151963-Thumbnail Image.png
Description
Currently, to interact with computer based systems one needs to learn the specific interface language of that system. In most cases, interaction would be much easier if it could be done in natural language. For that, we will need a module which understands natural language and automatically translates it to

Currently, to interact with computer based systems one needs to learn the specific interface language of that system. In most cases, interaction would be much easier if it could be done in natural language. For that, we will need a module which understands natural language and automatically translates it to the interface language of the system. NL2KR (Natural language to knowledge representation) v.1 system is a prototype of such a system. It is a learning based system that learns new meanings of words in terms of lambda-calculus formulas given an initial lexicon of some words and their meanings and a training corpus of sentences with their translations. As a part of this thesis, we take the prototype NL2KR v.1 system and enhance various components of it to make it usable for somewhat substantial and useful interface languages. We revamped the lexicon learning components, Inverse-lambda and Generalization modules, and redesigned the lexicon learning algorithm which uses these components to learn new meanings of words. Similarly, we re-developed an inbuilt parser of the system in Answer Set Programming (ASP) and also integrated external parser with the system. Apart from this, we added some new rich features like various system configurations and memory cache in the learning component of the NL2KR system. These enhancements helped in learning more meanings of the words, boosted performance of the system by reducing the computation time by a factor of 8 and improved the usability of the system. We evaluated the NL2KR system on iRODS domain. iRODS is a rule-oriented data system, which helps in managing large set of computer files using policies. This system provides a Rule-Oriented interface langauge whose syntactic structure is like any procedural programming language (eg. C). However, direct translation of natural language (NL) to this interface language is difficult. So, for automatic translation of NL to this language, we define a simple intermediate Policy Declarative Language (IPDL) to represent the knowledge in the policies, which then can be directly translated to iRODS rules. We develop a corpus of 100 policy statements and manually translate them to IPDL langauge. This corpus is then used for the evaluation of NL2KR system. We performed 10 fold cross validation on the system. Furthermore, using this corpus, we illustrate how different components of our NL2KR system work.
ContributorsKumbhare, Kanchan Ravishankar (Author) / Baral, Chitta (Thesis advisor) / Ye, Jieping (Committee member) / Li, Baoxin (Committee member) / Arizona State University (Publisher)
Created2013
151793-Thumbnail Image.png
Description
Linear Temporal Logic is gaining increasing popularity as a high level specification language for robot motion planning due to its expressive power and scalability of LTL control synthesis algorithms. This formalism, however, requires expert knowledge and makes it inaccessible to non-expert users. This thesis introduces a graphical specification environment to

Linear Temporal Logic is gaining increasing popularity as a high level specification language for robot motion planning due to its expressive power and scalability of LTL control synthesis algorithms. This formalism, however, requires expert knowledge and makes it inaccessible to non-expert users. This thesis introduces a graphical specification environment to create high level motion plans to control robots in the field by converting a visual representation of the motion/task plan into a Linear Temporal Logic (LTL) specification. The visual interface is built on the Android tablet platform and provides functionality to create task plans through a set of well defined gestures and on screen controls. It uses the notion of waypoints to quickly and efficiently describe the motion plan and enables a variety of complex Linear Temporal Logic specifications to be described succinctly and intuitively by the user without the need for the knowledge and understanding of LTL specification. Thus, it opens avenues for its use by personnel in military, warehouse management, and search and rescue missions. This thesis describes the construction of LTL for various scenarios used for robot navigation using the visual interface developed and leverages the use of existing LTL based motion planners to carry out the task plan by a robot.
ContributorsSrinivas, Shashank (Author) / Fainekos, Georgios (Thesis advisor) / Baral, Chitta (Committee member) / Burleson, Winslow (Committee member) / Arizona State University (Publisher)
Created2013
151653-Thumbnail Image.png
Description
Answer Set Programming (ASP) is one of the most prominent and successful knowledge representation paradigms. The success of ASP is due to its expressive non-monotonic modeling language and its efficient computational methods originating from building propositional satisfiability solvers. The wide adoption of ASP has motivated several extensions to its modeling

Answer Set Programming (ASP) is one of the most prominent and successful knowledge representation paradigms. The success of ASP is due to its expressive non-monotonic modeling language and its efficient computational methods originating from building propositional satisfiability solvers. The wide adoption of ASP has motivated several extensions to its modeling language in order to enhance expressivity, such as incorporating aggregates and interfaces with ontologies. Also, in order to overcome the grounding bottleneck of computation in ASP, there are increasing interests in integrating ASP with other computing paradigms, such as Constraint Programming (CP) and Satisfiability Modulo Theories (SMT). Due to the non-monotonic nature of the ASP semantics, such enhancements turned out to be non-trivial and the existing extensions are not fully satisfactory. We observe that one main reason for the difficulties rooted in the propositional semantics of ASP, which is limited in handling first-order constructs (such as aggregates and ontologies) and functions (such as constraint variables in CP and SMT) in natural ways. This dissertation presents a unifying view on these extensions by viewing them as instances of formulas with generalized quantifiers and intensional functions. We extend the first-order stable model semantics by by Ferraris, Lee, and Lifschitz to allow generalized quantifiers, which cover aggregate, DL-atoms, constraints and SMT theory atoms as special cases. Using this unifying framework, we study and relate different extensions of ASP. We also present a tight integration of ASP with SMT, based on which we enhance action language C+ to handle reasoning about continuous changes. Our framework yields a systematic approach to study and extend non-monotonic languages.
ContributorsMeng, Yunsong (Author) / Lee, Joohyung (Thesis advisor) / Ahn, Gail-Joon (Committee member) / Baral, Chitta (Committee member) / Fainekos, Georgios (Committee member) / Lifschitz, Vladimir (Committee member) / Arizona State University (Publisher)
Created2013
151471-Thumbnail Image.png
Description
In this dissertation I develop a deep theory of temporal planning well-suited to analyzing, understanding, and improving the state of the art implementations (as of 2012). At face-value the work is strictly theoretical; nonetheless its impact is entirely real and practical. The easiest portion of that impact to highlight concerns

In this dissertation I develop a deep theory of temporal planning well-suited to analyzing, understanding, and improving the state of the art implementations (as of 2012). At face-value the work is strictly theoretical; nonetheless its impact is entirely real and practical. The easiest portion of that impact to highlight concerns the notable improvements to the format of the temporal fragment of the International Planning Competitions (IPCs). Particularly: the theory I expound upon here is the primary cause of--and justification for--the altered (i) selection of benchmark problems, and (ii) notion of "winning temporal planner". For higher level motivation: robotics, web service composition, industrial manufacturing, business process management, cybersecurity, space exploration, deep ocean exploration, and logistics all benefit from applying domain-independent automated planning technique. Naturally, actually carrying out such case studies has much to offer. For example, we may extract the lesson that reasoning carefully about deadlines is rather crucial to planning in practice. More generally, effectively automating specifically temporal planning is well-motivated from applications. Entirely abstractly, the aim is to improve the theory of automated temporal planning by distilling from its practice. My thesis is that the key feature of computational interest is concurrency. To support, I demonstrate by way of compilation methods, worst-case counting arguments, and analysis of algorithmic properties such as completeness that the more immediately pressing computational obstacles (facing would-be temporal generalizations of classical planning systems) can be dealt with in theoretically efficient manner. So more accurately the technical contribution here is to demonstrate: The computationally significant obstacle to automated temporal planning that remains is just concurrency.
ContributorsCushing, William Albemarle (Author) / Kambhampati, Subbarao (Thesis advisor) / Weld, Daniel S. (Committee member) / Smith, David E. (Committee member) / Baral, Chitta (Committee member) / Davalcu, Hasan (Committee member) / Arizona State University (Publisher)
Created2012
152167-Thumbnail Image.png
Description
Contaminants of emerging concern (CECs) present in wastewater effluent can threat its safe discharge or reuse. Additional barriers of protection can be provided using advanced or natural treatment processes. This dissertation evaluated ozonation and constructed wetlands to remove CECs from wastewater effluent. Organic CECs can be removed by hydroxyl radical

Contaminants of emerging concern (CECs) present in wastewater effluent can threat its safe discharge or reuse. Additional barriers of protection can be provided using advanced or natural treatment processes. This dissertation evaluated ozonation and constructed wetlands to remove CECs from wastewater effluent. Organic CECs can be removed by hydroxyl radical formed during ozonation, however estimating the ozone demand of wastewater effluent is complicated due to the presence of reduced inorganic species. A method was developed to estimate ozone consumption only by dissolved organic compounds and predict trace organic oxidation across multiple wastewater sources. Organic and engineered nanomaterial (ENM) CEC removal in constructed wetlands was investigated using batch experiments and continuous-flow microcosms containing decaying wetland plants. CEC removal varied depending on their physico-chemical properties, hydraulic residence time (HRT) and relative quantities of plant materials in the microcosms. At comparable HRTs, ENM removal improved with higher quantity of plant materials due to enhanced sorption which was verified in batch-scale studies with plant materials. A fate-predictive model was developed to evaluate the role of design loading rates on organic CEC removal. Areal removal rates increased with hydraulic loading rates (HLRs) and carbon loading rates (CLRs) unless photolysis was the dominant removal mechanism (e.g. atrazine). To optimize CEC removal, wetlands with different CLRs can be used in combination without lowering the net HLR. Organic CEC removal in denitrifying conditions of constructed wetlands was investigated and selected CECs (e.g. estradiol) were found to biotransform while denitrification occurred. Although level of denitrification was affected by HRT, similar impact on estradiol was not observed due to a dominant effect from plant biomass quantity. Overall, both modeling and experimental findings suggest considering CLR as an equally important factor with HRT or HLR to design constructed wetlands for CEC removal. This dissertation provided directions to select design parameters for ozonation (ozone dose) and constructed wetlands (design loading rates) to meet organic CEC removal goals. Future research is needed to understand fate of ENMs during ozonation and quantify the contributions from different transformation mechanisms occurring in the wetlands to incorporate in a model and evaluate the effect of wetland design.
ContributorsSharif, Fariya (Author) / Westerhoff, Paul (Thesis advisor) / Halden, Rolf (Committee member) / Fox, Peter (Committee member) / Herckes, Pierre (Committee member) / Arizona State University (Publisher)
Created2013
152168-Thumbnail Image.png
Description
There has been a lot of research in the field of artificial intelligence about thinking machines. Alan Turing proposed a test to observe a machine's intelligent behaviour with respect to natural language conversation. The Winograd schema challenge is suggested as an alternative, to the Turing test. It needs inferencing capabilities,

There has been a lot of research in the field of artificial intelligence about thinking machines. Alan Turing proposed a test to observe a machine's intelligent behaviour with respect to natural language conversation. The Winograd schema challenge is suggested as an alternative, to the Turing test. It needs inferencing capabilities, reasoning abilities and background knowledge to get the answer right. It involves a coreference resolution task in which a machine is given a sentence containing a situation which involves two entities, one pronoun and some more information about the situation and the machine has to come up with the right resolution of a pronoun to one of the entities. The complexity of the task is increased with the fact that the Winograd sentences are not constrained by one domain or specific sentence structure and it also contains a lot of human proper names. This modification makes the task of association of entities, to one particular word in the sentence, to derive the answer, difficult. I have developed a pronoun resolver system for the confined domain Winograd sentences. I have developed a classifier or filter which takes input sentences and decides to accept or reject them based on a particular criteria. Once the sentence is accepted. I run parsers on it to obtain the detailed analysis. Furthermore I have developed four answering modules which use world knowledge and inferencing mechanisms to try and resolve the pronoun. The four techniques I use are : ConceptNet knowledgebase, Search engine pattern counts,Narrative event chains and sentiment analysis. I have developed a particular aggregation mechanism for the answers from these modules to arrive at a final answer. I have used caching technique for the association relations that I obtain for different modules, so as to boost the performance. I run my system on the standard ‘nyu dataset’ of Winograd sentences and questions. This dataset is then restricted, by my classifier, to 90 sentences. I evaluate my system on this 90 sentence dataset. When I compare my results against the state of the art system on the same dataset, I get nearly 4.5 % improvement in the restricted domain.
ContributorsBudukh, Tejas Ulhas (Author) / Baral, Chitta (Thesis advisor) / VanLehn, Kurt (Committee member) / Davulcu, Hasan (Committee member) / Arizona State University (Publisher)
Created2013