Matching Items (294)
150019-Thumbnail Image.png
Description
Currently Java is making its way into the embedded systems and mobile devices like androids. The programs written in Java are compiled into machine independent binary class byte codes. A Java Virtual Machine (JVM) executes these classes. The Java platform additionally specifies the Java Native Interface (JNI). JNI allows Java

Currently Java is making its way into the embedded systems and mobile devices like androids. The programs written in Java are compiled into machine independent binary class byte codes. A Java Virtual Machine (JVM) executes these classes. The Java platform additionally specifies the Java Native Interface (JNI). JNI allows Java code that runs within a JVM to interoperate with applications or libraries that are written in other languages and compiled to the host CPU ISA. JNI plays an important role in embedded system as it provides a mechanism to interact with libraries specific to the platform. This thesis addresses the overhead incurred in the JNI due to reflection and serialization when objects are accessed on android based mobile devices. It provides techniques to reduce this overhead. It also provides an API to access objects through its reference through pinning its memory location. The Android emulator was used to evaluate the performance of these techniques and we observed that there was 5 - 10 % performance gain in the new Java Native Interface.
ContributorsChandrian, Preetham (Author) / Lee, Yann-Hang (Thesis advisor) / Davulcu, Hasan (Committee member) / Li, Baoxin (Committee member) / Arizona State University (Publisher)
Created2011
150026-Thumbnail Image.png
Description
As pointed out in the keynote speech by H. V. Jagadish in SIGMOD'07, and also commonly agreed in the database community, the usability of structured data by casual users is as important as the data management systems' functionalities. A major hardness of using structured data is the problem of easily

As pointed out in the keynote speech by H. V. Jagadish in SIGMOD'07, and also commonly agreed in the database community, the usability of structured data by casual users is as important as the data management systems' functionalities. A major hardness of using structured data is the problem of easily retrieving information from them given a user's information needs. Learning and using a structured query language (e.g., SQL and XQuery) is overwhelmingly burdensome for most users, as not only are these languages sophisticated, but the users need to know the data schema. Keyword search provides us with opportunities to conveniently access structured data and potentially significantly enhances the usability of structured data. However, processing keyword search on structured data is challenging due to various types of ambiguities such as structural ambiguity (keyword queries have no structure), keyword ambiguity (the keywords may not be accurate), user preference ambiguity (the user may have implicit preferences that are not indicated in the query), as well as the efficiency challenges due to large search space. This dissertation performs an expansive study on keyword search processing techniques as a gateway for users to access structured data and retrieve desired information. The key issues addressed include: (1) Resolving structural ambiguities in keyword queries by generating meaningful query results, which involves identifying relevant keyword matches, identifying return information, composing query results based on relevant matches and return information. (2) Resolving structural, keyword and user preference ambiguities through result analysis, including snippet generation, result differentiation, result clustering, result summarization/query expansion, etc. (3) Resolving the efficiency challenge in processing keyword search on structured data by utilizing and efficiently maintaining materialized views. These works deliver significant technical contributions towards building a full-fledged search engine for structured data.
ContributorsLiu, Ziyang (Author) / Chen, Yi (Thesis advisor) / Candan, Kasim S (Committee member) / Davulcu, Hasan (Committee member) / Jagadish, H V (Committee member) / Arizona State University (Publisher)
Created2011
149681-Thumbnail Image.png
Description
The trend towards using recycled materials on new construction projects is growing as the cost for construction materials are ever increasing and the awareness of the responsibility we have to be good stewards of our environment is heightened. While recycled asphalt is sometimes used in pavements, its use as structural

The trend towards using recycled materials on new construction projects is growing as the cost for construction materials are ever increasing and the awareness of the responsibility we have to be good stewards of our environment is heightened. While recycled asphalt is sometimes used in pavements, its use as structural fill has been hindered by concern that it is susceptible to large long-term deformations (creep), preventing its use for a great many geotechnical applications. While asphalt/soil blends are often proposed as an alternative to 100% recycled asphalt fill, little data is available characterizing the geotechnical properties of recycled asphalt soil blends. In this dissertation, the geotechnical properties for five different recycled asphalt soil blends are characterized. Data includes the particle size distribution, plasticity index, creep, and shear strength for each blend. Blends with 0%, 25%, 50%, 75% and 100% recycled asphalt were tested. As the recycled asphalt material used for testing had particles sizes up to 1.5 inches, a large 18 inch diameter direct shear apparatus was used to determine the shear strength and creep characteristics of the material. The results of the testing program confirm that the creep potential of recycled asphalt is a geotechnical concern when the material is subjected to loads greater than 1500 pounds per square foot (psf). In addition, the test results demonstrate that the amount of soil blended with the recycled asphalt can greatly influence the creep and shear strength behavior of the composite material. Furthermore, there appears to be an optimal blend ratio where the composite material had better properties than either the recycled asphalt or virgin soil alone with respect to shear strength.
ContributorsSchaper, Jeffery M (Author) / Kavazanjian, Edward (Thesis advisor) / Houston, Sandra L. (Committee member) / Zapata, Claudia E (Committee member) / Arizona State University (Publisher)
Created2011
149728-Thumbnail Image.png
Description
In geotechnical engineering, measuring the unsaturated hydraulic conductivity of fine grained soils can be time consuming and tedious. The various applications that require knowledge of the unsaturated hydraulic conductivity function are great, and in geotechnical engineering, they range from modeling seepage through landfill covers to determining infiltration of water

In geotechnical engineering, measuring the unsaturated hydraulic conductivity of fine grained soils can be time consuming and tedious. The various applications that require knowledge of the unsaturated hydraulic conductivity function are great, and in geotechnical engineering, they range from modeling seepage through landfill covers to determining infiltration of water under a building slab. The unsaturated hydraulic conductivity function can be measured using various direct and indirect techniques. The instantaneous profile method has been found to be the most promising unsteady state method for measuring the unsaturated hydraulic conductivity function for fine grained soils over a wide range of suction values. The instantaneous profile method can be modified by using different techniques to measure suction and water content and also through the way water is introduced or removed from the soil profile. In this study, the instantaneous profile method was modified by creating duplicate soil samples compacted into cylindrical tubes at two different water contents. The techniques used in the duplicate method to measure the water content and matric suction included volumetric moisture probes, manual water content measurements, and filter paper tests. The experimental testing conducted in this study provided insight into determining the unsaturated hydraulic conductivity using the instantaneous profile method for a sandy clay soil and recommendations are provided for further evaluation. Overall, this study has demonstrated that the presence of cracks has no significant impact on the hydraulic behavior of soil in high suction ranges. The results of this study do not examine the behavior of cracked soil unsaturated hydraulic conductivity at low suction and at moisture contents near saturation.
ContributorsJacquemin, Sean Christopher (Author) / Zapata, Claudia (Thesis advisor) / Houston, Sandra (Committee member) / Kavazanjian, Edward (Committee member) / Arizona State University (Publisher)
Created2011
149794-Thumbnail Image.png
Description
Genes have widely different pertinences to the etiology and pathology of diseases. Thus, they can be ranked according to their disease-significance on a genomic scale, which is the subject of gene prioritization. Given a set of genes known to be related to a disease, it is reasonable to use them

Genes have widely different pertinences to the etiology and pathology of diseases. Thus, they can be ranked according to their disease-significance on a genomic scale, which is the subject of gene prioritization. Given a set of genes known to be related to a disease, it is reasonable to use them as a basis to determine the significance of other candidate genes, which will then be ranked based on the association they exhibit with respect to the given set of known genes. Experimental and computational data of various kinds have different reliability and relevance to a disease under study. This work presents a gene prioritization method based on integrated biological networks that incorporates and models the various levels of relevance and reliability of diverse sources. The method is shown to achieve significantly higher performance as compared to two well-known gene prioritization algorithms. Essentially, no bias in the performance was seen as it was applied to diseases of diverse ethnology, e.g., monogenic, polygenic and cancer. The method was highly stable and robust against significant levels of noise in the data. Biological networks are often sparse, which can impede the operation of associationbased gene prioritization algorithms such as the one presented here from a computational perspective. As a potential approach to overcome this limitation, we explore the value that transcription factor binding sites can have in elucidating suitable targets. Transcription factors are needed for the expression of most genes, especially in higher organisms and hence genes can be associated via their genetic regulatory properties. While each transcription factor recognizes specific DNA sequence patterns, such patterns are mostly unknown for many transcription factors. Even those that are known are inconsistently reported in the literature, implying a potentially high level of inaccuracy. We developed computational methods for prediction and improvement of transcription factor binding patterns. Tests performed on the improvement method by employing synthetic patterns under various conditions showed that the method is very robust and the patterns produced invariably converge to nearly identical series of patterns. Preliminary tests were conducted to incorporate knowledge from transcription factor binding sites into our networkbased model for prioritization, with encouraging results. Genes have widely different pertinences to the etiology and pathology of diseases. Thus, they can be ranked according to their disease-significance on a genomic scale, which is the subject of gene prioritization. Given a set of genes known to be related to a disease, it is reasonable to use them as a basis to determine the significance of other candidate genes, which will then be ranked based on the association they exhibit with respect to the given set of known genes. Experimental and computational data of various kinds have different reliability and relevance to a disease under study. This work presents a gene prioritization method based on integrated biological networks that incorporates and models the various levels of relevance and reliability of diverse sources. The method is shown to achieve significantly higher performance as compared to two well-known gene prioritization algorithms. Essentially, no bias in the performance was seen as it was applied to diseases of diverse ethnology, e.g., monogenic, polygenic and cancer. The method was highly stable and robust against significant levels of noise in the data. Biological networks are often sparse, which can impede the operation of associationbased gene prioritization algorithms such as the one presented here from a computational perspective. As a potential approach to overcome this limitation, we explore the value that transcription factor binding sites can have in elucidating suitable targets. Transcription factors are needed for the expression of most genes, especially in higher organisms and hence genes can be associated via their genetic regulatory properties. While each transcription factor recognizes specific DNA sequence patterns, such patterns are mostly unknown for many transcription factors. Even those that are known are inconsistently reported in the literature, implying a potentially high level of inaccuracy. We developed computational methods for prediction and improvement of transcription factor binding patterns. Tests performed on the improvement method by employing synthetic patterns under various conditions showed that the method is very robust and the patterns produced invariably converge to nearly identical series of patterns. Preliminary tests were conducted to incorporate knowledge from transcription factor binding sites into our networkbased model for prioritization, with encouraging results. To validate these approaches in a disease-specific context, we built a schizophreniaspecific network based on the inferred associations and performed a comprehensive prioritization of human genes with respect to the disease. These results are expected to be validated empirically, but computational validation using known targets are very positive.
ContributorsLee, Jang (Author) / Gonzalez, Graciela (Thesis advisor) / Ye, Jieping (Committee member) / Davulcu, Hasan (Committee member) / Gallitano-Mendel, Amelia (Committee member) / Arizona State University (Publisher)
Created2011
149919-Thumbnail Image.png
Description
The infrastructure is built in Unsaturated Soils. However, the geotechnical practitioners insist in designing the structures based on Saturated Soil Mechanics. The design of structures based on unsaturated soil mechanics is desirable because it reduces cost and it is by far a more sustainable approach. The research community has identified

The infrastructure is built in Unsaturated Soils. However, the geotechnical practitioners insist in designing the structures based on Saturated Soil Mechanics. The design of structures based on unsaturated soil mechanics is desirable because it reduces cost and it is by far a more sustainable approach. The research community has identified the Soil-Water Characteristic Curve as the most important soil property when dealing with unsaturated conditions. This soil property is unpopular among practitioners because the laboratory testing takes an appreciable amount of time. Several authors have attempted predicting the Soil-Water Characteristic Curve; however, most of the published predictions are based on a very limited soil database. The National Resources Conservation Service has a vast database of engineering soil properties with more than 36,000 soils, which includes water content measurements at different levels of suctions. This database was used in this study to validate two existing models that based the Soil-Water Characteristic Curve prediction on statistical analysis. It was found that although the predictions are acceptable for some ranges of suctions; they did not performed that well for others. It was found that the first model validated was accurate for fine-grained soils, while the second model was best for granular soils. For these reasons, two models to estimate the Soil-Water Characteristic Curve are proposed. The first model estimates the fitting parameters of the Fredlund and Xing (1994) function separately and then, the predicted parameters are fitted to the Fredlund and Xing function for an overall estimate of the degree of saturation. Results show an overall improvement on the predicted values when compared to existing models. The second model is based on the relationship between the Soil-Water Characteristic Curve and the Pore-Size Distribution of the soils. The process allows for the prediction of the entire Soil-Water Characteristic Curve function and proved to be a better approximation than that used in the first attempt. Both models constitute important tools in the implementation of unsaturated soil mechanics into engineering practice due to the link of the prediction with simple and well known engineering soil properties.
ContributorsTorres Hernández, Gustavo (Author) / Zapata, Claudia (Thesis advisor) / Houston, Sandra (Committee member) / Witczak, Matthew (Committee member) / Arizona State University (Publisher)
Created2011
149907-Thumbnail Image.png
Description
Most existing approaches to complex event processing over streaming data rely on the assumption that the matches to the queries are rare and that the goal of the system is to identify these few matches within the incoming deluge of data. In many applications, such as stock market analysis and

Most existing approaches to complex event processing over streaming data rely on the assumption that the matches to the queries are rare and that the goal of the system is to identify these few matches within the incoming deluge of data. In many applications, such as stock market analysis and user credit card purchase pattern monitoring, however the matches to the user queries are in fact plentiful and the system has to efficiently sift through these many matches to locate only the few most preferable matches. In this work, we propose a complex pattern ranking (CPR) framework for specifying top-k pattern queries over streaming data, present new algorithms to support top-k pattern queries in data streaming environments, and verify the effectiveness and efficiency of the proposed algorithms. The developed algorithms identify top-k matching results satisfying both patterns as well as additional criteria. To support real-time processing of the data streams, instead of computing top-k results from scratch for each time window, we maintain top-k results dynamically as new events come and old ones expire. We also develop new top-k join execution strategies that are able to adapt to the changing situations (e.g., sorted and random access costs, join rates) without having to assume a priori presence of data statistics. Experiments show significant improvements over existing approaches.
ContributorsWang, Xinxin (Author) / Candan, K. Selcuk (Thesis advisor) / Chen, Yi (Committee member) / Davulcu, Hasan (Committee member) / Arizona State University (Publisher)
Created2011
149953-Thumbnail Image.png
Description
The theme for this work is the development of fast numerical algorithms for sparse optimization as well as their applications in medical imaging and source localization using sensor array processing. Due to the recently proposed theory of Compressive Sensing (CS), the $\ell_1$ minimization problem attracts more attention for its ability

The theme for this work is the development of fast numerical algorithms for sparse optimization as well as their applications in medical imaging and source localization using sensor array processing. Due to the recently proposed theory of Compressive Sensing (CS), the $\ell_1$ minimization problem attracts more attention for its ability to exploit sparsity. Traditional interior point methods encounter difficulties in computation for solving the CS applications. In the first part of this work, a fast algorithm based on the augmented Lagrangian method for solving the large-scale TV-$\ell_1$ regularized inverse problem is proposed. Specifically, by taking advantage of the separable structure, the original problem can be approximated via the sum of a series of simple functions with closed form solutions. A preconditioner for solving the block Toeplitz with Toeplitz block (BTTB) linear system is proposed to accelerate the computation. An in-depth discussion on the rate of convergence and the optimal parameter selection criteria is given. Numerical experiments are used to test the performance and the robustness of the proposed algorithm to a wide range of parameter values. Applications of the algorithm in magnetic resonance (MR) imaging and a comparison with other existing methods are included. The second part of this work is the application of the TV-$\ell_1$ model in source localization using sensor arrays. The array output is reformulated into a sparse waveform via an over-complete basis and study the $\ell_p$-norm properties in detecting the sparsity. An algorithm is proposed for minimizing a non-convex problem. According to the results of numerical experiments, the proposed algorithm with the aid of the $\ell_p$-norm can resolve closely distributed sources with higher accuracy than other existing methods.
ContributorsShen, Wei (Author) / Mittlemann, Hans D (Thesis advisor) / Renaut, Rosemary A. (Committee member) / Jackiewicz, Zdzislaw (Committee member) / Gelb, Anne (Committee member) / Ringhofer, Christian (Committee member) / Arizona State University (Publisher)
Created2011
149831-Thumbnail Image.png
Description
The focus of this study is statistical characterization of the significant duration of strong ground motion time histories. The significant duration is defined as the time needed to build up between five and seventy five (SD575) and ninety five percent (SD595) of the energy of a strong motion record.

The focus of this study is statistical characterization of the significant duration of strong ground motion time histories. The significant duration is defined as the time needed to build up between five and seventy five (SD575) and ninety five percent (SD595) of the energy of a strong motion record. Energy is measured as the integral of the square of the acceleration time history and can be used to capture the potential destructiveness of an earthquake. Correlations of the geometric means of the two significant duration measures (SD575 and SD595) with source, path, and near surface site parameters have been investigated using the geometric mean of 2,690 pairs of recorded horizontal strong ground motion data from 129 earthquakes in active plate margins. These time histories correspond to moment magnitudes between 4.8 and 7.9, site to source distances up to 200 km, and near surface shear wave velocity ranging from 120 to 2250 m/s. Empirical relationships have been developed based upon the simple functional forms, and observed correlations. The coefficients of the independent variables in these empirical relationships have been determined through nonlinear regression analysis using a random effects model. It is found that significant duration measures correlate well with magnitude, site to source distance, and near surface shear wave velocity. The influence of the depth to top of rupture, depth to the shear wave velocity of 1000 m/s and the style of faulting were not found to be statistically significant. Comparison of the empirical relationship developed in this study with existing empirical relationships for the significant duration shows good agreement at intermediate magnitudes (M 6.5). However, at larger and smaller magnitude, the differences between the correlations developed in this study and those from previous studies are significant.
ContributorsGhanat, Simon T (Author) / Kavazanjian, Jr., Edward (Thesis advisor) / Houston, Sandra (Committee member) / Arrowsmith, Ramon (Committee member) / Arizona State University (Publisher)
Created2011
149822-Thumbnail Image.png
Description
It is estimated that wind induced soil transports more than 500 x 106 metric tons of fugitive dust annually. Soil erosion has negative effects on human health, the productivity of farms, and the quality of surface waters. A variety of different polymer stabilizers are available on the market for fugitive

It is estimated that wind induced soil transports more than 500 x 106 metric tons of fugitive dust annually. Soil erosion has negative effects on human health, the productivity of farms, and the quality of surface waters. A variety of different polymer stabilizers are available on the market for fugitive dust control. Most of these polymer stabilizers are expensive synthetic polymer products. Their adverse effects and expense usually limits their use. Biopolymers provide a potential alternative to synthetic polymers. They can provide dust abatement by encapsulating soil particles and creating a binding network throughout the treated area. This research into the effectiveness of biopolymers for fugitive dust control involved three phases. Phase I included proof of concept tests. Phase II included carrying out the tests in a wind tunnel. Phase III consisted of conducting the experiments in the field. Proof of concept tests showed that biopolymers have the potential to reduce soil erosion and fugitive dust transport. Wind tunnel tests on two candidate biopolymers, xanthan and chitosan, showed that there is a proportional relationship between biopolymer application rates and threshold wind velocities. The wind tunnel tests also showed that xanthan gum is more successful in the field than chitosan. The field tests showed that xanthan gum was effective at controlling soil erosion. However, the chitosan field data was inconsistent with the xanthan data and field data on bare soil.
ContributorsAlsanad, Abdullah (Author) / Kavazanjian, Edward (Thesis advisor) / Edwards, David (Committee member) / Zapata, Claudia (Committee member) / Arizona State University (Publisher)
Created2011