Matching Items (280)
149991-Thumbnail Image.png
Description
With the introduction of compressed sensing and sparse representation,many image processing and computer vision problems have been looked at in a new way. Recent trends indicate that many challenging computer vision and image processing problems are being solved using compressive sensing and sparse representation algorithms. This thesis assays some applications

With the introduction of compressed sensing and sparse representation,many image processing and computer vision problems have been looked at in a new way. Recent trends indicate that many challenging computer vision and image processing problems are being solved using compressive sensing and sparse representation algorithms. This thesis assays some applications of compressive sensing and sparse representation with regards to image enhancement, restoration and classication. The first application deals with image Super-Resolution through compressive sensing based sparse representation. A novel framework is developed for understanding and analyzing some of the implications of compressive sensing in reconstruction and recovery of an image through raw-sampled and trained dictionaries. Properties of the projection operator and the dictionary are examined and the corresponding results presented. In the second application a novel technique for representing image classes uniquely in a high-dimensional space for image classification is presented. In this method, design and implementation strategy of the image classification system through unique affine sparse codes is presented, which leads to state of the art results. This further leads to analysis of some of the properties attributed to these unique sparse codes. In addition to obtaining these codes, a strong classier is designed and implemented to boost the results obtained. Evaluation with publicly available datasets shows that the proposed method outperforms other state of the art results in image classication. The final part of the thesis deals with image denoising with a novel approach towards obtaining high quality denoised image patches using only a single image. A new technique is proposed to obtain highly correlated image patches through sparse representation, which are then subjected to matrix completion to obtain high quality image patches. Experiments suggest that there may exist a structure within a noisy image which can be exploited for denoising through a low-rank constraint.
ContributorsKulkarni, Naveen (Author) / Li, Baoxin (Thesis advisor) / Ye, Jieping (Committee member) / Sen, Arunabha (Committee member) / Arizona State University (Publisher)
Created2011
149681-Thumbnail Image.png
Description
The trend towards using recycled materials on new construction projects is growing as the cost for construction materials are ever increasing and the awareness of the responsibility we have to be good stewards of our environment is heightened. While recycled asphalt is sometimes used in pavements, its use as structural

The trend towards using recycled materials on new construction projects is growing as the cost for construction materials are ever increasing and the awareness of the responsibility we have to be good stewards of our environment is heightened. While recycled asphalt is sometimes used in pavements, its use as structural fill has been hindered by concern that it is susceptible to large long-term deformations (creep), preventing its use for a great many geotechnical applications. While asphalt/soil blends are often proposed as an alternative to 100% recycled asphalt fill, little data is available characterizing the geotechnical properties of recycled asphalt soil blends. In this dissertation, the geotechnical properties for five different recycled asphalt soil blends are characterized. Data includes the particle size distribution, plasticity index, creep, and shear strength for each blend. Blends with 0%, 25%, 50%, 75% and 100% recycled asphalt were tested. As the recycled asphalt material used for testing had particles sizes up to 1.5 inches, a large 18 inch diameter direct shear apparatus was used to determine the shear strength and creep characteristics of the material. The results of the testing program confirm that the creep potential of recycled asphalt is a geotechnical concern when the material is subjected to loads greater than 1500 pounds per square foot (psf). In addition, the test results demonstrate that the amount of soil blended with the recycled asphalt can greatly influence the creep and shear strength behavior of the composite material. Furthermore, there appears to be an optimal blend ratio where the composite material had better properties than either the recycled asphalt or virgin soil alone with respect to shear strength.
ContributorsSchaper, Jeffery M (Author) / Kavazanjian, Edward (Thesis advisor) / Houston, Sandra L. (Committee member) / Zapata, Claudia E (Committee member) / Arizona State University (Publisher)
Created2011
149754-Thumbnail Image.png
Description
A good production schedule in a semiconductor back-end facility is critical for the on time delivery of customer orders. Compared to the front-end process that is dominated by re-entrant product flows, the back-end process is linear and therefore more suitable for scheduling. However, the production scheduling of the back-end process

A good production schedule in a semiconductor back-end facility is critical for the on time delivery of customer orders. Compared to the front-end process that is dominated by re-entrant product flows, the back-end process is linear and therefore more suitable for scheduling. However, the production scheduling of the back-end process is still very difficult due to the wide product mix, large number of parallel machines, product family related setups, machine-product qualification, and weekly demand consisting of thousands of lots. In this research, a novel mixed-integer-linear-programming (MILP) model is proposed for the batch production scheduling of a semiconductor back-end facility. In the MILP formulation, the manufacturing process is modeled as a flexible flow line with bottleneck stages, unrelated parallel machines, product family related sequence-independent setups, and product-machine qualification considerations. However, this MILP formulation is difficult to solve for real size problem instances. In a semiconductor back-end facility, production scheduling usually needs to be done every day while considering updated demand forecast for a medium term planning horizon. Due to the limitation on the solvable size of the MILP model, a deterministic scheduling system (DSS), consisting of an optimizer and a scheduler, is proposed to provide sub-optimal solutions in a short time for real size problem instances. The optimizer generates a tentative production plan. Then the scheduler sequences each lot on each individual machine according to the tentative production plan and scheduling rules. Customized factory rules and additional resource constraints are included in the DSS, such as preventive maintenance schedule, setup crew availability, and carrier limitations. Small problem instances are randomly generated to compare the performances of the MILP model and the deterministic scheduling system. Then experimental design is applied to understand the behavior of the DSS and identify the best configuration of the DSS under different demand scenarios. Product-machine qualification decisions have long-term and significant impact on production scheduling. A robust product-machine qualification matrix is critical for meeting demand when demand quantity or mix varies. In the second part of this research, a stochastic mixed integer programming model is proposed to balance the tradeoff between current machine qualification costs and future backorder costs with uncertain demand. The L-shaped method and acceleration techniques are proposed to solve the stochastic model. Computational results are provided to compare the performance of different solution methods.
ContributorsFu, Mengying (Author) / Askin, Ronald G. (Thesis advisor) / Zhang, Muhong (Thesis advisor) / Fowler, John W (Committee member) / Pan, Rong (Committee member) / Sen, Arunabha (Committee member) / Arizona State University (Publisher)
Created2011
149728-Thumbnail Image.png
Description
In geotechnical engineering, measuring the unsaturated hydraulic conductivity of fine grained soils can be time consuming and tedious. The various applications that require knowledge of the unsaturated hydraulic conductivity function are great, and in geotechnical engineering, they range from modeling seepage through landfill covers to determining infiltration of water

In geotechnical engineering, measuring the unsaturated hydraulic conductivity of fine grained soils can be time consuming and tedious. The various applications that require knowledge of the unsaturated hydraulic conductivity function are great, and in geotechnical engineering, they range from modeling seepage through landfill covers to determining infiltration of water under a building slab. The unsaturated hydraulic conductivity function can be measured using various direct and indirect techniques. The instantaneous profile method has been found to be the most promising unsteady state method for measuring the unsaturated hydraulic conductivity function for fine grained soils over a wide range of suction values. The instantaneous profile method can be modified by using different techniques to measure suction and water content and also through the way water is introduced or removed from the soil profile. In this study, the instantaneous profile method was modified by creating duplicate soil samples compacted into cylindrical tubes at two different water contents. The techniques used in the duplicate method to measure the water content and matric suction included volumetric moisture probes, manual water content measurements, and filter paper tests. The experimental testing conducted in this study provided insight into determining the unsaturated hydraulic conductivity using the instantaneous profile method for a sandy clay soil and recommendations are provided for further evaluation. Overall, this study has demonstrated that the presence of cracks has no significant impact on the hydraulic behavior of soil in high suction ranges. The results of this study do not examine the behavior of cracked soil unsaturated hydraulic conductivity at low suction and at moisture contents near saturation.
ContributorsJacquemin, Sean Christopher (Author) / Zapata, Claudia (Thesis advisor) / Houston, Sandra (Committee member) / Kavazanjian, Edward (Committee member) / Arizona State University (Publisher)
Created2011
150348-Thumbnail Image.png
Description
Demands in file size and transfer rates for consumer-orientated products have escalated in recent times. This is primarily due to the emergence of high definition video content. Now factor in the consumer desire for convenience, and we find that wireless service is the most desired approach for inter-connectivity. Consumers expect

Demands in file size and transfer rates for consumer-orientated products have escalated in recent times. This is primarily due to the emergence of high definition video content. Now factor in the consumer desire for convenience, and we find that wireless service is the most desired approach for inter-connectivity. Consumers expect wireless service to emulate wired service with little to virtually no difference in quality of service (QoS). The background section of this document examines the QoS requirements for wireless connectivity of high definition video applications. I then proceed to look at proposed solutions at the physical (PHY) and the media access control (MAC) layers as well as cross-layer schemes. These schemes are subsequently are evaluated in terms of usefulness in a multi-gigabit, 60 GHz wireless multimedia system targeting the average consumer. It is determined that a substantial gap in published literature exists pertinent to this application. Specifically, little or no work has been found that shows how an adaptive PHYMAC cross-layer solution that provides real-time compensation for varying channel conditions might be actually implemented. Further, no work has been found that shows results of such a model. This research proposes, develops and implements in Matlab code an alternate cross-layer solution that will provide acceptable QoS service for multimedia applications. Simulations using actual high definition video sequences are used to test the proposed solution. Results based on the average PSNR metric show that a quasi-adaptive algorithm provides greater than 7 dB of improvement over a non-adaptive approach while a fully-adaptive alogrithm provides over18 dB of improvement. The fully adaptive implementation has been conclusively shown to be superior to non-adaptive techniques and sufficiently superior to even quasi-adaptive algorithms.
ContributorsBosco, Bruce (Author) / Reisslein, Martin (Thesis advisor) / Tepedelenlioğlu, Cihan (Committee member) / Sen, Arunabha (Committee member) / Arizona State University (Publisher)
Created2011
Description
Filtration for microfluidic sample-collection devices is desirable for sample selection, concentration, preprocessing, and downstream manipulation, but microfabricating the required sub-micrometer filtration structure is an elaborate process. This thesis presents a simple method to fabricate polydimethylsiloxane (PDMS) devices with an integrated membrane filter that will sample, lyse, and extract the DNA

Filtration for microfluidic sample-collection devices is desirable for sample selection, concentration, preprocessing, and downstream manipulation, but microfabricating the required sub-micrometer filtration structure is an elaborate process. This thesis presents a simple method to fabricate polydimethylsiloxane (PDMS) devices with an integrated membrane filter that will sample, lyse, and extract the DNA from microorganisms in aqueous environments. An off-the-shelf membrane filter disc was embedded in a PDMS layer and sequentially bound with other PDMS channel layers. No leakage was observed during filtration. This device was validated by concentrating a large amount of cyanobacterium Synechocystis in simulated sample water with consistent performance across devices. After accumulating sufficient biomass on the filter, a sequential electrochemical lysing process was performed by applying 5VDC across the filter. This device was further evaluated by delivering several samples of differing concentrations of cyanobacterium Synechocystis then quantifying the DNA using real-time PCR. Lastly, an environmental sample was run through the device and the amount of photosynthetic microorganisms present in the water was determined. The major breakthroughs in this design are low energy demand, cheap materials, simple design, straightforward fabrication, and robust performance, together enabling wide-utility of similar chip-based devices for field-deployable operations in environmental micro-biotechnology.
ContributorsLecluse, Aurelie (Author) / Meldrum, Deirdre (Thesis advisor) / Chao, Joseph (Thesis advisor) / Westerhoff, Paul (Committee member) / Arizona State University (Publisher)
Created2011
148174-Thumbnail Image.png
Description

Much of Nepal lacks access to clean drinking water, and many water sources are contaminated with arsenic at concentrations above both World Health Organization and local Nepalese guidelines. While many water treatment technologies exist, it is necessary to identify those that are easily implementable in developing areas. One simple treatment

Much of Nepal lacks access to clean drinking water, and many water sources are contaminated with arsenic at concentrations above both World Health Organization and local Nepalese guidelines. While many water treatment technologies exist, it is necessary to identify those that are easily implementable in developing areas. One simple treatment that has gained popularity is biochar—a porous, carbon-based substance produced through pyrolysis of biomass in an oxygen-free environment. Arizona State University’s Engineering Projects in Community Service (EPICS) has partnered with communities in Nepal in an attempt to increase biochar production in the area, as it has several valuable applications including water treatment. Biochar’s arsenic adsorption capability will be investigated in this project with the goal of using the biochar that Nepalese communities produce to remove water contaminants. It has been found in scientific literature that biochar is effective in removing heavy metal contaminants from water with the addition of iron through surface activation. Thus, the specific goal of this research was to compare the arsenic adsorption disparity between raw biochar and iron-impregnated biochar. It was hypothesized that after numerous bed volumes pass through a water treatment column, iron from the source water will accumulate on the surface of raw biochar, mimicking the intentionally iron-impregnated biochar and further increasing contaminant uptake. It is thus an additional goal of this project to compare biochar loaded with iron through an iron-spiked water column and biochar impregnated with iron through surface oxidation. For this investigation, the biochar was crushed and sieved to a size between 90 and 100 micrometers. Two samples were prepared: raw biochar and oxidized biochar. The oxidized biochar was impregnated with iron through surface oxidation with potassium permanganate and iron loading. Then, X-ray fluorescence was used to compare the composition of the oxidized biochar with its raw counterpart, indicating approximately 0.5% iron in the raw and 1% iron in the oxidized biochar. The biochar samples were then added to batches of arsenic-spiked water at iron to arsenic concentration ratios of 20 mg/L:1 mg/L and 50 mg/L:1 mg/L to determine adsorption efficiency. Inductively coupled plasma mass spectrometry (ICP-MS) analysis indicated an 86% removal of arsenic using a 50:1 ratio of iron to arsenic (1.25 g biochar required in 40 mL solution), and 75% removal with a 20:1 ratio (0.5 g biochar required in 40 mL solution). Additional samples were then inserted into a column process apparatus for further adsorption analysis. Again, ICP-MS analysis was performed and the results showed that while both raw and treated biochars were capable of adsorbing arsenic, they were exhausted after less than 70 bed volumes (234 mL), with raw biochar lasting 60 bed volumes (201 mL) and oxidized about 70 bed volumes (234 mL). Further research should be conducted to investigate more affordable and less laboratory-intensive processes to prepare biochar for water treatment.

ContributorsLaird, Ashlyn (Author) / Schoepf, Jared (Thesis director) / Westerhoff, Paul (Committee member) / Chemical Engineering Program (Contributor) / School of International Letters and Cultures (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05
149901-Thumbnail Image.png
Description
Query Expansion is a functionality of search engines that suggest a set of related queries for a user issued keyword query. In case of exploratory or ambiguous keyword queries, the main goal of the user would be to identify and select a specific category of query results among different categorical

Query Expansion is a functionality of search engines that suggest a set of related queries for a user issued keyword query. In case of exploratory or ambiguous keyword queries, the main goal of the user would be to identify and select a specific category of query results among different categorical options, in order to narrow down the search and reach the desired result. Typical corpus-driven keyword query expansion approaches return popular words in the results as expanded queries. These empirical methods fail to cover all semantics of categories present in the query results. More importantly these methods do not consider the semantic relationship between the keywords featured in an expanded query. Contrary to a normal keyword search setting, these factors are non-trivial in an exploratory and ambiguous query setting where the user's precise discernment of different categories present in the query results is more important for making subsequent search decisions. In this thesis, I propose a new framework for keyword query expansion: generating a set of queries that correspond to the categorization of original query results, which is referred as Categorizing query expansion. Two approaches of algorithms are proposed, one that performs clustering as pre-processing step and then generates categorizing expanded queries based on the clusters. The other category of algorithms handle the case of generating quality expanded queries in the presence of imperfect clusters.
ContributorsNatarajan, Sivaramakrishnan (Author) / Chen, Yi (Thesis advisor) / Candan, Selcuk (Committee member) / Sen, Arunabha (Committee member) / Arizona State University (Publisher)
Created2011
149953-Thumbnail Image.png
Description
The theme for this work is the development of fast numerical algorithms for sparse optimization as well as their applications in medical imaging and source localization using sensor array processing. Due to the recently proposed theory of Compressive Sensing (CS), the $\ell_1$ minimization problem attracts more attention for its ability

The theme for this work is the development of fast numerical algorithms for sparse optimization as well as their applications in medical imaging and source localization using sensor array processing. Due to the recently proposed theory of Compressive Sensing (CS), the $\ell_1$ minimization problem attracts more attention for its ability to exploit sparsity. Traditional interior point methods encounter difficulties in computation for solving the CS applications. In the first part of this work, a fast algorithm based on the augmented Lagrangian method for solving the large-scale TV-$\ell_1$ regularized inverse problem is proposed. Specifically, by taking advantage of the separable structure, the original problem can be approximated via the sum of a series of simple functions with closed form solutions. A preconditioner for solving the block Toeplitz with Toeplitz block (BTTB) linear system is proposed to accelerate the computation. An in-depth discussion on the rate of convergence and the optimal parameter selection criteria is given. Numerical experiments are used to test the performance and the robustness of the proposed algorithm to a wide range of parameter values. Applications of the algorithm in magnetic resonance (MR) imaging and a comparison with other existing methods are included. The second part of this work is the application of the TV-$\ell_1$ model in source localization using sensor arrays. The array output is reformulated into a sparse waveform via an over-complete basis and study the $\ell_p$-norm properties in detecting the sparsity. An algorithm is proposed for minimizing a non-convex problem. According to the results of numerical experiments, the proposed algorithm with the aid of the $\ell_p$-norm can resolve closely distributed sources with higher accuracy than other existing methods.
ContributorsShen, Wei (Author) / Mittlemann, Hans D (Thesis advisor) / Renaut, Rosemary A. (Committee member) / Jackiewicz, Zdzislaw (Committee member) / Gelb, Anne (Committee member) / Ringhofer, Christian (Committee member) / Arizona State University (Publisher)
Created2011
149817-Thumbnail Image.png
Description
Atmospheric particulate matter has a substantial impact on global climate due to its ability to absorb/scatter solar radiation and act as cloud condensation nuclei (CCN). Yet, little is known about marine aerosol, in particular, the carbonaceous fraction. In the present work, particulate matter was collected, using High Volume (HiVol) samplers,

Atmospheric particulate matter has a substantial impact on global climate due to its ability to absorb/scatter solar radiation and act as cloud condensation nuclei (CCN). Yet, little is known about marine aerosol, in particular, the carbonaceous fraction. In the present work, particulate matter was collected, using High Volume (HiVol) samplers, onto quartz fiber substrates during a series of research cruises on the Atlantic Ocean. Samples were collected on board the R/V Endeavor on West–East (March–April, 2006) and East–West (June–July, 2006) transects in the North Atlantic, as well as on the R/V Polarstern during a North–South (October–November, 2005) transect along the western coast of Europe and Africa. The aerosol total carbon (TC) concentrations for the West–East (Narragansett, RI, USA to Nice, France) and East–West (Heraklion, Crete, Greece to Narragansett, RI, USA) transects were generally low over the open ocean (0.36±0.14 μg C/m3) and increased as the ship approached coastal areas (2.18±1.37 μg C/m3), due to increased terrestrial/anthropogenic aerosol inputs. The TC for the North–South transect samples decreased in the southern hemisphere with the exception of samples collected near the 15th parallel where calculations indicate the air mass back trajectories originated from the continent. Seasonal variation in organic carbon (OC) was seen in the northern hemisphere open ocean samples with average values of 0.45 μg/m3 and 0.26 μg/m3 for spring and summer, respectively. These low summer time values are consistent with SeaWiFS satellite images that show decreasing chlorophyll a concentration (a proxy for phytoplankton biomass) in the summer. There is also a statistically significant (p<0.05) decline in surface water fluorescence in the summer. Moreover, examination of water–soluble organic carbon (WSOC) shows that the summer aerosol samples appear to have a higher fraction of the lower molecular weight material, indicating that the samples may be more oxidized (aged). The seasonal variation in aerosol content seen during the two 2006 cruises is evidence that a primary biological marine source is a significant contributor to the carbonaceous particulate in the marine atmosphere and is consistent with previous studies of clean marine air masses.
ContributorsHill, Hansina Rae (Author) / Herckes, Pierre (Thesis advisor) / Westerhoff, Paul (Committee member) / Hartnett, Hilairy (Committee member) / Arizona State University (Publisher)
Created2011