Matching Items (73)
Filtering by

Clear all filters

149563-Thumbnail Image.png
Description
This thesis explores the independent effects of the manipulation of rocks into alignments, prehistoric farming, and season on soil properties in two areas with a history of prehistoric agriculture in central Arizona, Pueblo la Plata within the Agua Fria National Monument (AFNM), and an archaeological site north of the Phoenix

This thesis explores the independent effects of the manipulation of rocks into alignments, prehistoric farming, and season on soil properties in two areas with a history of prehistoric agriculture in central Arizona, Pueblo la Plata within the Agua Fria National Monument (AFNM), and an archaeological site north of the Phoenix basin along Cave Creek (CC). Soil properties, annual herbaceous biomass and the physical properties of alignments and surface soils were measured and compared across the landscape, specifically on: 1) agricultural rock alignments that were near the archaeological site 2) geologically formed rock alignments that were located 0.5-1 km away from settlements; and 3) areas both near and far from settlements where rock alignments were absent. At AFNM, relatively well-built rock alignments have altered soil properties and processes while less-intact alignments at CC have left few legacies.
ContributorsTrujillo, Jolene Eve (Author) / Hall, Sharon J (Thesis advisor) / Collins, Scott L. (Committee member) / Spielmann, Katherine A. (Committee member) / Arizona State University (Publisher)
Created2011
149430-Thumbnail Image.png
Description
As an industrial society, humans have increasingly separated agricultural processes from natural ecosystems. Many areas of the Southwestern US, however, maintain traditional practices that link agricultural systems to the natural environment. One such practice, diverting river water into fields via earthen irrigation canals, allows ditch water to recharge

As an industrial society, humans have increasingly separated agricultural processes from natural ecosystems. Many areas of the Southwestern US, however, maintain traditional practices that link agricultural systems to the natural environment. One such practice, diverting river water into fields via earthen irrigation canals, allows ditch water to recharge groundwater and riparian vegetation to prosper along canal banks. As there is growing interest in managing landscapes for multiple ecosystem services, this study was undertaken to determine if irrigation canals function as an extension of the riparian corridor. I was specifically interested in determining if the processes within semi-arid streams that drive riparian plant community structure are manifested in earthen irrigation ditches. I examined herbaceous and woody vegetation along the middle Verde River, AZ, USA and three adjacent irrigation ditches across six months. I also collected sieved hydrochores--seeds dispersing through water--within ditches and the river twelve times. Results indicate that ditch vegetation was similar to streamside river vegetation in abundance (cover and basal area) due to surface water availability but more diverse than river streamside vegetation due to high heterogeneity. Compositionally, herbaceous vegetation along the ditch was most similar to the river banks, while low disturbance fostered woody vegetation along the ditches similar to high floodplain and river terrace vegetation. Hydrochore richness and abundance within the river was dependent on seasonality and stream discharge, but these relationships were dampened in the ditches. Species-specific strategies of hydrochory, however, did emerge in both systems. Strategies include pulse species, which disperse via hydrochory in strict accordance with their restricted dispersal windows, constant species, which are year round hydrochores, and combination species, which show characteristics of both. There was high overlap in the composition of hydrochores in the two systems, with obligate wetland species abundant in both. Upland species were more seasonally constant and abundant in the ditch water than the river. The consistency of river processes and similarity of vegetation suggest that earthen irrigation ditches do function as an extension of the riparian corridor. Thus, these man-made irrigation ditches should be considered by stakeholders for their multiple ecosystem services.
ContributorsBetsch, Jacqueline Michelle (Author) / Stromberg, Juliet C. (Thesis advisor) / Hall, Sharon J (Committee member) / Merritt, David M. (Committee member) / Arizona State University (Publisher)
Created2010
157623-Thumbnail Image.png
Description
Feature embeddings differ from raw features in the sense that the former obey certain properties like notion of similarity/dissimilarity in it's embedding space. word2vec is a preeminent example in this direction, where the similarity in the embedding space is measured in terms of the cosine similarity. Such language embedding models

Feature embeddings differ from raw features in the sense that the former obey certain properties like notion of similarity/dissimilarity in it's embedding space. word2vec is a preeminent example in this direction, where the similarity in the embedding space is measured in terms of the cosine similarity. Such language embedding models have seen numerous applications in both language and vision community as they capture the information in the modality (English language) efficiently. Inspired by these language models, this work focuses on learning embedding spaces for two visual computing tasks, 1. Image Hashing 2. Zero Shot Learning. The training set was used to learn embedding spaces over which similarity/dissimilarity is measured using several distance metrics like hamming / euclidean / cosine distances. While the above-mentioned language models learn generic word embeddings, in this work task specific embeddings were learnt which can be used for Image Retrieval and Classification separately.

Image Hashing is the task of mapping images to binary codes such that some notion of user-defined similarity is preserved. The first part of this work focuses on designing a new framework that uses the hash-tags associated with web images to learn the binary codes. Such codes can be used in several applications like Image Retrieval and Image Classification. Further, this framework requires no labelled data, leaving it very inexpensive. Results show that the proposed approach surpasses the state-of-art approaches by a significant margin.

Zero-shot classification is the task of classifying the test sample into a new class which was not seen during training. This is possible by establishing a relationship between the training and the testing classes using auxiliary information. In the second part of this thesis, a framework is designed that trains using the handcrafted attribute vectors and word vectors but doesn’t require the expensive attribute vectors during test time. More specifically, an intermediate space is learnt between the word vector space and the image feature space using the hand-crafted attribute vectors. Preliminary results on two zero-shot classification datasets show that this is a promising direction to explore.
ContributorsGattupalli, Jaya Vijetha (Author) / Li, Baoxin (Thesis advisor) / Yang, Yezhou (Committee member) / Venkateswara, Hemanth (Committee member) / Arizona State University (Publisher)
Created2019
161939-Thumbnail Image.png
Description
Traditional Reinforcement Learning (RL) assumes to learn policies with respect to reward available from the environment but sometimes learning in a complex domain requires wisdom which comes from a wide range of experience. In behavior based robotics, it is observed that a complex behavior can be described by a combination

Traditional Reinforcement Learning (RL) assumes to learn policies with respect to reward available from the environment but sometimes learning in a complex domain requires wisdom which comes from a wide range of experience. In behavior based robotics, it is observed that a complex behavior can be described by a combination of simpler behaviors. It is tempting to apply similar idea such that simpler behaviors can be combined in a meaningful way to tailor the complex combination. Such an approach would enable faster learning and modular design of behaviors. Complex behaviors can be combined with other behaviors to create even more advanced behaviors resulting in a rich set of possibilities. Similar to RL, combined behavior can keep evolving by interacting with the environment. The requirement of this method is to specify a reasonable set of simple behaviors. In this research, I present an algorithm that aims at combining behavior such that the resulting behavior has characteristics of each individual behavior. This approach has been inspired by behavior based robotics, such as the subsumption architecture and motor schema-based design. The combination algorithm outputs n weights to combine behaviors linearly. The weights are state dependent and change dynamically at every step in an episode. This idea is tested on discrete and continuous environments like OpenAI’s “Lunar Lander” and “Biped Walker”. Results are compared with related domains like Multi-objective RL, Hierarchical RL, Transfer learning, and basic RL. It is observed that the combination of behaviors is a novel way of learning which helps the agent achieve required characteristics. A combination is learned for a given state and so the agent is able to learn faster in an efficient manner compared to other similar approaches. Agent beautifully demonstrates characteristics of multiple behaviors which helps the agent to learn and adapt to the environment. Future directions are also suggested as possible extensions to this research.
ContributorsVora, Kevin Jatin (Author) / Zhang, Yu (Thesis advisor) / Yang, Yezhou (Committee member) / Praharaj, Sarbeswar (Committee member) / Arizona State University (Publisher)
Created2021
161967-Thumbnail Image.png
Description
Machine learning models can pick up biases and spurious correlations from training data and projects and amplify these biases during inference, thus posing significant challenges in real-world settings. One approach to mitigating this is a class of methods that can identify filter out bias-inducing samples from the training datasets to

Machine learning models can pick up biases and spurious correlations from training data and projects and amplify these biases during inference, thus posing significant challenges in real-world settings. One approach to mitigating this is a class of methods that can identify filter out bias-inducing samples from the training datasets to force models to avoid being exposed to biases. However, the filtering leads to a considerable wastage of resources as most of the dataset created is discarded as biased. This work deals with avoiding the wastage of resources by identifying and quantifying the biases. I further elaborate on the implications of dataset filtering on robustness (to adversarial attacks) and generalization (to out-of-distribution samples). The findings suggest that while dataset filtering does help to improve OOD(Out-Of-Distribution) generalization, it has a significant negative impact on robustness to adversarial attacks. It also shows that transforming bias-inducing samples into adversarial samples (instead of eliminating them from the dataset) can significantly boost robustness without sacrificing generalization.
ContributorsSachdeva, Bhavdeep Singh (Author) / Baral, Chitta (Thesis advisor) / Liu, Huan (Committee member) / Yang, Yezhou (Committee member) / Arizona State University (Publisher)
Created2021
168842-Thumbnail Image.png
Description
There has been an explosion in the amount of data on the internet because of modern technology – especially image data – as a consequence of an exponential growth in the number of cameras existing in the world right now; from more extensive surveillance camera systems to billions of people

There has been an explosion in the amount of data on the internet because of modern technology – especially image data – as a consequence of an exponential growth in the number of cameras existing in the world right now; from more extensive surveillance camera systems to billions of people walking around with smartphones in their pockets that come with built-in cameras. With this sudden increase in the accessibility of cameras, most of the data that is getting captured through these devices is ending up on the internet. Researchers soon took leverage of this data by creating large-scale datasets. However, generating a dataset – let alone a large-scale one – requires a lot of man-hours. This work presents an algorithm that makes use of optical flow and feature matching, along with utilizing localization outputs from a Mask R-CNN, to generate large-scale vehicle datasets without much human supervision. Additionally, this work proposes a novel multi-view vehicle dataset (MVVdb) of 500 vehicles which is also generated using the aforementioned algorithm.There are various research problems in computer vision that can leverage a multi-view dataset, e.g., 3D pose estimation, and 3D object detection. On the other hand, a multi-view vehicle dataset can be used for a 2D image to 3D shape prediction, generation of 3D vehicle models, and even a more robust vehicle make and model recognition. In this work, a ResNet is trained on the multi-view vehicle dataset to perform vehicle re-identification, which is fundamentally similar to a vehicle make and recognition problem – also showcasing the usability of the MVVdb dataset.
ContributorsGuha, Anubhab (Author) / Yang, Yezhou (Thesis advisor) / Lu, Duo (Committee member) / Banerjee, Ayan (Committee member) / Arizona State University (Publisher)
Created2022
168367-Thumbnail Image.png
Description
In recent years, there has been significant progress in deep learning and computer vision, with many models proposed that have achieved state-of-art results on various image recognition tasks. However, to explore the full potential of the advances in this field, there is an urgent need to push the processing of

In recent years, there has been significant progress in deep learning and computer vision, with many models proposed that have achieved state-of-art results on various image recognition tasks. However, to explore the full potential of the advances in this field, there is an urgent need to push the processing of deep networks from the cloud to edge devices. Unfortunately, many deep learning models cannot be efficiently implemented on edge devices as these devices are severely resource-constrained. In this thesis, I present QU-Net, a lightweight binary segmentation model based on the U-Net architecture. Traditionally, neural networks consider the entire image to be significant. However, in real-world scenarios, many regions in an image do not contain any objects of significance. These regions can be removed from the original input allowing a network to focus on the relevant regions and thus reduce computational costs. QU-Net proposes the salient regions (binary mask) that the deeper models can use as the input. Experiments show that QU-Net helped achieve a computational reduction of 25% on the Microsoft Common Objects in Context (MS COCO) dataset and 57% on the Cityscapes dataset. Moreover, QU-Net is a generalizable model that outperforms other similar works, such as Dynamic Convolutions.
ContributorsSanthosh Kumar Varma, Rahul (Author) / Yang, Yezhou (Thesis advisor) / Fan, Deliang (Committee member) / Yang, Yingzhen (Committee member) / Arizona State University (Publisher)
Created2021
168821-Thumbnail Image.png
Description
It is not merely an aggregation of static entities that a video clip carries, but alsoa variety of interactions and relations among these entities. Challenges still remain for a video captioning system to generate natural language descriptions focusing on the prominent interest and aligning with the latent aspects beyond observations. This work presents

It is not merely an aggregation of static entities that a video clip carries, but alsoa variety of interactions and relations among these entities. Challenges still remain for a video captioning system to generate natural language descriptions focusing on the prominent interest and aligning with the latent aspects beyond observations. This work presents a Commonsense knowledge Anchored Video cAptioNing (dubbed as CAVAN) approach. CAVAN exploits inferential commonsense knowledge to assist the training of video captioning model with a novel paradigm for sentence-level semantic alignment. Specifically, commonsense knowledge is queried to complement per training caption by querying a generic knowledge atlas ATOMIC, and form the commonsense- caption entailment corpus. A BERT based language entailment model trained from this corpus then serves as a commonsense discriminator for the training of video captioning model, and penalizes the model from generating semantically misaligned captions. With extensive empirical evaluations on MSR-VTT, V2C and VATEX datasets, CAVAN consistently improves the quality of generations and shows higher keyword hit rate. Experimental results with ablations validate the effectiveness of CAVAN and reveals that the use of commonsense knowledge contributes to the video caption generation.
ContributorsShao, Huiliang (Author) / Yang, Yezhou (Thesis advisor) / Jayasuriya, Suren (Committee member) / Xiao, Chaowei (Committee member) / Arizona State University (Publisher)
Created2022
171536-Thumbnail Image.png
Description
More people live in cities or metropolitan areas than ever before, which encompass many types of urbanization. These areas are culturally diverse and densely populated heterogeneous landscapes that are shaped by socio-ecological patterns. Cities support human and wildlife populations that are influenced indirectly and directly by human decisions. This process

More people live in cities or metropolitan areas than ever before, which encompass many types of urbanization. These areas are culturally diverse and densely populated heterogeneous landscapes that are shaped by socio-ecological patterns. Cities support human and wildlife populations that are influenced indirectly and directly by human decisions. This process can result in unequal access to environmental services and accessible green spaces. Additionally, biodiversity distribution is influenced by human decisions. Although neighborhood income can drive biodiversity in metropolitan areas (i.e., the ‘luxury effect’), other socio-cultural factors may also influence the presence and abundance of wildlife beyond simple measures of wealth. To understand how additional social factors shape distributions of wildlife, I ask, are patterns of wildlife distribution associated with neighborhood ethnicity, in addition to income and ecological landscape characteristics within metropolitan areas? Utilizing data from 38 wildlife cameras deployed in neighborhood public parks and non-built spaces in metro Phoenix, AZ (USA), I estimated occupancy and activity patterns of coyotes (Canis latrans), desert cottontail rabbits (Sylvilagus audubonii), and domestic cats (Felis catus) across gradients of median household income and neighborhood ethnicity, estimated by the proportion of Latinx residents. Neighborhood ethnicity appeared in the top models for all species, and neighborhood % of Latinx residents was inversely associated with presence of native Sonoran Desert animals (coyotes and cottontail rabbits). Furthermore, daily activity patterns of coyotes differed in neighborhoods with higher vs. lower proportion of Latinx residents. My results suggest that socio-cultural variables beyond income are associated with wildlife distributions, and that factors associated with neighborhood ethnicity may be an informative correlate of city-wide ecological patterns. In this research, I unraveled predictive social variables and differentiated wildlife distribution across neighborhood gradients of income and ethnic composition, bringing attention to the potentially unequal distribution of mammals in cities.
ContributorsCocroft, Alexandreana (Author) / Hall, Sharon J (Thesis advisor) / Lerman, Susannah B (Committee member) / Lewis, Jesse (Committee member) / Arizona State University (Publisher)
Created2022
171495-Thumbnail Image.png
Description
Multimodal reasoning is one of the most interesting research fields because of the ability to interact with systems and the explainability of the models' behavior. Traditional multimodal research problems do not focus on complex commonsense reasoning (such as physical interactions). Although real-world objects have physical properties associated with them,

Multimodal reasoning is one of the most interesting research fields because of the ability to interact with systems and the explainability of the models' behavior. Traditional multimodal research problems do not focus on complex commonsense reasoning (such as physical interactions). Although real-world objects have physical properties associated with them, many of these properties (such as mass and coefficient of friction) are not captured directly by the imaging pipeline. Videos often capture objects, their motion, and the interactions between different objects. However, these properties can be estimated by utilizing cues from relative object motion and the dynamics introduced by collisions. This thesis introduces a new video question-answering task for reasoning about the implicit physical properties of objects in a scene, from videos. For this task, I introduce a dataset -- CRIPP-VQA (Counterfactual Reasoning about Implicit Physical Properties - Video Question Answering), which contains videos of objects in motion, annotated with hypothetical/counterfactual questions about the effect of actions (such as removing, adding, or replacing objects), questions about planning (choosing actions to perform to reach a particular goal), as well as descriptive questions about the visible properties of objects. Further, I benchmark the performance of existing video question-answering models on two test settings of CRIPP-VQA: i.i.d. and an out-of-distribution setting which contains objects with values of mass, coefficient of friction, and initial velocities that are not seen in the training distribution. Experiments reveal a surprising and significant performance gap in terms of answering questions about implicit properties (the focus of this thesis) and explicit properties (the focus of prior work) of objects.
ContributorsPatel, Maitreya Jitendra (Author) / Yang, Yezhou (Thesis advisor) / Baral, Chitta (Committee member) / Lee, Kookjin (Committee member) / Arizona State University (Publisher)
Created2022