Matching Items (3)
Filtering by

Clear all filters

152006-Thumbnail Image.png
Description
When people look for things in their environment they use a target template - a mental representation of the object they are attempting to locate - to guide their attention around a scene and to assess incoming visual input to determine if they have found that for which they are

When people look for things in their environment they use a target template - a mental representation of the object they are attempting to locate - to guide their attention around a scene and to assess incoming visual input to determine if they have found that for which they are searching. However, unlike laboratory experiments, searchers in the real-world rarely have perfect knowledge regarding the appearance of their target. In five experiments (with nearly 1,000 participants), we examined how the precision of the observer's template affects their ability to conduct visual search. Specifically, we simulated template imprecision in two ways: First, by contaminating our searchers' templates with inaccurate features, and second, by introducing extraneous features to the template that were unhelpful. In those experiments we recorded the eye movements of our searchers in order to make inferences regarding the extent to which attentional guidance and decision-making are hindered by template imprecision. We also examined a third way in which templates may become imprecise; namely, that they may deteriorate over time. Overall, our findings support a dual-function theory of the target template, and highlight the importance of examining template precision in future research.
ContributorsHout, Michael C (Author) / Goldinger, Stephen D (Thesis advisor) / Azuma, Tamiko (Committee member) / Homa, Donald (Committee member) / Reichle, Erik (Committee member) / Arizona State University (Publisher)
Created2013
152859-Thumbnail Image.png
Description
Previous research has shown that people can implicitly learn repeated visual contexts and use this information when locating relevant items. For example, when people are presented with repeated spatial configurations of distractor items or distractor identities in visual search, they become faster to find target stimuli in these repeated contexts

Previous research has shown that people can implicitly learn repeated visual contexts and use this information when locating relevant items. For example, when people are presented with repeated spatial configurations of distractor items or distractor identities in visual search, they become faster to find target stimuli in these repeated contexts over time (Chun and Jiang, 1998; 1999). Given that people learn these repeated distractor configurations and identities, might they also implicitly encode semantic information about distractors, if this information is predictive of the target location? We investigated this question with a series of visual search experiments using real-world stimuli within a contextual cueing paradigm (Chun and Jiang, 1998). Specifically, we tested whether participants could learn, through experience, that the target images they are searching for are always located near specific categories of distractors, such as food items or animals. We also varied the spatial consistency of target locations, in order to rule out implicit learning of repeated target locations. Results suggest that participants implicitly learned the target-predictive categories of distractors and used this information during search, although these results failed to reach significance. This lack of significance may have been due the relative simplicity of the search task, however, and several new experiments are proposed to further investigate whether repeated category information can benefit search.
ContributorsWalenchok, Stephen C (Author) / Goldinger, Stephen D (Thesis advisor) / Azuma, Tamiko (Committee member) / Homa, Donald (Committee member) / Hout, Michael C (Committee member) / Arizona State University (Publisher)
Created2014
154219-Thumbnail Image.png
Description
ABSTRACT

The present studies investigated the separate effects of two types of visual feedback delay – increased latency and decreased updating rate – on performance – both actual (e.g. response time) and subjective (i.e. rating of perceived input device performance) – in 2-dimensional pointing tasks using a mouse as an input

ABSTRACT

The present studies investigated the separate effects of two types of visual feedback delay – increased latency and decreased updating rate – on performance – both actual (e.g. response time) and subjective (i.e. rating of perceived input device performance) – in 2-dimensional pointing tasks using a mouse as an input device. The first sub-study examined the effects of increased latency on performance using two separate experiments. In the first experiment the effects of constant latency on performance were tested, wherein participants completed blocks of trials with a constant level of latency. Additionally, after each block, participants rated their subjective experience of the input device performance at each level of latency. The second experiment examined the effects of variable latency on performance, where latency was randomized within blocks of trials.

The second sub-study investigated the effects of decreased updating rates on performance in the same manner as the first study, wherein experiment one tested the effect of constant updating rate on performance as well as subjective rating, and experiment two tested the effect of variable updating rate on performance. The findings suggest that latency is negative correlated with actual performance as well as subjective ratings of performance, and updating rate is positively correlated with actual performance as well as subjective ratings of performance.
ContributorsBrady, Kyle J (Author) / Wu, Bing (Thesis advisor) / Hout, Michael C (Committee member) / Branaghan, Russell (Committee member) / Arizona State University (Publisher)
Created2015