Abstract
When and how do we decide whether the target is present in the scene or not? There are surprisingly few quantitative models that address this question. We propose a generative model of visual search. First, the observer obtains stimulus information (corrupted by additive white noise of variance σ) at each location and time instant. Second, information across scene locations is combined using optimal Bayesian inference to compute the instantaneous evidence for target presence in the scene (log likelihood ratio of target presence vs. absence at time t). Third, the observer accumulates evidence over time and decides ‘yes’ if the accumulated evidence exceeds criterion ζ+Δ, ‘no’ if the evidence falls below criterion −ζ+Δ, and otherwise continues to observe the scene. Δ is a criterion-shift which depends on the target frequency, and the reward for different responses. This model has only 2 free parameters (noise σ, criterion ζ), compared to many parameters in drift diffusion models. The model can explain several search phenomena including how accuracy and RT are affected by set-size, target-distractor discriminability, distractor heterogeneity, target frequency and reward. It can capture the shape of the RT distribution in target present and absent trials for different tasks like feature, conjunction and spatial configuration search. It explains that rare targets are missed (Wolfe et. al, 2005) because decreasing target frequency (e.g., from 50% to 2%) shifts the starting point of the decision process closer to ‘no’ than ‘yes’ criterion, leading to high miss error rates, and fast abandoning of search. Our model predicts that increasing penalties on miss errors will decrease these errors and increase RTs. We validated these predictions through psychophysics experiments with 4 human subjects. To summarize, we have proposed a generative model of visual search that with only 2 free parameters, can explain a wide range of search phenomena.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.