Published in

American Psychological Association, Journal of Experimental Psychology: Learning, Memory, and Cognition, 3(40), p. 703-723, 2014

DOI: 10.1037/a0035549

Links

Tools

Export citation

Search in Google Scholar

Evidence Evaluation: Measure Z Corresponds to Human Utility Judgments Better Than Measure L and Optimal-Experimental-Design Models

This paper is available in a repository.
This paper is available in a repository.

Full text: Download

Green circle
Preprint: archiving allowed
Green circle
Postprint: archiving allowed
Red circle
Published version: archiving forbidden
Data provided by SHERPA/RoMEO

Abstract

Evidence evaluation is a crucial process in many human activities, spanning from medical diagnosis to impression formation. The present experiments investigated which, if any, normative model best conforms to people’s intuition about the value of the obtained evidence. Psychologists, epistemologists, and philosophers of science have proposed several models to account for people’s intuition about the utility of the obtained evidence with respect either to a focal hypothesis or to a constellation of hypotheses. We pitted against each other the so called optimal-experimental-design models (i.e., Bayesian diagnosticity, log10 diagnosticity, information gain, Kullback-Leibler distance, probability gain, and impact) and measures L and Z to compare their ability to describe humans’ intuition about the value of the obtained evidence. Participants received words-and-numbers scenarios concerning two hypotheses and binary features. They were asked to evaluate the utility of “yes” and “no” answers to questions about some features possessed in different proportions (i.e., the likelihoods) by two types of extraterrestrial creatures (corresponding to two mutually exclusive and exhaustive hypotheses). Participants evaluated either how an answer was helpful or how an answer decreased/increased their beliefs with respect either to a single hypothesis or to both hypotheses. We fitted mixed-effects models and we used the Akaike information criterion (AIC) and the Bayesian information criterion (BIC) values to compare the competing models of the value of the obtained evidence. Overall, the experiments showed that measure Z was the best-fitting model of participants’ judgments of the value of obtained answers. We discussed the implications for the human hypothesis-evaluation process.