ABSTRACT

Statistical inference can be described as the process of drawing conclusions about a population or process based on sample data. This chapter outlines the logic of "classical" or "frequentist" methods for such inference. Three commonly used concepts for assessing statistical error are confidence intervals, p-values, and null hypothesis tests. We explain the reasoning behind these devices without focusing unduly on the computational steps. We also outline the logic underlying resampling (simulation) methods. We identify common misinterpretations of computed quantities (such as transposing the conditional), and we discuss some of the comparative advantages and disadvantages of using confidence intervals, p-values, classical hypothesis tests, and likelihood ratios for various purposes in forensic science. Along with idealized, simple examples of probabilistic processes, we use two principal examples from forensic science to illustrate the frequentist reasoning. The first involves an experiment to ascertain the validity and false positive probability of identifications made by latent fingerprint examiners. The second involves measurements of the refractive index of glass fragments.