Introduction to Statistics for Forensic Scientists
eBook - ePub

Introduction to Statistics for Forensic Scientists

David Lucy

Share book
  1. English
  2. ePUB (mobile friendly)
  3. Available on iOS & Android
eBook - ePub

Introduction to Statistics for Forensic Scientists

David Lucy

Book details
Book preview
Table of contents
Citations

About This Book

Introduction to Statistics for Forensic Scientists is an essential introduction to the subject, gently guiding the reader through the key statistical techniques used to evaluate various types of forensic evidence. Assuming only a modest mathematical background, the book uses real-life examples from the forensic science literature and forensic case-work to illustrate relevant statistical concepts and methods.

Opening with a brief overview of the history and use of statistics within forensic science, the text then goes on to introduce statistical techniques commonly used to examine data obtained during laboratory experiments. There is a strong emphasis on the evaluation of scientific observation as evidence and modern Bayesian approaches to interpreting forensic data for the courts. The analysis of key forms of evidence are discussed throughout with a particular focus on DNA, fibres and glass.

An invaluable introduction to the statistical interpretation of forensic evidence; this book will be invaluable for all undergraduates taking courses in forensic science.

  • Introduction to the key statistical techniques used in the evaluation of forensic evidence
  • Includes end of chapter exercises to enhance student understanding
  • Numerous examples taken from forensic science to put the subject into context

Frequently asked questions

How do I cancel my subscription?
Simply head over to the account section in settings and click on “Cancel Subscription” - it’s as simple as that. After you cancel, your membership will stay active for the remainder of the time you’ve paid for. Learn more here.
Can/how do I download books?
At the moment all of our mobile-responsive ePub books are available to download via the app. Most of our PDFs are also available to download and we're working on making the final remaining ones downloadable now. Learn more here.
What is the difference between the pricing plans?
Both plans give you full access to the library and all of Perlego’s features. The only differences are the price and subscription period: With the annual plan you’ll save around 30% compared to 12 months on the monthly plan.
What is Perlego?
We are an online textbook subscription service, where you can get access to an entire online library for less than the price of a single book per month. With over 1 million books across 1000+ topics, we’ve got you covered! Learn more here.
Do you support text-to-speech?
Look out for the read-aloud symbol on your next book to see if you can listen to it. The read-aloud tool reads text aloud for you, highlighting the text as it is being read. You can pause it, speed it up and slow it down. Learn more here.
Is Introduction to Statistics for Forensic Scientists an online PDF/ePUB?
Yes, you can access Introduction to Statistics for Forensic Scientists by David Lucy in PDF and/or ePUB format, as well as other popular books in Medizin & Biostatistik. We have over one million books available in our catalogue for you to explore.

Information

Publisher
Wiley
Year
2013
ISBN
9781118700105
Edition
1
Topic
Medizin
Subtopic
Biostatistik

1

A short history of statistics in the law

The science of statistics refers to two distinct, but linked, areas of knowledge. The first is the enumeration of types of event and counts of entities for economic, social and scientific purposes, the second is the examination of uncertainty. It is in this second guise that statistics can be regarded as the science of uncertainty. It is therefore natural that statistics should be applied to evidence used for legal purposes as uncertainty is a feature of any legal process where decisions are made upon the basis of evidence. Typically, if a case is brought to a court it is the role of the court to discern, using evidence, what has happened, then decide what, if anything, has to be done in respect of the alleged events. Courts in the common law tradition are not in themselves bodies which can directly launch investigations into events, but are institutions into which evidence is brought for decisions to be made. Unless all the evidence points unambiguously towards an inevitable conclusion, different pieces of evidence will carry different implications with varying degrees of force. Modern statistical methods are available which are designed to measure this ‘weight’ of evidence.

1.1 History

Informal notions of probability have been a feature of decision making which date to at least as far in the past as the earliest writing. Many applications were, as related by Franklin (2001), to the process of law. Ancient Egypt seems to have two strands, one of which relates to the number of reliable witnesses willing to testify for or against a case, evidence which remains important today. The other is the use of oracles and is no longer in use. Even in the ancient world there seems to have been scepticism about the information divulged by oracles, sometimes two or three being consulted and the majority opinion followed. Subsequently the Jewish tradition made the assessment of uncertainty central to many religious and legal practices. Jewish law is notable in that it does not admit confession, a wholly worthy feature which makes torture useless. It also required a very high standard of proof which differed according to the seriousness of any alleged offence. Roman law had the concept of onus of proof, but the wealthier sections of Roman society were considered more competent to testify then others. The Roman judiciary were allowed some latitude to judge in accordance with the evidence. In contrast to Jewish law, torture was widespread in Roman practice. In fact in some circles the evidence from a tortured witness was considered of a higher quality than had the same witness volunteered the evidence, particularly if they happened to be a member of the slave classes.
European Medieval law looked to the Roman codes, but started to take a more abstract view of law based on general principles. This included developments in the theory of evidence such as half, quarter and finer grades of proof, and multiple supporting strands forming what we today would call a case. There seem to have been variable attitudes to the use of torture. Ordeal was used in the earlier period to support the civil law in cases which were otherwise intractable. An important tool for evidence evaluation with its beginnings in the Western European Medieval was the development of a form of jury which has continued uninterrupted until the present day. It is obvious that the ancient thinkers had some idea that the evidence with which they were dealing was uncertain, and devised many ingenious methods of making some sort of best decision in the face of the uncertainties, usually revolving around some weighting scheme given to the various individual pieces of evidence, and some process of summation. Nevertheless, it is apparent that uncertainty was not thought about in the same way in which we would think about it today.
Informal enumeration types of analyses were applied as early as in the middle of the 17th century to observational data with John Gaunt’s analysis of the London Mortality bills (Gaunt, 1662, cited in Stigler, 1986), and it is at this point in time that French mathematicians such as De MĂ©rĂ©, Roberval, Pascal and Fermet started to work on a more recognizably modern notion of probability in their attempts to solve the problem of how best to divide up the stakes on interrupted dice games.
From there, ideas of mathematical probability were steadily developed into all areas of science using large run, or frequentist, type approaches. They were also applied to law, finding particular uses in civil litigation in the United States of America where the methods of statistics have been used, and continue to be used, to aid courts in their deliberations in such areas as employment discrimination and antitrust legislation (Fienberg, 1988).
First suggested in the latter part of the nineteenth century by Poincaré, Darboux and Appell (Aitken and Taroni, 2004, p. 153) was an intuitive and intellectually satisfying method for placing a simple value on evidence. This employed a measure called a likelihood ratio, and was the beginning of a more modern approach to evidence evaluation in forensic science. A likelihood ratio is a statistical method which can be used directly to assess the worth of observations, and is currently the predominant measure for numerically based evidence.
Since the inception of DNA evidence in forensic science in the courts of the mid 1980s, lawyers, and indeed forensic scientists themselves, have looked towards statistical science to provide precise evaluation of the worth of evidence which follows the explicitly probabilistic approach to the evidential value of DNA matches.

1.2 Some recent uses of statistics in forensic science

A brief sample of the Journal of Forensic Sciences between the years 1999 and 2002 shows that about half of the papers have some sort of statistical content. These can be classified into: regression and calibration, percentages, classical hypothesis tests, means, standard deviations, classification and other methods. This makes knowledge of numerical techniques at some level essential, either for publication in the literature, or knowledgeable and informed reading.
The statistical methods used in the surveyed papers were:
1. Regression and calibration – regression is finding the relationship between one thing and another. For example, Thompson et al. (1999) wished to compare amounts of explosive residue detected by GC-MS with that detected by LC-UV. To do this they undertook a regression analysis which told them that the relationship was almost 1:1, that is, they would have more or less the same measurement from either method. Calibration is in some senses the complement of regression in that what you are trying to do is make an estimate of one quantity from another. Migeot and De Kinder (2002) used calibration to make estimates of how many shots an assault rifle had fired since its piston was last cleaned by the number of carbon particles on the piston.
2. Percentages and enumeration statistics – counts and proportions of objects, employed universally as summary statistics.
3. Means, standard deviations and t-tests – a mean is a measure of location†. For example, Solari and Abramovitch (2002) used stages in the development of teeth to estimate ages for Hispanic detainees in Texas. They assigned known age individuals to 10 stages of third molar development and calculated the mean age for the individuals falling in that age category. What they were then able to do was to assign any unknown individual to a developmental category, thus suggesting an expected age for that individual.
Standard deviations are measures of dispersion about a mean. In the example above, Solari and Abramovitch (2002) also calculated the standard deviation for age for each of their developmental categories. They were then able to gain some idea of how wrong they would be in assigning any age to an unknown individual. t-tests tell you how different are two samples based on the means and standard deviations of those samples. For example, Koons and Buscaglia (2002) used t-tests on elemental compositions from glass found at a crime scene to that found on a suspect to tell whether the two samples of glass possibly came from the same source.
4. Classification – this allows the researcher to assign categories on the basis of some measurement. Stojanowski and Siedemann (1999) used neck bone measurements from known sex skeletons and a discriminant function analysis to calculate a feature rule which would allow them to categorize skeletal remains as male, or female.
5. Other methods – these include χ2‡ tests and Bayesian methods.

1.3 What is probability?

When we speak of probability what is it we mean? Everybody uses the expression ‘probably’ to express belief favouring one possible outcome, or world state, over other possible outcomes, but does the term probability confer other meanings?
Examining the sorts of things which constitute mathematical ideas of probability there seem to be two different sorts. The first are the aleatory§ probabilities, such events as the outcomes from dice throwing and coin tossing. Here the system is known, and the probabilities deduced from knowledge of the system. For instance, with a fair coin I know that in any single toss it will land with probability 0.5 heads, and probability 0.5 tails. I also know that in a long run of tosses roughly half will be heads, and roughly half tails.
A second type of probability is epistemic. This is where we have no innate knowledge of the system from which to deduce probabilities for outcomes, but can by observation induce knowledge of the system. Suppose one were to examine a representative number of people and found that 60% of them were mobile telephone users. Then we would have some knowledge of the structure of mobile telephone ownership amongst the population, but because we had not examined every member of the population to see whether or not they were a mobile telephone user, our estimate based on those we had looked at would be subject to a quantifiable uncertainty.
Scientists often use this sort type of generalization to suggest possible mechanisms which underly the observations. This type of empiricism employs, by necessity, some form of the uniformitarian assumption. The uniformitarian assumption implies that processes observed in the present will have been in operation in the past, and will be in operation in the future. A form of the uniformitarian assumption is, to some extent, an inevitable feature of all sciences based upon observation, but it is the absolute cornerstone of statistics. Without accepting the assumption that the processes which cause some members of a population to take on certain characteristics are at work in the wider population, any form of statistical inference, or estimation, is impossible.
To what extent probabilities from induced and deduced systems are different is open to some debate. The deduced probability cannot ever be applied to anything other than a notional system. A die may be specified as fair, but any real die will always have minor inconsistencies and flaws which will make it not quite fair. To some extent the aleatory position is artificial and tautological. When a fair die is stipulated then we know the properties in some absolute sense of the die. It is not possible to have this absolute knowledge about any actual observable system. We simply use the notion as a convenient framework from which to develop a calculus of probability, which, whenever it is used, must be applied to probability systems which are fundamentally epistemic. Likewise, because all inferences made about populations are based on the observation of a few members of that population, some degree of deduced aleatory uncertainty is inevitable as part of that inference.
As all real probabilities are induced by observation, and are essentially frequencies, does this mean that a probability can only ever be a statement about the relative proportions of observations in a population? And, if so, is it nonsense to speak of the probability for a single event of special interest?
An idea of a frequency being attached to an outcome for a single event is ridiculous as the outcome of interest either happens or does not happen, From a single throw of a six-sided die we cannot have an outcome in which the die lands 1/6 with its six face uppermost, it either lands with the six face uppermost, or it does not. There is no possible physical state of affairs which correspond to a probability of 1/6 for a single event. Were one to throw the six-sided die 12 times then the physical state corresponding to a probability of 1/6 would be the observation of two sixes. But there can be no single physical event which corresponds to a probability of 1/6.
The only way in which a single event can be quantified by a probability is to conceive of that probability as a product of mind, in short to hold an idealist interpretation of probability (Hacking, 1966). This is what statisticians call subjective probability (O’Hagen, 2004) and is an interpretation of probability which stipulates that probability is a function of, and only exists in, the mind of those interested in the event in question. This is why they are subjective, not because they are somehow unfounded or made up, but because they rely upon idealist interpretations of probability.
A realist interpretation of probability would be one which is concerned with frequencies and numbers of outcomes in long runs of events, and making inferences about the proportions of outcomes in wider populations. A realist interpretation of probability would not be able to make statements about the outcome of a single event as any such statement must necessarily be a belief as it cannot exist in the observable world, and therefore requires some ideal notion of probability. Realist positions imply that there is something in the observed world which is causing uncertainty, uncertainty being a property external to the mind of the observer. Some might argue that these external probabilities are propensities of the system in question to behave in a specific way. Unfortunately the propensity theory of probability generates the same problem for a realist conception when applied to a single event because a propensity cannot be observed directly, and would have to be a product of mind. In many respects realist interpretations can be more productive for the scientist because of the demands that some underlying explanatory factor be hypothesized or found. This is in contrast to idealist positions where a cause for uncertainty is desirable, but not absolutely necessary, as the uncertainty resides in the mind.
This distinction between realist and idealist is not one which is seen in statistical sciences, and indeed the terms are not used. There are no purely realist statisticians; all statisticians are willing to make probabilistic statements about single events, so all statisticians are to some degree idealistic about their conception of probability. However, a debate in statistics which mirrors the realist/idealist positions is that of the frequentist/Bayesian approaches. There is a mathematical theorem of probability called Bayes’ theorem, which we will encounter in Section 9.2, and Bayesians are a school of statisticians named after the theorem. The differences between Bayesians and frequentists are not mathematical, Bayes’ theorem is a mathematical theorem and, given the tenets of probability theory, Bayes’ theorem is correct. The differences are in this interpretation of the nature of probability. Frequentists tend to argue against subjective probabilities, and for long-run frequency based interpretations of probability. Bayesians are in favour of subjective notions of probability, and think that all quantities which are uncertain can be expressed in probabilistic terms.
This leads to a rather interesting position for forensic scientists. On the one hand they do experimental work in the laboratory where long runs of repeated results are possible; on the other hand they have to interpret data as evidence which relates to singular events. The latter aspect of the work of the forensic scientist is explicitly idealistic because events in a criminal or civil case happened once and only once, and require a subjective interpretation of probability to interpret probabilities as degrees of belief. The experimental facet of forensic science can easily accommodate a more realist view of probability.
The subjective view of probability is the one which most easily fits commonsense notions of probability, and the only one which can be used to quantify uncertainty about single events. There are some fears amongst scientists that a subjective probability is an undemonstrated probability without foundation or empirical support, and indeed a subjective probability can be that. But most subjective probabilities are based on frequencies observed empirically, and are not, as the term subjective might imply, somehow snatched out of the air, or made up.
There is a view of the nature of probability which can side-step many of the problems and debates about the deeper meaning of just what probability is. This is an instrumentalist position (Hacking, 1966) where one simply does not care about the exact interpretation of probability, but rather one simply views it as a convenient intellectual devise to enable calculations to be made about uncertainty. The instrumentalist’s position implies a loosely idealist background, where probability is a product of mind, and not a fundamental component of the material world.
† Location in this context is a measure of any central tendency, for instance, male stature in the United Kingdom tends towards 5â€Č8″.
‡ Pronounced ‘chi-squared’.
§ Aleatory just means by chance and is not a word specific to statistics.

2

Data types, location and dispersion

All numeric data can be classified into one or more types. For most types of data the most basic descriptive statistics are a measure of central tendency, called location, and some measure of dispersion, which to some extent is a measure of how good is a description the measure of central tendency. The concepts of location and dispersion do not apply to all data types.

2.1 Types of data

There are three fundamental types of data:
1. Nominal data are simply classified into discrete categories, the ordering having no significance. Biological sex usually comes in male/female, whereas gender can be male/female/other. Things such as drugs can be classified by geographical area such as South American, Afghan, Northern Indian or Oriental. Further descriptions by some measure of location, and dispersion, are not really relevant to data of this type.
2. Ordinal data are again classified into discrete categories; this time the ordering does have significance. The development of the third molar (Solari and Abramovitch, 2002) was classified into 10 stages. Each class related to age, and therefore the order in which the classes appear is important.
3. Continuous data types can take on any value in an allowed range. The concentration of magnesium in glass is a continuous data type which can range from 0% to about 5% before the glass becomes a substance which is not glass. Within that range magnesium can adopt any value s...

Table of contents