This edition first published 2016 © 2016 by John Wiley & Sons, Ltd
Registered office: John Wiley & Sons, Ltd, The Atrium, Southern Gate, Chichester, West Sussex, PO19 8SQ, UK
Editorial offices: 9600 Garsington Road, Oxford, OX4 2DQ, UK
The Atrium, Southern Gate, Chichester, West Sussex, PO19 8SQ, UK
For details of our global editorial offices, for customer services and for information about how to apply for permission to reuse the copyright material in this book please see our website at www.wiley.com/wiley-blackwell.
The right of the authors to be identified as the authors of this work has been asserted in accordance with the UK Copyright, Designs and Patents Act 1988.
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by the UK Copyright, Designs and Patents Act 1988, without the prior permission of the publisher.
Designations used by companies to distinguish their products are often claimed as trademarks. All brand names and product names used in this book are trade names, service marks, trademarks or registered trademarks of their respective owners. The publisher is not associated with any product or vendor mentioned in this book.
Limit of Liability/Disclaimer of Warranty: While the publisher and author(s) have used their best efforts in preparing this book, they make no representations or warranties with respect to the accuracy or completeness of the contents of this book and specifically disclaim any implied warranties of merchantability or fitness for a particular purpose. It is sold on the understanding that the publisher is not engaged in rendering professional services and neither the publisher nor the author shall be liable for damages arising herefrom. If professional advice or other expert assistance is required, the services of a competent professional should be sought.
Library of Congress Cataloging-in-Publication Data applied for.
ISBN: 9781118492482
A catalogue record for this book is available from the British Library.
Wiley also publishes its books in a variety of electronic formats. Some content that appears in print may not be available in electronic books.
Cover image: © Dmitrii Kotin/gettyimages
This book started as part of a wider project, the examination of the applicability of logical and probabilistic reasoning to evidence generally. This has been the subject of vigorous discussion in the legal literature and is one of the main threads of the ‘New Evidence Scholarship’.
Forensic science suggested itself as a case study as there seemed to be some degree of consensus that forensic scientific evidence should be thought about in probabilistic terms, but when we surveyed the field it appeared to be a mess.
The law seemed in equal confusion.
It also seemed to us that courts and textbook writers were keener to bandy words such as ‘reliability’ and ‘regard the evidence with caution’ than to explain what ideas lay behind these phrases.
The time had clearly come for some fundamental re-evaluation of forensic science. As we studied the matter, we realised that the application of a few basic logical principles solved the problems of description and prescription with which we were faced. That is not to say that solutions came easily; the application of these principles requires hard thinking and we cannot pretend to offer answers to all the questions. The results lead to advice about how to think about evidence of much more practical value than an admonition to ‘regard the evidence with caution’.
While preparing this book we found some forensic scientists who had been thinking along the same lines and had published papers in the scientific literature. The most prolific current writer is Dr Ian Evett of the British Home Office Forensic Science Service. Gradually, and despite opposition from within the scientific and legal fraternities, these ideas have begun to appear in legal literature and to influence the giving of evidence.
The result is that while the insights in this book will seem to some readers as revelatory as they first did to us, this book is, in fact, part of a movement throughout the forensic scientific world to re-evaluate scientific evidence and, at the same time, to encourage a greater sense of unity and professionalism amongst forensic scientists. So far as we know, however, this book is the first to be written as a single book-length work on the subject.
The task of convincing forensic scientists that they must rethink their interpretation of scientific evidence is one for scientists writing in scientific journals. At some point, however, the scientist has to communicate with a lawyer and perhaps with a jury. Likewise, the lawyer who understands the law and is an expert at communicating with juries has to be able to understand the scientist. It is evident that in the past there has been a sad failure of communication.
This book attempts to remedy that. It is designed to be read by both lawyers and forensic scientists so that each will better understand the other and they will be better equipped to work together to explain the evidence to the court.
We intend that the book will also be of value to academics and students. The basic logical principles we apply provide the intellectual tool-kit for re-evaluating the law relating to expert evidence and indeed to evidence generally. We believe that this is a classic example of deep theoretical thinking appropriate to university courses providing far more practical solutions to practitioners' problems than the ad hoc reasoning which has been applied to expert evidence in the past.
In completing this task we have been helped and encouraged enormously by academic colleagues and forensic scientists including, through the wonders of electronic mail, those from the United States and the United Kingdom. Particular mention must be made of Dr Evett, who has not only been of invaluable technical assistance but who chivvied us vigorously when we were slacking on the job. Valuable comments on drafts were provided in the later stages by Richard Friedman, David Kaye and Dennis Lindley and by David Wilson of John Wiley and Sons Ltd who supported us enthusiastically. We have also benefited from discussion at many conference and staff seminar presentations at our own and other universities, and from a presentational point of view we have even benefited from the outright hostility we have met on occasions. We have conducted thoroughly enjoyable (to us at any rate) Masters and Honours courses in which a number of enthusiastic students have contributed ideas and sharpened up our presentation. Some are mentioned by name at appropriate points in the book.
We have been generously supported by research grants from the Victoria University of Wellington Internal Grants Committee, which have enabled us to employ several research assistants as the project ground through its various phases. Isobel Egerton, Andrew Fairfax, Victoria Heine, Michael Sleigh and Victoria Wicks-Brown have all contributed during vacations and term time.
Certain passages are adapted versions of papers which we have published elsewhere. More than one passage is extracted from our paper ‘Expert evidence: law, practice and probability’ (1992) 12 Oxford Journal of Legal Studies 392; the passage on stylometry is adapted from ‘Stylometric Evidence’ [1994] Crim L R 645, of which Isobel Egerton was co-author; much of Chapter 7 is to be found, differently arranged, in ‘DNA Evidence: Wrong Answers or Wrong Questions’ (1995) 96 Genetica 145; the section on fingerprints is adapted from ‘The Interpretation of Fingerprints’ (1994) 3 Expert Evidence 3. The assistance we have had from the editors and publishers of those journals is also gratefully acknowledged.
This book is based on a logical argument and the state of the law in any particular jurisdiction is not important for its thesis. Nonetheless, we have endeavoured to state the law correctly where we give examples and citations and hope that the law is correct as of 1 January 1995.
Bernard Robertson
Palmerston North
G. A. (Tony) Vignaux
Wellington
1995
It has been 20 years since the first edition of Interpreting Evidence. It was written in such a way that neither changes in the law nor advances in technology would invalidate the discussion. Since then, however, there have been substantial advances in the application of the principles discussed to new areas of forensic science. At the same time, there has been some confused reaction in the courts and little sign of great increase in understanding in the legal profession or academia.
The original authors had been asked by several scholars to prepare a new edition both to update the book so that it remained as comprehensive as possible and also because there was a need to get the book back into circulation. One of these, Charles Berger, a forensic scientist, not only urged the writing of a new edition but offered to participate and provide much-needed insight into recent advances. The authors were therefore delighted to recruit him to the team.
While the principles and Chapters 2 and 3 remain largely the same, a number of improvements have been made:
We have benefited from feedback about the first edition from forensic scientists and lawyers around the world. We are especially grateful for comments and help while preparing this edition from Colin Aitken, Niko Brümmer, John Buckleton, Christophe Champod, Ian Evett, Tacha Hicks Champod, Daniel Ramos, and Marianne Vignaux, none of whom, of course, are responsible for the views expressed or any errors made.
Bernard Robertson
Wellington
G. A. (Tony) Vignaux
Wellington
Charles E. H. Berger
The Hague
1 June 2016
Forensic scientific evidence can help us to establish:
There is a whole range of techniques used for forensic purposes, and new methods are continually being added to the arsenal of the forensic scientist. Our purpose is not to discuss the technical details of these methods, which rapidly become dated. We propose to concentrate on how such evidence should be interpreted and incorporated into the court process.1
Traditionally, several ideas have been proposed as principles for forensic science:
Edmond Locard (1877–1966) was a French forensic scientist. He proposed that we should always consider whether traces of the victim or crime scene can be found on the accused and whether traces of the accused can be found on the crime scene or victim. After an assault, for example, we might find skin and blood under a deceased's fingernails and infer that they come from the attacker. We might arrest a suspect on the basis of other evidence and find, on him or his clothing, fibres which might come from the deceased's clothes, blood which might come from the deceased or soil and plant material which might come from the scene.
The combination of these two ideas together might seem to have enormous potential value to the forensic scientist. If every contact provides ample opportunity for the transfer of traces, and every trace is different that seems to be cause for optimism. However, if no two objects are identical, then, for example, no two fingerprint impressions will be identical even if they are taken from the same finger; no two samples of handwriting by the same author will be identical. The question is whether two marks have the same source, and how much our observations help us in answering that question.
We describe these two statements as proposed principles rather than laws because neither meets the standard definition of a law of science. The philosopher Karl R. Popper (1902–1994) said that for a law to be regarded as scientific it must be potentially falsifiable, that is, it must be possible, at least in theory, to design an experiment which would disprove it.3
This ‘principle’ has a long history in forensic science, as can be seen from the following quotes that span the 20th century:
The principles which underlie all proof by comparison of handwritings are very simple, and, when distinctly enunciated, appear to be self-evident. To prove that two documents were written by the same hand, coincidences must be shown to exist in them which cannot be accidental.4
When any two items have characteristics in common of such number and significance as to preclude their simultaneous occurrence by chance, and there are no inexplicable differences, then it may be concluded that they are the same, or from the same source.5
…we look for unique characteristics in the items under examination. If we find a sufficient number of characteristics to preclude the possibility or probability of their having occurred by coincidence in two different objects, we are able to form a conclusion of individualization. It's as simple as that.6
This popular so-called principle, while simple, is fraught with problems. The possibility of a coincidence can never be completely excluded, which precludes categorical statements of individualisation. There is no general criterion possible for the number of coincidences needed to decide an individualisation; whatever level is chosen is purely arbitrary. How certain we would want to be for a decision would depend on the gravity of the crime involved (e.g. capital murder versus shoplifting). How certain we could be would also depend on other evidence and information in the case. Clearly, such issues and decisions are not up to the forensic scientist but rather the trier of fact. The role of the forensic scientist is not to decide the issue, but to describe what the evidence is worth. This ‘principle’ should therefore not be used.
In 1894, Alfred Dreyfus (1859–1935), an officer in the French army, was charged with treason in what was to become one of the most famous criminal trials in history. The charges were espionage and passing information to Germany. The espionage had definitely taken place and one of the central items of evidence was the comparison of the handwriting in an incriminating note with Dreyfus's own handwriting. A prominent witness for the prosecution was Alphonse Bertillon (1853–1914).
Bertillon was a Paris police officer who rose to found a police laboratory for the identification of criminals. He was well known for proposing a system of anthropometry, which became known as Bertillonage. Anthropometry simply means the measurement of humans. Bertillonage required taking a photograph and recording a series of measurements of bone features which were known not to change after adolescence. Later, fingerprints were added to the features recorded. The basis of the system was that it would be unlikely that any two people would have the same measurements over the whole range of features.
Bertillonage suffered from a number of problems. The method was slow and expensive and was far from error free. The officers taking the measurements had to be specially trained; this involved more expense, and even then, at the levels of accuracy called for, no two would take the same measurements from the same series of features. Nor could the system be applied to juveniles.
The purpose of the system was to determine whether or not a person had the same measurements as a person who had earlier been arrested. This can be very useful, for example, when someone is arrested on suspicion of failing to attend court or when a person being sentenced denies that previous convictions relate to him. However, Bertillonage could not help investigators by providing evidence that a particular person had been, for example, at the scene of a crime.
Although fingerprints were later taken as one of the Bertillonage measurements and Bertillon himself solved a crime using fingerprints in 1902, there was no formal classification system for them. Once such systems were developed (by Galton and Henry in England and India, and Vucetich in Argentina) it was possible to quickly exclude the majority of the fingerprint collection (i.e. the other classes) on each search. Fingerprints became a far quicker and simpler method of identification than anthropometry. In the first full year of operation by the London Metropolitan Police, fingerprints identified 3 times as many persons as anthropometry and, 2 years later, 10 times as many. Not only were fingerprints far simpler and cheaper to obtain and record but they could also help investigators identify the perpetrators of crimes. Bertillonage was dropped.
Bertillon gave evidence in the Dreyfus case as a handwriting expert and claimed that Dreyfus had written the incriminating document. His evidence referred to certain similarities and multiplied together the probabilities of each of the similarities occurring by chance to arrive at a very low probability of them occurring together by chance. His evidence was subjected to devastating critique by a number of people including Poincaré, an eminent mathematician.7 Poincaré made three important points about Bertillon's evidence. The first was that Bertillon had simply erred in that the figure he produced was the probability of getting the four similarities amongst four examined characteristics. There were far more characteristics examined, and so the chances of finding four similarities were actually much greater than Bertillon's figure. The second point Poincaré made was that events that have actually occurred might be seen beforehand as highly improbable. The example he gave was the drawing of a particular number or set of numbers in a lottery. The probability that any particular set of numbers will be drawn is extremely low. Once it has been drawn, however, that low probability does not mean that the draw has been dishonest.
Most importantly of all, Poincaré discussed what is called the inverse probability problem, the difference between calculating in advance the probability of an effect and calculating after the event the most probable cause of an effect:
As an example of probability of effects, we usually choose an urn containing 90 white balls and 10 black balls; if we randomly draw a ball from this urn, what is the probability for this ball to be black; it is evidently 1/10.
The problems of probability of causes are far more complicated, but far more interesting.
Let us suppose for example two urns of identical exterior; we know that the first contains 90 white balls and 10 black balls, and the second contains 90 black balls and 10 white balls. We draw arbitrarily a ball from one of the urns, without knowing from which, and we observe that it is white. What is the probability that it came from the first urn?
In this new problem, the effect is known, we observed that the ball drawn was white; but the cause is unknown, we do not know from which urn we made the draw.
The problem that we are concerned with here is of the same nature: the effect is known, the indicated coincidences on the document, and it is the cause (forgery or natural writing) that is to be determined.8
Poincaré identifies a crucial point for forensic science and, indeed, all reasoning about evidence in court. This is a central theme of this book and will be explained in the following chapters. Courts are not concerned with the probability that some observation would be made. They are concerned with what can be inferred from the fact that the observation has been made. The question for the court then is what inferences can be drawn as to the guilt of the accused.
Poincaré went on to make the point that single items of evidence enable us to alter our assessment of the probability of an event but they cannot determine the probability of an event on their own9
:
To be able to calculate, from an observed event, the probability of a cause, we need several data:
Photographs are still used to help identify criminals and are recorded with the details of their convictions. They have a number of advantages: they can be transmitted and reproduced easily and can enable people to be recognised at a distance. In most cases, a photograph will settle a question of identity. Where this is seriously challenged, however, a photograph is of questionable value, particularly if much time has passed since it was taken.10 Similarly, physical descriptions can be broadcast on police radios and even the most rudimentary description will eliminate a large proportion of the population. However, when identity is seriously challenged, descriptions and even eyewitness identification are of questionable value, perhaps because the question has become whether the perpetrator was the accused or someone else of similar appearance.
The limitations of Bertillonage prompt us to consider the features of an ideal scientific system for identifying people. These would include:
Inevitably, few systems will satisfy all these requirements and in particular there may be a trade-off between the last requirement and the others. Each of the systems that we examine later will satisfy some of these requirements but not all.
If we can establish features that help distinguish between individuals or groups, it becomes useful to maintain a database of observed features of known individuals. Large databases of DNA profiles have now been established, as happened with fingerprint collections over the last century. In investigations, such databases allow police to search for individuals that could have left a crime scene trace.
If a suspect has been identified and the observed features of this known person are, for example, similar to those of the traces from the crime scene, we need to evaluate what those observed similarities are worth. If the suspect had nothing to do with the crime, what would be the probability of finding those similarities? That probability can be assessed with the help of databases of features that are representative of some population. It does not require the contributors to the database to be known.
As we have seen, evidence should not be expected to give certainty. This does not make evidence ‘unreliable’ or inadmissible. Lawyers often tend to ignore evidence that does not claim to provide certainty, but by doing so, they lose relevant and probative evidence.11 Uncertainty is inherent to the limited amount of information that can be extracted from traces, which may be minute, old and contaminated. Poincaré did not tell us to simply discard such evidence, but to assess the probability of the observed effects for the possible causes.
It follows that a scientific witness will not, in principle, be able to say that two samples came from the same person. The evidence can only lead to an assessment of the probabilities that the evidence would be found if the prosecution case was true and if the defence case was true. The legal system has not been successful in dealing with this kind of evidence, and our purpose is to explain how such evidence should be given and integrated into the case.
Rather than think rigorously about these problems, the legal system has been prone to ask questions such as ‘how reliable is this evidence?’. This question is difficult to answer since ‘reliable’ appears to have no fixed meaning. We discuss its different possible meanings and the consequences of each in Chapter 5.
We adopt a structure different from that of most other books on forensic scientific evidence. Those intended for scientists are usually built round the different techniques available. Those for lawyers are often structured round rules such as the Basis Rule, the Field of Expertise Rule, the Qualifications Rule and the Ultimate Issue Rule. That such a structure is unsatisfactory is shown by the extent to which these ‘rules’ are intertwined. Courts sometimes refer to one, sometimes to another. Cases that are decided on the basis of one rule are often explicable by reference to another. In this book, we