Forensic scientist Andrew Feiter works in the D.C. crime lab in 2016. (Photo by Bill O’Leary/The Washington Post)
Opinion writer

CORRECTION: This post links to reports from the Minnesota Department of Public Safety Bureau of Criminal Apprehension (BCA) crime lab, and included a quote from the director of that lab from one of those reports. The links and quote were included in a list of examples of crime labs that have been accredited despite a history of misconduct and sloppiness. However, the crime lab problems in St. Paul occurred at the crime lab of the St. Paul Police Department, not the St. Paul branch of the BCA lab. The two labs are not affiliated. My apologies for the error.

Also, a clarification: I wrote in this post that, “The links to the pages listing revoked/suspended labs and labs on probation now go to a site called the ANSI-ANQ National Accreditation Board, which apparently merged with ASLD’s accrediting arm. But neither website appears to lists any crime labs on probation, currently in suspension, or whose accreditation has been revoked.” 

A reader points out that while the pages are tough to find, there are lists of both “suspensions” and “withdrawals.” The lists are difficult to decipher but the vast majority of “withdrawals” were voluntary — likely the result of a lab shutting down, letting its accreditation lapse, or switching to another type of testing. In the last 10 years, just three labs have had accreditation withdrawn for “nonconforming” or “nonconformance.” And because ANSI-ANQ accredits lots of labs other than crime labs, and because the site only lists labs by three or four initials, it isn’t even clear that these three were crime labs. (The three labs are listed as TP, EQAM, and ICL. An Internet search suggests two of the three are not crime labs. I couldn’t positively identify the third.) An archive of the page that appears to list revocations specifically for crime labs that was active as of last year simply says, “There are no laboratories for which accreditation has been revoked.”

There have been significantly more suspensions for “nonconforming,” but here again, from what I can tell, very few of them are crime labs. Here again, if we look at the Internet Archive, a cached page from the old side from April of last year says, “There are no laboratories for which accreditation is currently suspended.”

So I was incorrect to write that the new site doesn’t list labs for which accreditation has been suspended or revoked. It does. But the lists themselves aren’t particularly helpful or user-friendly, and to the extent that one can decode them, it appears that very few — if any — crime labs have had to worry about their accreditation.

Last October, the Fordham University Law School hosted a symposium on forensics and how they’re used in the courtyard. Presenters then wrote up their presentations for publication in the March 2018 Fordham Law Review.

I want to address an essay by Alice R. Isenberg and Cary T. Oien, titled “Scientific Excellence in the Forensic Science Community.” Isenberg is the deputy assistant director of the Laboratory Division at the FBI. Oien is a senior forensic scientist at the FBI. As you might imagine, I have some issues with their essay.

Generally, the piece makes the case that the alarm about forensics is much ado about very little. It’s also a defense of how forensics is used by the federal government, and by the FBI specifically.

So let’s jump right in:

The practice of forensic science has existed for centuries. Each year, hundreds of thousands of cases are closed, suspects cleared, and offenders convicted through routine, accurate, and reliable forensic testing. Forensic testing includes chemical analysis to determine the nature of seized drugs; examinations performed on physical materials such as fibers, glass, and spent bullet casings; and examination of biological materials such as DNA. Tests performed for each of these examinations, regardless of the materials examined, are strictly prescribed by laboratory policies, supported by peer-reviewed research, and lead to accurate and reliable results.

The first fingerprint case in the United States was in 1911. Some early forebears of what we call “forensics” today were used in some European courts going back to the 19th century, but most fields of modern forensics were developed and first used in the Progressive Era. So while the U.S. court system has been around for centuries (just long enough to use the plural), forensics as we know it today is about 100 years old.

As for “peer-reviewed research,” it depends on how you define the term. Some forensics journals claim to be peer-reviewed. But that review is generally done by other forensic specialists. The truth is, most fields of forensics weren’t subjected to rigorous scientific testing until the last 10 to 15 years. That testing has shown that, in many fields, expert witnesses have been giving testimony that is either completely unsupported by science or they’ve been significantly exaggerating the significance of their findings.

“Accurate and reliable results?” Well, that depends. The most problematic fields of forensics are those known as the pattern matching fields. This includes any specialty that requires an analyst to look at one sample and “match” it to another. Think hair and carpet-fiber analysis, bite-mark analysis, shoe-print and tire-tread analysis, blood-spatter analysis and fingerprint matching. The degree to which these fields are problematic vary quite a bit (bite-mark matching is probably on the least reliable end of the spectrum, with fingerprint matching at the other end), but all at their core are subjective. (Fingerprint matching breaks down the moment you start looking at partial prints.) That means they cannot calculate a margin for error. It means analysts will often disagree about conclusions, sometimes in ways that directly contradict one another. And by definition, any method of analysis that results in experts coming to contradictory conclusions about the same piece of evidence can’t possibly be accurate (one of them is obviously wrong) or reliable.

This means that these fields aren’t science. That doesn’t mean they have no evidentiary value at all. But it does mean that analysts need to be extremely careful about how they present this sort of evidence to juries. The language they use needs to be standardized and then explained to juries, so that the amount of emphasis the jury puts on it is based on the evidence’s actual significance and not other factors, such as the charisma or persuasiveness of the analyst. This hasn’t been happening.

A casual reader of recent media reports might be led to believe that forensic science lacks any scientific credibility. However, this narrative is completely inaccurate and at odds with the scientific excellence that exists throughout the forensic science community. Forensic disciplines are grounded in diverse sciences such as chemistry, biology, and physics, and every forensic discipline practiced in an accredited forensic laboratory must demonstrate that it is reliable, accurate, and fit for its intended use.

There are a lot of subjective and undefined terms in this paragraph. To date, a number of commissions, panels and other bodies staffed with actual scientists have reviewed some of the most common fields of forensics and found them lacking in scientific merit. These include the National Academy of Sciences, the Texas Forensic Science Commission, the President’s Council of Advisors on Science and Technology and the National Commission on Forensic Science.

To say that several forensic disciplines are “grounded in diverse sciences” tells us little about whether the actual methods of analysis that the practitioners use to reach their conclusions are scientifically sound and reliable. One important test of scientific merit, for example, is repeatability. If a method of forensic analysis is scientifically reliable, two trained analysts using the same experiment with the same evidence should reliably get the same result. In a scientifically proven field such as DNA testing, two trained scientists testing the same blood, hair, semen, skin cells or other biological material will reliably produce the same DNA profile. The same can’t always be said of blood-spatter analysts, tool-mark analysts or bite-mark analysts. 

One could say, for example, that blood-spatter analysis is “grounded in” physics — the physics of what a liquid such as blood does as it is forced from the body, travels through the air and strikes a wall or floor. But simply referencing some bit of scientific knowledge doesn’t mean that the end product of your analysis will be scientifically reliable. Astrologers, for example, utilize some principles and knowledge from astronomy, which is a legitimate field of science. This doesn’t mean that astrologers’ predictions are “grounded in science.” Just last week, the New York Times ran an editorial on this, based on a two-part investigation into a conviction based on blood-spatter evidence, written by Pamela Colloff.

Joe Bryan was convicted on the word of a detective named Robert Thorman, who testified before the jury as an expert in what is known as bloodstain-pattern analysis . . . People like Detective Thorman got certified as bloodstain-pattern analysts after taking a weeklong course that now costs as little as a few hundred dollars.

Pamela Colloff, who wrote the articles on the Times Magazine/ProPublica investigation of Joe Bryan’s case, enrolled in one of these courses, where the instructor told her, “We’re not really going to focus on the math and physics; it just kind of bogs things down.” Ms. Colloff passed the final exam, as did everyone in the class.

Thanks in part to such dubious standards, the interpretation of bloodstain evidence has become notoriously ambiguous. The same patterns can, like a Rorschach test, be read in very different ways; some trials feature two bloodstain “experts,” one on each side, who testify to opposite conclusions. A 2009 report by the National Academy of Sciences found that “the opinions of bloodstain-pattern analysts are more subjective than scientific,” and, “The uncertainties associated with bloodstain pattern analysis are enormous.”

And yet judges in many states have accepted these experts’ testimony as scientifically valid — not because of any concrete evidence that it is, but because other courts have accepted it before.

This of course is how courts routinely adjudicate challenges to the scientific validity of expert testimony. They look to see what other courts have done. This means that once a scientifically dubious field enters the criminal justice system, it becomes really difficult to reverse the process.

Let’s get back to Isenberg and Oien.

Accreditation and quality assurance systems assure the public that accredited organizations are competent and their results can be relied upon. Many groups — such as the National Commission on Forensic Science, the National Academy of Sciences, the President’s Council of Advisors on Science and Technology (PCAST), and the Department of Justice (DOJ)— recognize that accreditation is critically important. In fact, in December 2015, the Attorney General directed that all DOJ forensic laboratories must obtain or maintain accreditation.

About that 2015 directive. It has a pretty big loophole. It only states that federal prosecutors use accredited crime labs “when practicable.” As Frontline reported at the time, if finding an accredited lab would result in too much of a delay or too great an expense, federal prosecutors can continue to use unaccredited labs.

Accreditation is an external assessment of a laboratory’s technical competence to perform specific types of testing. Accreditation demonstrates that a laboratory is performing its work correctly and consistent with appropriate standards. To maintain this recognition, a laboratory is periodically reevaluated to ensure its ongoing compliance with accreditation requirements. Laboratory accreditation is internationally regarded as a reliable indicator of technical competence, and it provides credibility and public confidence in laboratory operations. An accredited laboratory’s quality assurance system must include written standard operating procedures, proficiency testing, training programs, processes for technical review of reports, testimony monitoring, and many other requirements.

All else being equal, some accreditation is certainly better than no accreditation. But accreditation is merely a baseline. It doesn’t ensure competency. It doesn’t ensure that a crime lab is operating within the parameters of sound science. In the end, an accreditation is only as valuable as the rigor, reputation and thoroughness of the accrediting organization. And as I’ll get to in a moment, the history of these groups leaves a lot to be desired.

Accrediting bodies also typically exist within the existing culture of forensics. Typically, when we talk about accreditation, we aren’t talking about outside scientists who are assessing the scientific credibility and rigor of a crime lab and its analysts. These accrediting organizations usually begin their evaluations of crime labs from the perspective that even the more subjective and controversial fields of forensics, when applied according to the prevailing standards of those fields, are legitimate and reliable. In other words, accrediting organizations evaluate subjective fields such as blood-spatter, shoe-print, or hair-fiber analysis on their own respective terms. If you’re meeting the minimum standards put forth by the blood-spatter, shoe-print, or hair-fiber analysis community, you’re probably going to be accredited. Accreditation does not mean that these fields as a whole are scientific, credible or reliable.

According to the Bureau of Justice Statistics, 88 percent of the 409 publicly-funded forensic crime labs in the United States are accredited. Unaccredited labs are often very small — less than ten people — and offer services in a limited number of disciplines. In addition to forensic laboratories, laboratories performing other types of tests are accredited according to the same international standard. This includes environmental labs checking for levels of lead in groundwater, chemistry labs preparing chemicals for consumer use, or food labs ensuring the safety of our food supply.

This is a strange comparison. The fact that a crime lab is subjected to similar standards as labs that perform other types of tests says nothing about the validity of the crime lab’s analysis.

But let’s look at that 88 percent figure. It seems impressive. But it seems impressive only until you start to look at the seemingly endless parade of crime lab scandals we’ve seen over the past 15 to 20 years, and you begin to notice how many of those labs . . . were already accredited.

Here’s a quick list of such cases, which isn’t remotely comprehensive:

  • Last October, Massachusetts officials fired the head of the Office of Alcohol Testing within the Massachusetts State Police Crime Laboratory after discovering that analysts at the lab routinely withheld exculpatory information about blood-alcohol tests from defense attorneys. The revelation could affect thousands of drunk-driving tests dating back to 2011. On its website, the Massachusetts State Police Crime Laboratory describes itself as “fully accredited.”
  • In 2016, police officials in Austin shut down the DNA division of the city’s police-run crime lab indefinitely after discovering that analysts were using analytical tools that were badly outdated. The lab also had problems with cross-contamination. Despite multiple red flags about the lab dating back to 2008, the lab was repeatedly accredited by the appropriate bodies within the field of forensics. It took an inspection by the Texas Forensic Science Commission — an oversight body from outside the forensics world — to finally bring these problems to light.
  • In 2013, the New York City Medical Examiner’s Office underwent a review of 800 rape cases after investigators found 26 instances in which DNA evidence was mishandled or overlooked by an analyst at the lab. The cases spanned 2001 to 2011. In fact, the same year that the city’s Medical Examiner’s Office announced this review, it also boasted in a news release that its DNA lab received “perfect scores” from an international accrediting organization.
  • In their various incarnations, the crime labs in Houston have been the scene of multiple scandals, including 2014 revelations that an analyst had tampered with evidence; a 2005 report finding incompetence, cheating on proficiency tests, and even possible perjury and faking of test results; and the shuttering of the fingerprint lab in 2009 after an audit showed exceptionally high error rates. (This is not at all a comprehensive list of the crime lab problems in Houston.) Through all of this, the Harris County and Houston labs have boasted of accreditation from both national and international forensic organizations. The American Society of Crime Laboratory Directors, for example, accredited the Houston PD lab in 2006. Incredibly, when a 2008 investigation into cheating on proficiency tests led to the closing of the lab’s DNA testing unit, the resignation of the head of that unit and the suspension of two other analysts, the head of the organization that accredited the lab said his group had no plans to revoke the lab’s accreditation.
  • A 2013 audit of the crime lab in St. Paul, Minn., found widespread ineptitude, including “sloppy documentation, dirty equipment, faulty techniques and ignorance of basic scientific procedures.” The report recommended that the lab cease operations until it could be overhauled from the ground up. Yet in its 2007, 2008, 2009 and even 2012 annual reports, the Minnesota Department of Public Safety touted the St. Paul crime lab’s accreditation by the American Society of Crime Lab Directors. In 2009, three years before the scandal broke, crime lab director Frank Dolejsi boasted that the accreditation “is an affirmation of the quality of all aspects of the [crime lab’s] forensic science services.” After the audit, state officials vowed to win back public trust by promising that the St. Paul crime lab would — you guessed it — get itself accredited. (*SEE CORRECTION ABOVE)
  • In 2014, a crime lab analyst in Florida was arrested for stealing seized drugs and tampering with evidence, possibly tainting up to 2,600 cases from at least 80 different law enforcement agencies. His lab was accredited.
  • In 2012, an audit of the North Carolina crime lab done after an exoneration found that analysts in the serology unit had been withholding exculpatory evidence from defense attorneys for at least 16 years. The audit found that analysts routinely overstated their findings and that training manuals were adamantly pro-prosecution, to the point of referring to defense experts as “whores.” The manuals were likely pro-prosecution because the analysts themselves reported to prosecutors, who wrote their year-end reviews and determined raises and promotions. The North Carolina lab had been accredited since 1988.

These are just a few examples. In 2013, the ABA Journal reviewed dozens of crime lab scandals across the country. The report found a lot of problems in both accredited and non-accredited labs. Of course, there’s Isenberg and Oien’s own FBI crime lab, which is arguably the most accredited and respected crime lab in the world. But the FBI lab, too, has been no stranger to scandal. Just a few years ago, the agency admitted that it’s hair-fiber analysts had overstated their findings nearly every time they testified. This occurred over a period spanning decades and implicated thousands of convictions. Those analysts then trained analysts at state and local labs, potentially corrupting untold thousands of more cases. And that’s just one scandal of several. There’s the Brandon Mayfield debacle. There’s the FBI’s long-held and since disproved claim about “compositional bullet lead analysis,” which also affected thousands of criminal cases. That’s quite a record from one of the most respected labs in the world.

Back to Isenberg and Oien.

Forensic examiners must complete extensive training to be qualified to perform casework in accredited laboratories. Training programs can be one to two years, or longer, and require examiners to perform analyses on samples with a known correct answer. The examiner must also demonstrate a thorough understanding of the science behind the method employed and an understanding of lab policies, procedures, legal rules, evidence handling, etc. The examiner must undergo oral examinations, mock trials, and competency tests for which the correct answer is known. In addition, all examiners must demonstrate competency to apply the processes accurately and reliably before they are assigned actual cases. Once qualified to conduct casework in an accredited laboratory, every examiner undergoes continual competency monitoring through proficiency tests administered at least once per year.

This is difficult to square with the examples above, or with the other scandals at accredited labs far too numerous to list in their entirety.

The most common credentialing agency for crime labs is the aforementioned American Society of Crime Laboratory Directors (ASCLD). Isenberg and Oien even cite this organization’s website in laying out the allegedly rigorous process to becoming an accredited forensic examiner. But while ASCLD may claim that these are the criteria crime labs need to meet to earn and keep their credentials, those requirements don’t appear to be all that tightly enforced. Labs that don’t meet them still get accredited. And previously accredited labs that don’t meet them don’t seem to lose their status.

From that 2013 report by the ABA Journal:

If the accreditation process is so rigorous and demanding, critics wonder, then why have so few labs been sanctioned? ASCLD/LAB’s website lists the status of all accredited labs and shows that no lab’s accreditation is currently revoked or suspended; there are also no labs on probation. And [ASCLD Executive Director Ralph] Keaton says he can count on one hand the number of labs whose accreditation has ever been revoked or suspended, though he says it would probably take two hands to count the number of labs that have ever been placed on probation.

Keaton says that has a lot to do with the overall quality of accredited labs. But critics say it has more to do with the chummy nature of the inspection process, which creates a tendency to “go along to get along” among inspectors, and the agency’s own interest in keeping labs accredited. . . .

New York City criminal defense lawyer Marvin Schechter, a member of the committee that produced the NAS report, is one of ASCLD/LAB’s biggest critics. Schechter, also a member of the New York State Commission on Forensic Science, wrote a lengthy memo to his fellow commissioners in 2011 recommending that they look for a new accreditor. He characterized ASCLD/LAB as an organization more interested in protecting its members’ images than in promoting accountability.

“In fact, ASCLD/LAB could more properly be described as a product service organization,” Schechter wrote, “which sells for a fee a ‘seal of approval’ covering diverse laboratory systems, which laboratories can utilize to bolster their credibility through in-court testimony by technicians, plus ancillary services such as protection from outside inquiry, shielding of internal activities and, where necessary, especially in the event of public condemnation, a spokesperson to buffer the laboratory from media inquiry.”

The links to the pages listing revoked/suspended labs and labs on probation now go to a site called the ANSI-ANQ National Accreditation Board, which apparently merged with ASLD’s accrediting arm. But neither website appears to lists any crime labs on probation, currently in suspension, or whose accreditation has been revoked. If even egregious incompetence and misconduct rarely if ever results in a revocation or suspension, what does accreditation really mean?

More from Isenberg and Oien: 

Testimony monitoring is also a requirement for accredited laboratories. The Federal Bureau of Investigation (FBI) Laboratory requires that examiners request a transcript for each testimony provided. FBI examiners also must follow approved standards for scientific testimony and reports, which document the acceptable range of conclusions expressed in both laboratory reports and testimony. The DOJ is developing similar documents called Uniform Language for Testimony and Reports, as well as a testimony-monitoring framework, which will apply to all DOJ laboratories. The purpose of these testimony-monitoring activities is to prevent examiner testimony from exceeding scientific limitations.

This sounds great at first. But it’s all dependent on the answers to several important questions. First, who determines the “acceptable range of conclusions” for lab reports and testimony? Is it someone who already practices in that particular field? What does “approved standards for scientific testimony” mean? Who is setting those standards? Who is approving them? Who decides if an analyst has “exceeded scientific limitations?”

The entire point of the National Commission on Forensic Science empaneled by President Barack Obama was to bring in actual scientists to make such determinations. Attorney General Jeff Sessions allowed that commission’s charter to expire shortly after taking office. (It’s worth noting that while Obama deserves credit for setting up the commission, his record in this area was mixed at best.) Last August, Deputy Attorney General Rod J. Rosenstein announced a new forensics review and monitoring process at DOJ. He also revealed that the new system would be overseen by a man named Ted Hunt.

Ted Hunt is not a scientist. He is a longtime prosecutor. His supporters point out that he was part of the very Forensic Science Commission that Obama empaneled and that reformers lament was allowed to expire. But Hunt was often a voice of dissent on the commission. Here’s a look at Hunt’s record on the commission from Mother Jones:

In March 2016, the commission recommended that then-Attorney General Loretta Lynch direct forensic experts and attorneys working on behalf of the Justice Department to stop using the phrase “to a reasonable degree of scientific certainty.” The phrase is commonly used on witness stands and in lab reports and gives juries and judges a sense of factuality, but it is subjective and lacks any agreed-upon meaning across the sciences. Hunt was one of two commission members who opposed the recommendation, which Lynch adopted last September.

Lynch also adopted a recommendation by the commission requiring forensic testing labs that work with the department and its attorneys to publicize their internal procedures, from equipment maintenance to estimations of uncertainty, in order to foster transparency, trust, and best practices in the industry. Hunt was one of four commissioners who opposed it.

Last September, when the commission released a document supporting stricter accreditation standards for forensic labs, Hunt voted against it. And when the commission recommended that the National Institute of Standards and Technology conduct scientific evaluations of the “technical merit of test methods and practices used in forensic science disciplines,” he opposed that, too. At its final meeting, when members had already been informed that the group would be coming to an end, several commissioners pushed for a resolution encouraging experts to use more quantitative language to convey the accuracy of forensic testimony. The resolution narrowly failed, with Hunt among the nays.

So DOJ did away with a transparent, external commission mostly populated and led by scientists, and replaced it with an internal commission led by a career prosecutor who has opposed efforts to increase transparency, increase accountability, increase scientific accuracy and strengthen the requirements for accreditation. This is likely why Hunt’s appointment was hailed by ASCLD and forensics groups such as the International Association for Identification and the American Academy of Forensic Sciences, and why the new DOJ venture has been viewed with skepticism by groups such as the Innocence Project.

In the remainder of their essay, Isenberg and Oien take aim at the report by the President’s Council of Advisors on Science and Technology (PCAST) issued in the waning months of the Obama administration. That report was perhaps the most strident and explicitly critical report on forensics to date. The authors of and researchers for the report were scientists, not forensics practitioners or lawyers (though a panel of judges and attorneys did contribute in an advisory role).

The essay’s discussion of the PCAST report is a bit technical to delve into here with too much detail, but in general, the PCAST authors posited that because most pattern-matching fields of forensics are so subjective — an analyst basically just “eyeballs” two samples and comes up with language to describe how similar or different they are — the only way to determine if these fields are reliable is what’s known as a “black box” test. We can’t evaluate a particular analyst’s process for determining matches because so much of it is done in his or her head. Instead, we have to look at results. That means administering competency tests. Give analysts a sample from a case in which the culprit is known, give them samples from multiple suspects, and record who gets it right and who doesn’t. As you might guess, many forensic analysts and the professional groups they belong to are reluctant to submit to such tests. Most of these fields have already been accepted by the courts. They have little to gain, and a lot to lose.

In their essay, Isenberg and Oien dispute the notion that black-box tests are the only real way to measure the reliability of a given field. They argue that relying too much on such tests “ignores much peer-reviewed research, overlooks critical aspects of many studies, and fails to acknowledge the empirical value of these studies.” But these studies are, again, typically reviewed only by other practitioners from the same field, using the same procedures and standards within the field. Here again, you’re evaluating the legitimacy of an entire field based on criteria that already accepts that field’s legitimacy.

In conclusion, Isenberg and Oien write:

Science continuously evolves and is built upon observation, study, and experience that spans hundreds of years. The justice system would not be well served by the exclusion of reliable forensic methods and techniques that provide valuable information to a wide range of stakeholders.

Of course, whether some of these fields are reliable is precisely what’s in dispute.

But at issue here is how we evaluate new or questionable expertise for use in the courtroom. What do we do going forward? I think the answer is that we only allow this sort of analysis and expertise after it has been subjected to the rigors of scientific inquiry. Before we let a new drug hit the market, we require its manufacturer to prove that the drug is safe and effective. Those standards are enforced by scientists with expertise in pharmaceuticals. When a drug slips through and causes harm, we pull it from the market. It isn’t a perfect system, and there are interesting debates at its parameters. But in general, we test new technology before we put it to everyday use.

Contrast this to forensics. The decision whether to allow a new field of forensics into court is made by a judge, not a scientist, or even a fellow practitioner. Judges typically look for guidance on these questions not from scientists, but from other judges. The briefs in such challenges are written by lawyers. Judges then tend to err on the side of letting evidence in, on the assumption that our adversarial system will sort it out. (In his speech last year, Rosenstein touted this flaw as a feature.) Even once we discover that a field is scientifically suspect, it’s difficult to get the courts to even acknowledge it, much less stop it from being used again, much less correct the cases that may have already been tainted.

Like the other defenders of these fields, Isenberg and Oien say we should simply trust the internal procedures at FBI and DOJ to get forensics right. Even at face value, that’s a difficult argument to accept. FBI agents and federal prosecutors aren’t evaluated for their allegiance to scientific principles. They’re evaluated on their ability to close cases and win convictions. Therefore, there’s a strong incentive for them to sacrifice sound science for expedience. That incentive will affect even the most conscientious and fair-minded at DOJ.

But the case Isenberg and Oien are making is even harder to accept given the history of these disciplines and the ongoing onslaught of forensics and crime lab scandals. “Just trust us” hasn’t worked in the past. Why should we think it would work now?