TORTINI

For your delectation and delight, desultory dicta on the law of delicts.

On The Quaint Notion That Gatekeeping Rules Do Not Apply to Judges

April 27th, 2014

In In re Zurn Pex Plumbing Prods. Liab. Litig., 644 F.3d 604 (8th Cir. 2011), the United States Court of Appeals for the Eighth Circuit rejected the defendant’s argument that a “full and conclusive” Rule 702 gatekeeping procedure was required before a trial court could certify a class action under the Federal Rules. The Circuit remarked that “[t]he main purpose of Daubert exclusion is to protect juries from being swayed by dubious scientific testimony,” an interest “not implicated at the class certification stage where the judge is the decision maker.”  Id. at 613.

Surely, one important purpose of Rule 702 is to protect juries against dubious scientific testimony, but judges are not universally less susceptible to dubious testimony.  There are many examples of judges being misled by fallacious scientific evidence, especially when tendentiously presented by advocates in court.  No jury need be present for dubious science testimony + “zealous” advocacy to combine to create major errors and injustice.  See, e.g., Wells v. Ortho Pharmaceutical Corp., 615 F. Supp. 262 (N.D. Ga. 1985)(rendering verdict for plaintiffs after bench trial), aff’d and rev’d in part on other grounds, 788 F.2d 741 (11th Cir.), cert. denied, 479 U.S.950 (1986); Hans Zeisel & David Kaye, Prove It With Figures: Empirical Methods in Law and Litigation § 6.5 n.3, at 271 (1997) (characterizing Wells as “notorious,” and noting that the case became a “lightning rod for the legal system’s ability to handle expert evidence.”).  Clearly Rule 702 does not exist only to protect juries.

Nemo iudex in causa sua! Perhaps others should judge the competence of judges’ efforts at evaluating scientific evidence.  At the very least, within the institutional framework of our rules of civil procedure and evidence, Rule 702 creates a requirement of structured inquiry into expert opinion testimony before the court.  That gatekeeping inquiry, and its requirement of a finding, subject to later appellate review and to public and professional scrutiny, are crucial to the rendering of intellectual due process in cases that involve scientific and technical issues.  The Eighth Circuit was unduly narrow in its statement of the policy bases for Rule 702, and their applicability to class certification.

The case of Obrey v. Johnson, 400 F.3d 691 (9th Cir. 2005) provides another cautionary tale about the inadequacies of judges in the evaluation of scientific and statistical evidence.  The plaintiff, Mr. Obrey, sued the Navy on a claim of race discrimination in promoting managers at the Pearl Harbor Naval Shipyard.  The district court refused plaintiff’s motion to admit the testimony of a statistician, Mr. James Dannemiller, President of the SMS Research & Marketing Services, Inc. The district court also excluded much of plaintiff’s anecdotal evidence, and entered summary judgment.  Id. at 691 – 93.

On appeal, Obrey claimed that Dannemiller’s report showed “a correlation between race and promotion.” Id. at 693. This vague claim seemed good enough for the Ninth Circuit, which reversed the district court’s grant of summary judgment and remanded for trial.

The Ninth Circuit’s opinion does not tell us what sort of correlation was supposedly shown by Mr. Dannemiller. Was it Pearson’s r?  Or Jaspen’s multi-serial coefficient? Spearman’s ρ?  Perhaps Kendall’s τ? Maybe the appellate court was using correlation loosely, and Mr. Dannemiller had conducted some other sort of statistical analysis. The district court’s opinion is not published and is not available on Westlaw.  It is all a mystery. More process is due the litigants and the public.

Even more distressing than the uncertainty as to the nature of the correlation is that the Ninth Circuit does not tell us what the correlation “effect size” was, or whether the correlation was statistically significant.  If the Circuit did not follow strict hypothesis testing, perhaps it might have told us the extent of random error in the so-called correlation.  The Circuit did not provide any information about the extent or the precision of the claim of a “correlation”; nor did the Circuit assess the potential for bias or confounding in Mr. Dannemiller’s analysis.

Indeed, the Ninth Circuit seemed to suggest that Mr. Dannemiller never even showed a correlation; rather the court described Mr. Dannemiller as having opined that there was “no statistical evidence in these data that the selection process for GS-13 through GS-15 positions between 1999 and 2002 was unbiased with respect to race.” Id. at 694. Reading between the lines, it seems that the statistical evidence was simply inconclusive, and Mr. Dannemiller surreptitiously shifted the burden of proof and offered an opinion that the Navy had not ruled out bias. The burden, of course, was on Mr. Obrey to establish a prima facie case, but the appellate court glossed over this fatal gap in plaintiff’s evidence.

On appeal, the Navy pressed its objections to the relevance and reliability of Mr. Dannemiller’s opinions. Brief of the Navy, 2004 WL 1080083, at *1 (April 7, 2004).  There seemed to be no dispute that Mr. Dannemiller’s “study” was based entirely upon “statistical disparities,” which failed to take into account education, experience, and training.  Mr. Dannemiller appeared to have simplistically compared race make up of the promoted workers, ignoring the Navy’s showing of the relevancy of education, experience, and training.  Id. at *13, 18.

The Ninth Circuit not only ignored the facts of the case, it ignored its own precedents.  See Obrey v. Johnson, 400 F.3d at 696 (citing and quoting from Coleman v. Quaker Oats Co., 232 F.3d 1271, 1283 (9th Cir. 2000) (“Because [the statistics] failed to account for many factors pertinent to [the plaintiff], we conclude that the statistics are not enough to take this case to trial.”). The court, in Obrey, made no effort to distinguish its treatment of the parties in Coleman, or to justify its decision as to why the unspecified, unquantified, mysterious statistical analysis of Mr. Dannemiller sufficed under Rule 702. The Circuit cryptically announced that “Obrey’s evidence was not rendered irrelevant under Rule 402 simply because it failed to account for the relative qualifications of the applicant pool.”  Obrey, 400 F.3d at 695.  Citing pre-Daubert decisions for the most part (such as Bazemore), the Ninth Circuit persuaded itself that Rule 702 requires nothing more than simple relevancy. Had the Circuit taken even a cursory look at Bazemore, it would have seen that the case involved a much more involved multiple regression than whatever statistical analysis Mr. Dannemiller propounded.  And the Ninth Circuit would have seen that even the Bazemore decision acknowledged that there may be

“some regressions so incomplete as to be inadmissible as irrelevant… .”

478 U.S. 385, 400 n.10 (1986). It is difficult to imagine a discrimination claim analysis more incomplete than one that did not address education, training, and experience.

Sadly, neither the Navy’s nor Mr. Obrey’s brief, 2004 WL 545873 (Feb. 4, 2004) provided any discussion of the nature, quality, findings, or limits of Mr. Dannemiller’s statistical analysis.  The Navy’s brief referred to Mr. Dannemiller as a “purported” expert.  His resume, available online, shows that Mr. Dannemiller studied history as an undergraduate, and has a master’s degree in sociology. He is the president of SMS Research, a consulting company.

The taxpayers deserved better advocacy from the Department of Justice, and greater attention to statistical methodology from its appellate judges.  See ATA Airlines, Inc. v. Federal Exp. Corp., 665 F.3d 882, 888-96 (2011) (Posner, J.) (calling for lawyers and judges to do better in understanding and explaining, in plain English, the statistical analyses that are essential to their cases). Judges at level need to pay greater attention to the precepts of Rule 702, even when there is no jury around to be snuckered.

A Black Swan Case – Bayesian Analysis on Medical Causation

March 15th, 2014

Last month, I posted about an article that Professor Greenland wrote several years ago about his experience as a plaintiffs’ expert witness in a fenfluramine case. “The Infrequency of Bayesian Analyses in Non-Forensic Court Decisions (Feb. 16, 2014).” Greenland chided a defense expert for having declared that Bayesian analyses are rarely or never used in analyzing clinical trials or in assessments of pharmaco-epidemiologic data.  Greenland’s accusation of ludicrousness appeared mostly to blow back on him, but his stridency for Bayesian analyses did raise the question, whether such analyses have ever moved beyond random-match probability analyses in forensic evidence (DNA, fingerprint, paternity, etc.) or in screening and profiling cases.  I searched Google Scholar and Westlaw for counter-examples and found none, but I did solicit references to “Black Swan” cases. Shortly after I posted about the infrequency of Bayesian analyses, I came across a website that was dedicated to collecting legal citations of cases in which Bayesian analyses were important, but this website appeared to confirm my initial research.

Some months ago, Professor Brian Baigrie, of the Jackman Humanities Institute, at the University of Toronto, invited me to attend a meeting of an Institute working group on The Reliability of Evidence in Science and the Law.  The Institute fosters interdisciplinary scholarship, and this particular working group has a mission statement close to my interests:

The object of this series of workshops is to formulate a clear set of markers governing the reliability of evidence in the life sciences. The notion of evidence is a staple in epistemology and the philosophy of science; the notion of this group will be the way the notion of ‘evidence’ is understood in scientific contexts, especially in the life sciences, and in judicial form as something that ensures the objectivity of scientific results and the institutions that produce these results.

The Reliability of Evidence in Science and the Law. The faculty on the working group represent disciplines of medicine (Andrew Baines), philosophy (James R. Brown, Brian Baigrie), and law (Helena Likwornik, Hamish Stewart), with graduate students in the environmental science (Amy Lemay), history & philosophy of science and technology (Karolyn Koestler, Gwyndaf Garbutt ), and computer science (Maya Kovats).

Coincidentally, in preparation for the meeting, Professor Baigrie sent me links to a Canadian case, Goodman v. Viljoen, which turned out to be a black swan case! The trial court’s decision, in this medical malpractice case focused mostly on a disputed claim of medical causation, in which the plaintiffs’ expert witnesses sponsored a Bayesian analysis of the available epidemiologic evidence; the defense experts maintained that causation was not shown, and they countered with the unreliability of the proffered Bayesian analysis. The trial court resolved the causation dispute in favor of the plaintiffs, and their witnesses’ Bayesian approach. Goodman v. Viljoen, 2011 ONSC 821 (CanLII), aff’d, 2012 ONCA 896 (CanLII).  The Court of Appeals’ affirmance was issued over a lengthy, thoughtful dissent. The Canadian Supreme Court denied leave to appeal.

Goodman was a medical practice case. Mrs. Goodman alleged that her obstetrician deviated from the standard of care by failing to prescribe corticosteroids sufficiently early in advance of delivery to avoid or diminish the risk of cerebral palsy in her twins.  Damages were stipulated, and the breach of duty turned on a claim that Mrs. Goodman, in distress, called her obstetrician.  Given the decade that passed between the event and the lawsuit, the obstetrician was unable to document a response.  Duty and breach were disputed, but were not the focus of the trial.

The medical causation claim, in Goodman, turned upon a claim that the phone call to the obstetrician should have led to an earlier admission to the hospital, and the administration of antenatal corticosteroids.  According to the plaintiffs, the corticosteroids would have, more probably than not, prevented the twins from developing cerebral palsy, or would have diminished the severity of their condition.  The plaintiffs’ expert witnesses relied upon studies that suggested a 40% reduction and risk, and a probabilistic argument that they could infer from this risk ratio that the plaintiffs’ condition would have been avoided.  The case thus raises the issue whether evidence of risk can substitute for evidence of causation.  The Canadian court held that risk sufficed, and it went further, contrary to the majority of courts in the United States, to hold that a 40% reduction in risk sufficed to satisfy the more-likely-than-not standard.  See, e.g., Samaan v. St. Joseph Hosp., 670 F.3d 21 (1st Cir. 2012) (excluding expert witness testimony based upon risk ratios too small to support opinion that failure to administer intravenous tissue plasminogen activator (t-PA) to a patient caused serious stroke sequelae); see also “Federal Rule of Evidence 702 Requires Perscrutations — Samaan v. St. Joseph Hospital (2012)” (Feb. 4, 2012).

The Goodman courts, including the dissenting justice on the Ontario Court of Appeals, wrestled with a range of issues that warrant further consideration.  Here are some that come to mind from my preliminary read of the opinions:

1. Does evidence of risk suffice to show causation in a particular case?

2. If evidence of risk can show causation in a particular case, are there requirements that the magnitude of risk be quantified and of a sufficient magnitude to support the inference of causation in a particular case?

3. The judges and lawyers spoke of scientific “proof.”  When, if ever, is it appropriate to speak of scientific proof of a medical causal association?

4. Did the judges incorrectly dichotomize legal and scientific standards of causation?

5. Did the judges, by rejecting the need for “conclusive proof,” fail to articulate a meaningful standard for scientific evidence in any context, including judicial contexts?

6. What exactly does the “the balance of probabilities” mean, especially in the face of non-quantitative evidence?

7. What is the relationship between “but for” and “substantial factor” standards of causation?

8. Can judges ever manage to define “statistical significance” correctly?

9. What is the role of “common sense” in drawing inferences by judges and expert witnesses in biological causal reasoning?  Is it really a matter of common sense that if a drug did not fully avert the onset of a disease, it would surely have led to a less severe case of the disease?

10. What is the difference between “effect size” and the measure of random or sampling error?

11. Is scientific certainty really a matter of being 95% certain, or is this just another manifestation of the transposition fallacy?

12. Are Bayesian analyses acceptable in judicial settings, and if so, what information about prior probabilities must be documented before posterior probabilities can be given by expert witnesses and accepted by courts?

13. Are secular or ecological trends sufficiently reliable data for expert witnesses to rely upon in court proceedings?

14. Is the ability to identify biological plausibility sufficient to excuse the lack of statistical significance and other factors that are typically needed to support the causality of a putative association?

15. What are the indicia of reliability of meta-analyses used in judicial proceedings?

16. Should courts give full citations to scientific articles that are heavily relied upon as part of the requirement that they publicly explain and justify their decisions?

These are some of the questions that come to mind from my first read of the Goodman case.  The trial judge attempted to explain her decision in a fairly lengthy opinion. Unfortunately, the two judges, of the Ontario Court of Appeals, who voted to affirm, did not write at length. Justice Doherty wrote a thoughtful dissent, but the Supreme Court denied leave to appeal.  Many of the issues are not fully understandable from the opinions, but I hope to be able to read the underlying testimony before commenting.

Thanks to Professor Baigrie for the reference to this case.

“Judges and other lawyers must learn how to deal with scientific evidence and inference.”

March 1st, 2014

Late last year, a panel of 7th Circuit reversed an Administrative Law Judge (ALJ) who had upheld a citation and fine against Caterpillar Logistics, Inc. (Cat).  The panel, in a wonderfully succinct, but meaty decision by Judge Easterbrook, wrote of the importance of judges’ and lawyers’ learning to deal with scientific and statistical evidence. Caterpillar Logistics, Inc. v. Perez, 737 F.3d 1117 (7th Cir. 2013)

Pseudonymous MK, a worker in Cat’s packing department, developed epidcondylitis (tennis elbow).  Id. at 1118. OSHA regulations require employers to report injuries  “the work environment either caused or contributed to the resulting condition”. 29 C.F.R. § 1904.5(a). MK’s work required her to remove items from containers and place items in shipping cartons. The work was repetitive, but MK acknowledged that the work involved little or no impact or force.  Apparently, Cat gave some rather careful consideration to whether MK’s epidcondylitis was work related; it assembled a panel of three specialists in musculoskeletal disorders and two generalists to consider the matter.  The panel, relying upon NIOSH and AMA guidelines, rejected MK’s claim of work relatedness.  Both the NIOSH and the AMA guidelines conclude that repetitive motion in the absence of weight or impact does not cause epicondylitis. Id.

MK called an expert witness, Dr. Robert Harrison, a clinical professor of medicine, at the University of California, San Francisco.  Id. at 1118-1119.  Harrison unequivocally attributed MK’s condition to her work at Cat, but he failed to explain why no one else in Cat’s packing department ever developed the condition.  Id. at 1119.

Harrison acknowledged that epidemiologic evidence could confirm his opinion, but he dismissed such evidence as being able to disconfirm his opinion.  The ALJ echoed Dr. Harrison in holding epidemiologic evidence to be irrelevant:

“none of these [other] people are [sic] MK. Similar to the concept of the ‘eggshell skull’ plaintiff in civil litigation, you take your workers as they are.”

Id. at 1119-20, citing ALJ, at 2012 OSAHRC LEXIS 118 at *32.

Judge Easterbrook found this attempt to disqualify any opposing evidence to lie beyond the pale:

“Judges and other lawyers must learn how to deal with scientific evidence and inference.”

Id. (citing Jackson v. Pollion, 733 F.3d 786 (7th Cir. 2013).

Judge Easterbrook called out the ALJ for misunderstanding the nature of epidemiology and the role of statistics, in the examination of causation of health outcomes that have a baseline incidence or prevalence in the population:

“The way to test whether Harrison is correct is to look at data from thousands of workers in hundreds of workplaces—or at least to look at data about hundreds of worker-years in Caterpillar’s own workplace. Any given worker may have idiosyncratic susceptibility, though there’s no evidence that MK does. But the antecedent question is whether Harrison’s framework is sound, and short of new discoveries about human physiology only statistical analysis will reveal the answer. Any large sample of workers will contain people with idiosyncratic susceptibilities; the Law of Large Numbers ensures that their experience is accounted for. If studies of large numbers of workers show that the incidence of epicondylitis on jobs that entail repetitive motion but not force is no higher than for people who do not work in jobs requiring repetitive motion, then Harrison’s view has been refuted.”

Id. at 1120.

Judge Easterbrook acknowledged that Cat’s workplace evidence may have been a sample too small from which to draw a valid statistical inference, given the low base rate of epicondylitis in the general population.  Dr. Harrison’s and the ALJ’s stubborn refusal, however, to consider any disconfirming evidence, obviating the need to consider sample size and statistical power issues.

Finally,  Judge Easterbrook chastised the ALJ for dismissing Cat’s experience as irrelevant because many other employers will not have sufficient workforces or record keeping to offer similar evidence.  In Judge Easterbrook’s words:

“This is irrational. If the camera in a police car captures the events of a highspeed chase, the judiciary would not ignore that video just because other police cars lack cameras; likewise, if the police record an interrogation, courts will consider that information rather than wait for the day when all interrogations are recorded.”

Id. This decision illustrates why some commentators at places such as the Center for Progressive Reform get their knickers in a knot over the prospect of applying the strictures of Rule 702 to agency fact finding; they know it will make a difference.

As for the “idiosyncratic gambit,” this argument is made all too frequently in tort cases, with similar lack of predicate.  Plaintiffs claim that there may be a genetic or epigenetic susceptibility in a very small subset of the population, and that epidemiologic studies may miss this small, sequestered risk.  Right, and the light in the refrigerator may stay on when you close the door.  Prove it!

The Infrequency of Bayesian Analyses in Non-Forensic Court Decisions

February 16th, 2014

Sander Greenland is a well-known statistician, and no stranger to the courtroom.  I first encountered him as a plaintiffs’ expert witness in the silicone gel breast implant litigation, where he testified for plaintiffs in front of a panel of court-appointed expert witnesses (Drs. Diamond, Hulka, Kerkvliet, and Tugwell).  Professor Greenland has testified for plaintiffs in vaccine, neurontin, fenfluramine, anti-depressant and other pharmaceutical cases.  Although usually on the losing side, Greenland has written engaging post-mortems of several litigations, to attempt to vindicate his positions he took, or deconstruct positions taken by adversary expert witnesses.

In one attempt to “correct the record,” Greenland criticized a defense expert witness for stating that Bayesian methods are rarely used in medicine or in the regulation of medicines. Sander Greenland, “The Need for Critical Appraisal of Expert Witnesses in Epidemiology and Statistics,” 39 Wake Forest Law Rev. 291, 306 (2004).  According to Greenland, his involvement as a plaintiff’s expert witness in a fenfluramine case allowed him to observe a senior professor in Yale University, who served as Wyeth’s statistics expert, make a “ludicrous claim,” id. (emphasis added), that

“the Bayesian method is essentially never used in the medical literature or in the regulatory environments (such as the FDA) for interpreting study results. . . .”

Id. (quoting from Supplemental Affidavit of Prof. Robert Makuch, App. Ex. 114, ¶5, in Smith v. Wyeth-Ayerst Labs., 278 F.Supp. 2d 684 (W.D.N.C. 2003)). Greenland criticizes Professor Makuch’s affidavit as “provid[ing] another disturbing case study of misleading expert testimony regarding current standards and practice.” 39 Wake Forest Law Rev. at 306.

“Ludicrous,” “disturbing,” “misleading,” and “demonstrably quite false”?  Really?

Greenland notes, as a matter of background, that many leading statisticians recommend and adopt Bayesian statistics.  Id. (citing works by Donald Berry, George Box, Bradley Carlin, Andrew Gelman James Berger, and others). Remarkably, however, Greenland failed to cite a single new or supplemental drug application, or even one FDA summary of safety or efficacy, or FDA post-market safety or efficacy review.  At the time Greenland was preparing his indictment, there really was little or no evidence of FDA’s embrace of Bayesian methodologies.  Six years later, in 2010, the agency did promulgate a guidance that set recommended practices for Bayesian analyses in medical device trials. FDA Guidance for the Use of Bayesian Statistics in Medical Device Clinical Trials (February 5, 2010); 75 Fed. Reg. 6209 (February 8, 2010); see also Laura A. Thompson, “Bayesian Methods for Making Inferences about Rare Diseases in Pediatric Populations” (2010); Greg Campbell, “Bayesian Statistics at the FDA: The Trailblazing Experience with Medical Devices” (Presentation give by Director, Division of Biostatistics Center for Devices and Radiological Health at Rutgers Biostatistics Day, April 3, 2009).  Even today, Bayesian analysis remains uncommon at the U.S. FDA.

Having ignored the regulatory arena, Greenland purported to do a study of the biomedical journals, “to check the expert’s claim in detail.” 39 Wake Forest Law Rev. at 306. Greenland searched on the word “Bayesian” in the Journal of Clinical Oncology for issues published from 1994-2003, and “found over fifty publications that contain the word in that journal alone.” Greenland does not tell us why he selected this one journal, which was not in the subject matter area of the litigation in which he was serving as a partisan expert witness.  For most the time surveyed, the Journal of Clinical Oncology published 24 issues a year, and occasional supplements. Most volumes contained over 4,000 pages per year.  Finding 50 uses of the word “Bayesian” in over 40,000 pages hardly constitutes resounding evidence to support his charges of “ludicrous,” “misleading,” “disturbing,” and “quite false.”  Greenland further tells us looking at these 50 or so articles “revealed several,” which “had used Bayesian methods to explore statistically nonsignificant results.” 39 Wake Forest Law Rev. at 306-07 & n.61 (citing only one paper, Lisa Licitra et al., Primary Chemotherapy in Resectable Oral Cavity Squamous Cell Cancer: A Randomized Controlled Trial, 21 J. Clin. Oncol. 327 (2003)). So in over 40,000 pages, Greenland found “several” Bayesian analyses, apparently post hoc looks to explore results that did not achieve pre-specified levels of statistical significance. Given the historical evolution of Bayesian analyses at FDA, and Greenland’s own evidence, the posterior odds that Greenland was correct in his charges seem to be disturbingly low.

Greenland tells us that the number of Bayesian analyses could be increased by looking at additional journals, and the Bayesian textbooks he cites.  No doubt this is true, as is his statement that respected statisticians, in prestigious journals, have called for Bayesian analyses to replace frequentist methods. Of course, increasing the scope of his survey, Greenland would be dramatically increasing the denominator of total journal papers with statistical methods.  Odds are that the frequency would remain very low.  Greenland’s empirical evidence hardly contradicts his bête noire for making the quoted purely descriptive statement about the infrequent use of Bayesian analysis in biomedical journals and in regulatory applications.

In lodging charges of ludicrousness, Greenland might have presented a more balanced view from more carefully conducted surveys of the biomedical literature, in the relevant time period.  See, e.g., J. Martin Bland & Douglas G. Altman, “Bayesians and frequentists,” 317 Brit. Med. J. 1151, 1151 (1998) (“almost all the statistical analyses which appear in the British Medical Journal are frequentist”); David S. Moore, “Bayes for Beginners? Some Reasons to Hesitate,” 51 The Am. Statistician 254, 254 (“Bayesian methods are relatively rarely used in practice”); J.D. Emerson & Graham Colditz, “Use of statistical analysis in the New England Journal of Medicine,” in John Bailar & Frederick Mosteler, eds., Medical Uses of Statistics 45 (1992) (surveying 115 original research studies for statistical methods used; no instances of Bayesian approaches counted); Douglas Altman, “Statistics in Medical Journals: Developments in the 1980s,” 10 Statistics in Medicine 1897 (1991); B.S. Everitt, “Statistics in Psychiatry,” 2 Statistical Science 107 (1987) (finding only one use of Bayesian methods in 441 papers with statistical methodology).

Perhaps the balance between frequentist and Bayesian analysis is shifting today, but when Professor Makuch made his affidavit in 2002 or so, he was clearly correct, factually and statistically.

In the legal arena, Bayesian analyses are frequently used in evaluating forensic claims about DNA, paternity, lead-isotopes, and other issues of identification.  Remarkably, Bayesian analyses play virtually no role in litigation of health effects claims, whether based upon medicines, or upon occupational or environmental exposures.  In searching Google scholar and Westlaw I found no cases outside of forensics. Citations to black-swan cases are welcomed.

The Historical Intersection of Law and Epidemiology: Miller v National Cabinet (NY Court of Appeals 1960)

January 3rd, 2014

The history of statistics, epidemiology, and products liability are intertwined in ways that call for greater attention.  The 1950s and 1960s witnessed increasingly sophisticated statistical approaches to epidemiologic evidence. Starting in 1950, and continuing throughout the 1950s, Sir Richard Doll and Sir Austin Bradford Hill began their epidemiologic exploration of lung cancer among smokers. See, e.g., Richard Doll & A. Bradford Hill, “Smoking and Carcinoma of the Lung,” 2 Br. Med. J. 739 (1950); Richard Doll & A. Bradford Hill, “The mortality of doctors in relation to their smoking habits; a preliminary report,” 1 Br. Med. J. 1451 (1954).  In 1955, Sir Richard Doll published his important paper that suggested an association between asbestosis and lung cancer.  Richard Doll, “Mortality from Lung Cancer in Asbestos Workers,”  12 Br. J. Indus. Med. 81 (1955).  No disparity between observed and expected rates of lung cancer was observed among workers without asbestosis. Measures of p-values were used to assess the strength of the evidence against a null hypothesis of no association. As important an advance as was Doll’s paper, and as careful an investigator as he was, it is remarkable that Doll neglected to consider the potential role of smoking in producing the excess lung cancer rates among the factory workers with asbestosis. 

Starting in the 1960s, Dr. Irving Selikoff began publishing his epidemiologic studies of American asbestos insulators. See, e.g., Irving J. Selikoff , Jacob Churg,  and E. Cuyler Hammond, “Asbestos exposure and neoplasia,” 188 J. Am. Med. Ass’n 22 (1964). Selikoff neglected to stratify his observational data by the presence or absence of clinical asbestosis (although his later studies suggested that there was a very high prevalence of asbestosis after 20 years from first employment).  In addition, these insulator studies used crude measures of smoking, which lumped the very rare non-smoking insulators in with those who “never smoked regularly.”

In 1965, Sir Austin Bradford Hill published his lecture to the Royal Society of Medicine, in which he gave a spirited defense of inferring causality from observational epidemiologic studies. Austin Bradford Hill, “The Environment and Disease: Association or Causation?” 58 Proc. Royal Soc’y Med. 295 (1965). Hill was justly proud of the success of observational epidemiology, at least for very large effect sizes that made residual confounding and bias unlikely.

The same years as Hill’s lecture, the American Law Institute published the Restatement (Second) of Torts, with its controversial Section 402A. Before 1965, employee-plaintiff lawsuits against remote suppliers of raw materials and products to employers were a rarity in American law.  Bradford Hill’s lecture on causal assessments of “clear-cut” statistical associations came just as epidemiologic, statistical evidence was working its way into tort cases involving smoking and lung cancer.  Not surprisingly, some of the first uses of epidemiologic evidence occurred in cases involving claims that tobacco caused lung cancer. See, e.g., Lartigue v. R.J. Reynolds Tobacco Co., 317 F.2d 19 (1963) (affirming defense verdict in case noted for plaintiffs’ use of epidemiologic evidence) (“The plaintiff contends that the jury’s verdict was contrary to the manifest weight of the evidence. … The jury had the benefit of chemical studies, epidemiological studies, reports of animal experiments, pathological evidence, reports of clinical observations, and the testimony of renowned doctors. The plaintiff made a convincing case, in general, for the causal connection between tobacco and cancer and, in particular, for the causal connection between Lartigue’s smoking and his cancer. The defendants made a convincing case for the lack of any causal connection.”), cert. denied, 375 U.S. 865 (1963), and cert. denied, 379 U.S. 869 (1964).

Epidemiologic and statistical evidence in tort cases has become a commonplace, even when it is distorted and abused by litigants and judges. Recent decisions involving claims that benzene caused various cancers are illustrative.  See, e.g., Milward v. Acuity Specialty Products Group, Inc., 664 F.Supp. 2d 137 (D. Mass. 2009) (granting motion to exclude opinions that substantially distorted epidemiologic evidence under the vague rubric of “weight of the evidence”), rev’d, 639 F.3d 11 (1st Cir. 2011) (closing off scrutiny of expert witness’s abuse of epidemiologic evidence in one of the most controversial, reactionary decisions involving federal gatekeeping decisions of recent years), cert. denied, U.S. Steel Corp. v. Milward, ___ U.S. ___, 2012 WL 33303 (2012). See also David E. Bernstein, “The Misbegotten Judicial Resistance to the Daubert Revolution,” 89 Notre Dame L. Rev. 27 (2013); “WOE-fully Inadequate Methodology – An Ipse Dixit By Another Name” (Sept. 2, 2011); “Milward — Unhinging the Courthouse Door to Dubious Scientific Evidence” (Sept. 2, 2011)

Given the problematic Milward decision, we might wonder what challenges to benzene-leukemia cases looked like just before Hill’s defense of inferring causality from observational studies began to infuse the witches’ brew of Rule 402A. What did statistical evidence look like before Hill’s paper?  In court cases, typically, statistical evidence was presented crudely or not at all.

In 1960, there was little opportunity to challenge causation opinions on admissibility grounds; rather sufficiency of the evidence to support a verdict or judgment was the primary means to gain review of an adverse decision.  Case reports of leukemia in workers very heavily exposed to benzene appeared in the 1920s, but it was not until the 1960s that analytical epidemiologic evidence (case-control and cohort studies) of association between leukemia and benzene were published. See generally Deborah Glass, Christopher Gray, Damien Jolley, Carl Gibbons, and Malcolm Sim, “The health watch case – control study of leukemia and benzene: the story so far,” 1076 Ann. N.Y. Acad. Sci. 80-89 (2006).  Thus, when the New York Court of Appeals decided a case involving a claim of benzene-induced leukemia, in 1960, the judicial decision was driven largely by the absence of specific quantification of risk of leukemia among workers occupationally exposed to benzene.[1]  Miller v. National Cabinet Co., 8 N.Y.2d 277, 281, 168 N.E.2d 811, 813, 204 N.Y.S.2d 129, 132, modified on other grounds, 8 N.Y.2d 1025, 70 N.E.2d 214, 206 N.Y.S.2d 795 (1960). The New York high court wrestled with the formalistic aspects of the expert witnesses’ testimony, including whether they expressed themselves in terms of “possibilities” or “probabilities.” Miller, 8 N.Y.2d at 284, 168 N.E.2d  at 814-15, 204 N.Y.S.2d at 134.

Focusing in one of the more knowledgeable of plaintiffs’ expert witnesses, Dr. Reznikoff, the Miller court was impressed that this witness disclaimed any intent to support an inference, from statistical analyses, to the plaintiff’s decedent.  Id. at 283. Furthermore, the court suggested that Reznikoff’s evidence might have been sufficient were it not for concession:

“I am sorry I can’t give you any statistics, but we don’t have them.”

Id. at 283.  The court appeared also to be concerned that Reznikoff’s approach provided no mechanistic insight or understanding into why many cases of  leukemia followed benzene exposure. Id.  

Reznikoff’s qualifications to speak to the subject were not dispositive of the question; the court was looking for data that were not available in the 1950s, when the case was tried:

“Not every supposition of a witness concerning what might be has the force of evidence, even though he has been licensed to practice medicine. If the witness is unfamiliar with any statistical data in the medical literature or in his own practice to give an inkling either to himself or to the court or board of how high the incidence of these cases is in situations of this kind, then the doctor’s assumption that it is ‘quite high’ is without significance. The lack of any kind of statistical data, which in the absence of scientific understanding is all that there would be to go on, is the more inexplicable if the claim is well founded in view of the large number of persons who die of leukemia and of workers in industry who are exposed to benzol. If there were any observed correlation between the two, it is certain that a physician of Dr. Reznikoff’s standing would be in possession of the information.”

Id. at 283-84. The court did not excuse the claimant’s evidentiary display with the indulgent, “this was the best evidence available,” when the evidence was inadequate.  Nor did the court engage in soothsaying that causality would someday be demonstrated.  We might feel some frustration today, looking back, that the court missed this opportunity, but case reports, and even case reports and epidemiologic studies, have generated many false-positive associations.  Clearly, more is required, and the New York Court of Appeals recognized the necessity for more.

Traumatic Cancers Distinguished

In 1960, the courts still indulged the proto-scientific opinion that traumatic injury caused cancer.[2]  Some medical writers supported this opinion, but by 1960, the opinion was already falling out of favor due to an improved understanding of carcinogenesis.

There is much irony, therefore, in the Miller court drawing’s an invidious distinction between Reznikoff’s proto-epidemiologic evidence and traumatic cancer cases that were still prevalent in the 1960s.  Id. at 285-86. In the traumatic cancer cases, in the 1960s, and even in the 1970s, courts sustained verdicts for cancer claimants who had shown that their cancers were diagnosed very shortly after a traumatic blow to the precise portion of the body where cancer manifested. The Miller court referred to these traumatic cancer cases as presenting the kind of causal inferences that could be understood and made by judges and juries.  Today, 40 years later, we see those causal inferences as mostly rubbish, based upon incorrect, inadequate, and discarded theories of carcinogenesis.

The prospect of cancer cases sustained by epidemiologic (statistical) evidence clearly troubled the New York court:

“The courts have been confronted before with cancer cases, and this is not likely to be the last. This is not an isolated situation. Questions of causation are common to actions based on warranty, tort or workmen’s compensation proceedings. Would, for example, evidence that there are 4 to 11 times as many cases of lung cancer among cigarette smokers as among nonsmokers be sufficient to establish a cause of action for breach of warranty in the sale of cigarettes? … There appear to be no decisions upholding causation in so complex a variety of the disease as leukemia. The cancer decisions in the courts where recovery has been allowed have dealt almost entirely with trauma, and there only in instances where the trauma occurred in the spot in the body where the pre-existing cancer was and the symptoms of its aggravation were immediately apparent … . In all of those cases the immediacy of the symptoms of aggravation of the cancer by a traumatic injury suffered in the area where the cancer was located was accepted as a substitute for scientific evidence or understanding of cause and effect. Absent that, damage claims of this nature have been dismissed on the law for lack of evidence of causation.”

Id. at 285-86 (internal citations omitted). The Miller court went on to note that New York law required that the cancer must develop at the exact location of the injury.  Furthermore, latency between the traumatic blow and the clinical recognition of the cancer was fatal to the claim, even in the face of opinion testimony that a plaintiff’s cancer was a “very slow growing” tumor.  Today, we understand that latency between first-exposure and clinical manifestation is necessitated by the length of induction periods and the doubling time of solid tumor cancers.  As a result of the Miller court’s reliance upon some dodgy notions of cancer causation, it held that Mr. Miller’s latency period disqualified the case from the immediate impact rule of traumatic cancer cases. Id. at 287-88 (distinguishing Hagy v. Allied Chemical & Dye Corp., 122 Cal. App. 2d 361, 265 P.2d 86 (1953), which involved a diagnosis of laryngeal cancer following immediately upon exposure to sulfuric acid mists).

The majority in Miller further expressed its concern that the understanding of cancer causation was marked by such uncertainty that the mere possibilities of chemical carcinogenesis should not be tolerated in this and similar cases:

“… [F]or so long as the causes of a disease — like cancer — are unknown to science, everyone contracting the disease could secure medical testimony that it is ‘possible’ that the disease is contracted from a wide variety of causes, choosing in each instance the particular possibility having the greatest promise of holding liable some responsible defendant. Any cancer expert could readily state that cancer could be caused by virus infection or by exposure to automobile exhaust fumes, sunlight, radiation, smog, smoking, hormone imbalance or according to any other theory which has been entertained by researchers or specialists as a possibility. Is a malpractice suit pending against some doctor who has given cortisone or ACTH as medicine? Then appears a medical witness who testifies that possibly cancer is caused by hormone imbalance induced thereby. Is it an action for breach of an implied warranty in the sale of cigarettes? Then the medical witness will testify that cigarettes could be a cause of lung cancer. Is it X ray or working in a garage where there have been exhaust fumes? Then the ‘possibility’ doctrine is adapted to creating questions of fact in those fields — and the same with benzol exposure and leukemia. Such a doctrine would overturn the rule that the burden is on the party asserting that a disease is based on actionable facts to prove causation. It would mean that, wherever such a cause is possible, the burden rests on the opposite party to prove that the disease resulted from something else. Consequently, for so long as the causes of the disease are unknown to medical science, the claimant or plaintiff can always recover — if the trier of the fact is favorably disposed — since no one can prove that the disease had other causes. This is a perversion of the normal rule that the disease must have resulted from the occupation and that the burden of proving causation is upon the party asserting it. The law does not intend that the less that is known about a disease the greater shall be the opportunity of recovery in court.”

Id. at 289.

The Miller decision provoked a dissent, mostly on formalistic grounds.  Id. at 290. The dissenting judges asserted, without much analysis, that there was substantial evidence to support causality. Given that qualified expert witnesses showed up for the claimant seemed sufficient on this score for the dissenters.  To the extent that the claimant’s expert witnesses expressed themselves in terms of possibilities, the dissenters opined that possibilities are sufficient, especially in the context of workman’s compensation cases, in which the burden of proof standards are lower than in common law civil liability cases.

The majority opinion stands as an eloquent expression of concern about the need for quantitative evidence of statistical risk in chemical exposure cancer cases. The court also presciently saw what would become a plague of litigation involving claims of cancer causation.  In 1960, for benzene and leukemia, the evidence was clearly, even by the standards of the day, inadequate, and the claimant’s expert witnesses were appropriately modest about what inferences could be drawn both with respect to general and specific causation.  The 1970s would witness a growing immodesty among available expert witnesses, as well as an explosive growth in the techniques and applications of analytical epidemiology to many problems, including the relationship between benzene and leukemia.



[1] A decade or two later, the scientific community recognized high levels of exposure to benzene as a cause of certain kinds of leukemia, by virtue of epidemiologic studies. See, e.g., Fusun Yaris, Mustafa Dikici, Turhan Akbulut, Ersin Yaris, Hilmi Sabuncu, “Story of benzene and leukemia: epidemiologic approach of Muzaffer Aksoy,” 46 J. Occup. Health 244 (2004); Abdul Khalade, Maritta S Jaakkola, Eero Pukkala and Jouni JK Jaakkola, “Exposure to benzene at work and the risk of leukemia: a systematic review and meta-analysis,” 9 Envt’l Health 31 (2010).  See also Michael D. Green, The Paradox of Statutes of Limitations in Toxic Substances Litigation, 76 Cal. L. Rev. 965, 974 (1988).

[2] William B. Coley & Norman L. Higinbotham, “Injury as a Causative Factor in the Development of Malignant Tumors,” 98 Annals of Surgery 991 (1933); Shields Warren, “Minimal Criteria Required to Prove Causation of Traumatic or Occupational Neoplasms,” 117 Annals of Surgery 585 (April 1943); Shields Warren, “Criteria Required to Prove Causation of Occupational or Traumatic Tumors,” 10 U. Chi. L. Rev. 313, 318-20 (1943); Russell & Clark, “Medico-Legal Considerations of Trauma and Other External Influences in Relationship to Cancer,” 6 Vand. L. Rev. 868, 875 (1953); Arden R. Hedge, “Can a Single Injury Cause Cancer?” 90 California Medicine 55 (1959); Auster, “The Role of Trauma in Oncogenesis: A Juridical Consideration,” 175 J. Am. Med. Ass’n 940, 949 (1961); Comment, “Sufficiency of Proof in Traumatic Cancer Cases,” 46 Cornell L.Q. 581, 581-82 (1961); Comment, “Sufficiency of Proof in Traumatic Cancer: A Medico-Legal Quandary,” 16 Arkansas L. Rev. 243, 256 67 (1962); Dyke, “Traumatic Cancer,” 15 Clev.Mar. L. Rev 472, 484-94 (1977).  See also Comment, “Judicial Attitudes Towards Legal and Scientific Proof of Cancer Causation,” 3 Columbia J. Envt’l L. 344, 354-68 (1977).

 

The Seventh Circuit Regresses on Rule 702

October 29th, 2013

Earlier this month, a panel of the Seventh Circuit of the United States Court of Appeal decided a relatively straight forward case by reversing the trial court’s exclusion of a forensic accountant’s damages calculation.  Manpower, Inc. v. Insurance Company of the State of Pennsylvania, No. 12‐2688 (7th Cir. Oct. 16, 2013).  In reversing, the appellate court disregarded a congressional statute, Supreme Court precedent, and Circuit decisional law.

The case involved a dispute over insurance coverage dispute and an economic assessment of Manpower, Inc.’s economic losses that followed a building collapse.  The trial court excluded Manpower’s accounting expert witness, Sullivan, who projected a growth rate (7.76%) for the plaintiff by comparing total revenues for a five month period in 2006 to the same five months in the previous year.  Id. at 8.  The historical performance, however, included a negative annual growth rate of 4.79% , over the years 2003 to 2009.  Over the five months immediately preceding Sullivan’s chosen period in 2006, the growth rate was merely 3.8%, less than half his projected growth rate.  Id.  Sullivan tried to justify his rather his extreme selectivity in data reliance by adverting to information that he obtained from the company about its having initiated new policies and installed new managers by the end of 2005.  Id.

The trial court held that Sullivan, who was not an expert on business management, had uncritically accepted the claimant’s proffered explanation for a very short-term swing in profitability and revenue.  Id. at 9.  While suggesting that Sullivan’s opinion was not “bulletproof,” the panel of the Seventh Circuit reversed.  The panel, which should have been reviewing the district court for potential “abuse of discretion,” appears to have made its own independent determination that Sullivan opinion was “sufficiently reliable to present to a jury.” Id. at 17.  In reversing, the panel explained that “the district court exercised its gatekeeping role under Daubert with too much vigor.” Id.

The panel attempted to justify its reversal by suggesting that a district court “usurps the role of the jury, and therefore abuses its discretion, if it unduly scrutinizes the quality of the expert’s data and conclusions rather than the reliability of the methodology the expert employed.” Id. at 18.  The panel’s reversal illustrates several methodological and legal confusions that make this case noteworthy beyond its mundane subject matter.

Of course, the most striking error in the panel’s approach is citing to a Supreme Court case, Daubert, which has been effectively superseded by a Congressional statute, Federal Rule of Evidence 702, in 2000:

“A witness who is qualified as an expert … may testify in the form of an opinion or otherwise if:

(a) the expert’s scientific, technical, or other specialized knowledge will help the trier of fact to understand the evidence or to determine a fact in issue;

(b) the testimony is based on sufficient facts or data;

(c) the testimony is the product of reliable principles and methods; and

(d) the expert has reliably applied the principles and methods to the facts of the case.”

Pub. L. 93–595, § 1, Jan. 2, 1975, 88 Stat. 1937; Apr. 17, 2000 (eff. Dec. 1, 2000); Apr. 26, 2011, eff. Dec. 1, 2011.)  Ironically, the Supreme Court’s Daubert case itself, had the Manpower panel paid attention to it, reversed the Ninth Circuit for applying a standard, the so-called Frye test, which predated the adoption of the Federal Rules of Evidence in 1975.  Rather than following the holding of the Daubert case, the panel got mired down in its dicta about a distinction between methodology and conclusion.  The Supreme Court itself abandoned his distinction a few years later in General Electric Co. v. Joiner, when it noted that

“conclusions and methodology are not entirely distinct from one another.”

522 U.S. 136, 146 (1997).

The panel of the Seventh Circuit concluded, without much real analysis, that the district court had excluded Sullivan’s opinions on a basis that implicated his conclusion and data selection, not his methodology.  Id. at 19-20.  The problem, of course, is that how one selects data of past performance to project future performance is part and parcel of the methodology of making the economic projection.  The supposed distinction advanced by the panel is illusory, and contrary to post-Daubert decisions, and the Congressional revision of the statute, which requires attention to whether “the testimony is based on sufficient facts or data; the testimony is the product of reliable principles and methods; and, the expert has reliably applied the principles and methods to the facts of the case.” Rule 702.

To make matters worse, the appellate court in Manpower proceeded to attempt to justify its reversal on grounds of “[t]he latitude we afford to statisticians employing regression analysis, a proven statistical methodology used in a wide variety of contexts.” Id. at 21. Here the appellate court suggests that if expert witnesses use a statistical test or analysis, such as regression analysis, it does not matter how badly they apply the test, or how worthless their included data are.  Id. at 22.  According to the Manpower panel:

“the Supreme Court and this Circuit have confirmed on a number of occasions that the selection of the variables to include in a regression analysis is normally a question that goes to the probative weight of the analysis rather than to its admissibility. See, e.g.,Bazemore v. Friday, 478 U.S. 385, 400 (1986) (reversing lower court’s exclusion of regression analysis based on its view that the analysis did not include proper selection of variables); Cullen v. Indiana Univ. Bd. of Trustees, 338 F.3d 693, 701‐02 & n.4 (7th Cir. 2003) (citing Bazemore in rejecting challenge to expert based on omission of variables in regression analysis); In re High Fructose Corn Syrup Antitrust Litigation, 295 F.3d 651, 660‐61 (7th Cir. 2002) (detailing arguments of counsel about omission of variables and other flaws in application of the parties’ respective regression analyses and declining to exclude analyses on that basis); Adams v. Ameritech Servs., Inc., 231 F.3d 414, 423 (7th Cir. 2000) (citing Bazemore in affirming use of statistical analysis based solely on correlations—in other words, on a statistical comparison that employed no regression analysis of any independent variables at all). These precedents teach that arguments about how the selection of data inputs affect the merits of the conclusions produced by an accepted methodology should normally be left to the jury.”

Id. at 22.

Again, the Seventh Circuit’s approach in Manpower is misguided. Bazemore involved a multivariate regression analysis in the context of a discrimination case.  Neither the Supreme Court nor the Fourth Circuit considered the regression at issue in Bazemore as evidence; rather the analysis was focused upon whether, within the framework of discrimination law, the plaintiffs’ regression satisfied their burden of establishing a prima facie case that shifted the burden to the defendant. No admissibility challenge was made to the regression in Bazemore under Rule 702.  Of course, the Bazemore litigation predates the Supreme Court’s decision in Daubert by several years.  Furthermore, even the Bazemore decision acknowledged that there may be

“some regressions so incomplete as to be inadmissible as irrelevant… .”

478 U.S. 385, 400 n.10 (1986).

The need for quantitative analysis of race and other suspect class discrimination under the equal protection clause no doubt led the Supreme Court, and subsequent lower courts to avoid looking too closely at regression analyses.  Some courts, such as the Manpower panel view Bazemore as excluding regression analysis from gatekeeping of statistical evidence, which magically survives Daubert. The better reasoned cases, however, even within the Seventh Circuit fully apply the principles of Rule 702 to statistical inference and analyses. See, e.g., ATA Airlines, Inc. v. Fed. Express Corp., 665 F.3d 882, 888–89 (2011) (Posner, J.) (reversing on grounds that plaintiff’s regression analysis should never have been admitted), cert. denied, 2012 WL 189940 (Oct. 7, 2012); Zenith Elecs. Corp. v. WH-TV Broad. Corp., 395 F.3d 416 (7th Cir.) (affirming exclusion of expert witness opinion whose extrapolations were mere “ipse dixit”), cert. denied, 125 S. Ct. 2978 (2005); Sheehan v. Daily Racing Form, Inc. 104 F.3d 940 (7th Cir. 1997) (Posner, J.) (discussing specification error).  See also Munoz v. Orr, 200 F.3d 291 (5th Cir. 2000).  For a more enlightened and educated view of regression and the scope and application of Rule 702, from another Seventh Circuit panel, Judge Posner’s decision in ATA Airlines, supra, is an essential starting place. SeeJudge Posner’s Digression on Regression” (April 6, 2012).

There is yet one more flaw in the Manpower decision and its rejection of the relevancy of data quality for judicial gatekeeping.  Federal Rule of Evidence 703 specifically addresses the bases of an expert witness’s opinion testimony.  The Rule, in relevant part, provides that:

“If experts in the particular field would reasonably rely on those kinds of facts or data in forming an opinion on the subject, they need not be admissible for the opinion to be admitted.”

Here the district court had acted prudently in excluding an expert witness who accepted the assertions of new management that it had, within a very short time span, turned a company from a money loser into a money earner.  As any observer of the market knows, there are too many short-term “fixes,” such as cutting personnel, selling depreciated property, and the like, to accredit any such short-term data as “reasonably relied upon.”  See In re Agent Orange Product Liability Lit., 611 F. Supp. 1223, 1246 (E.D.N.Y. 1985) (excluding opinions under Rule 703 of proffered expert witnesses who relied upon checklists of symptoms prepared by the litigants; “no reputable physician relies on hearsay checklists by litigants to reach a conclusion with respect to the cause of their affliction”), aff’d on other grounds, 818 F.2d 187 (2d Cir. 1987), cert. denied, 487 U.S. 1234 (1988).

Manpower represents yet another example of Court of Appeals abrogating gatekeeping by reversing a district judge who attempted to apply the Rules and the relevant Supreme Court precedent.  The panel in Manpower ignored Congressional statutory enactments and precedents of its own Circuit, and it relied upon cases superseded and overruled by later Supreme Court cases.  That’s regression for you.

Urging Review and Reversal, Scientists File Amicus Brief in the Harkonen Case

September 7th, 2013

Earlier this week, Professors Kenneth Rothman and Timothy Lash, and I, filed our Brief by Scientists And Academics as Amici Curiae, in the case, Harkonen v. United States.  As noted previously, Dr. Harkonen has petitioned the Supreme Court for review of the Ninth Circuit’s affirmance of his conviction for Wire Fraud.  Other amici will likely file on Monday, September 9, 2013.

Aaron Kesselheim’s Presentation on FDA Regulation of Manufacturer Speech

September 3rd, 2013

On August 5, 2013, Dr. Scott Harkonen filed his petition for a writ of certiorari with the United States Supreme Court. As noted in some previous posts, Dr. Harkonen was acquitted of misbranding, but convicted of wire fraud, for his role in issuing a press release about the results of a clinical trial of interferon gamma 1b, in patients with idiopathic pulmonary fibrosis.  (See Multiplicity versus Duplicity – The Harkonen Conviction; The Matrixx Motion in U.S. v. Harkonen; The (Clinical) Trial by Franz Kafka).

Dr. Harkonen’s petition presents two questions:

“1. Whether a conclusion about the meaning of scientific data, one on which scientists may reasonably disagree, satisfies the element of a “false or fraudulent” statement under the wire fraud statute, 18 U.S.C. § 1343?

2. Whether applying 18 U.S.C. § 1343 to scientific conclusions drawn from accurate data violates the First Amendment’s proscription against viewpoint discrimination, or renders the statute, as applied , unconstitutionally vague.”

Both questions are important given that the government has conceded that Dr. Harkonen’s press release accurately presented the raw data and calculated p-values.  The crime, if crime it be, lay in Dr. Harkonen’s drawing a causal inference from a subgroup, p = 0.004, which was not prespecified, in a specified secondary endpoint of survival (p = 0.08), when the subgroup was clearly based upon the goals of the trials, and there was other corroborative evidence in the form of two previous trials, clinical practice, and strong mechanistic evidence.

The government argued that NO inferences could be drawn from a trial that “failed” on its primary endpoint.  The government’s embrace of this statistical orthodoxy greatly misrepresented scientific practice to the courts below.  The only “failed” trial is one that is not conducted.

There are many who would go to great lengths to distort the facts of the Harkonen case in order to demonize the pharmaceutical industry, or to arm the Justice Department with a weapon that can shut down scientific speech about pharmaceutical interventions.  The expansion of the Wire Fraud Act, seen in the Harkonen case, to achieve these political goals will not only affect pharmaceutical company scientists, but also government and academic scientists.  The standard for falsity, drawn from an out-dated, tendentious, and overly rigid conception of hypothesis testing will apply equally to non-industry scientists in False Claim Act cases.  Perhaps in future posts, I can provide some good examples, on condition that any qui tam relators share their bounty with me.

Back in May, Aaron Kesselheim presented (by video) a paper, written with Michelle Mello, of the Harvard School of Public Health, on “The Prospect of Continued FDA Regulation of Manufacturer Promotion in an Era of Expanding Commercial Speech.”  Kesselheim went out of his way to misrepresent the facts of the Harkonen case, as part of his brief against off-market promotion.

By way of background, Aaron S. Kesselheim is a physician and a lawyer, and an Assistant Professor of Medicine at Harvard Medical School.  He is also a faculty member in the Division of Pharmacoepidemiology and Pharmacoeconomics in the Department of Medicine at Brigham and Women’s Hospital.   Given his position and his training in two professions, as well as the extraordinary stakes involved in allowing the government to prosecute scientists for drawing allegedly false conclusions about facts that the government concedes are accurate, Dr. Kesselheim should have exercised much greater care in checking his own assertions more closely.

Dr. Kesselheim focused primarily on the Second Circuit’s recent decision in United States v. Caronia, 703 F.3d 149 (2d Cir. 2012) , which reversed a judgment of conviction for off-label promotion, on First Amendment grounds.  About nine minutes into his presentation, Kesselheim turned to alternative strategies for the government to use to squelch off-label promotion.  One of his suggestions was to follow the model of the Harkonen prosecution, and to prosecute off-label promotion as false and misleading speech.

In his discussion of his suggested strategy, Kesselheim suggested that Dr. Harkonen had made misleading “conclusory, unsubstantiated claims for efficacy,“ and “without reference to supporting evidence.”  It is Kesselheim, however, who seriously mislead his listeners and readers by stating that Dr. Harkonen had made “conclusory, unsubstantiated claims for efficacy.”  The Press Release that was the subject of the government’s indictment set out accurately actual count data and calculated p-values.  No data were fabricated or falsified.  Within the limited space and the informal context of a Press Release, Dr. Harkonen had provided a substantial account of the data from InterMune’s clinical trial, as well as citing a previous, independent clinical trial and its extension, clinical experience, and mechanism research on the action of interferon γ-1b.  Unfortunately, it is Kesselheim who is speaking in conclusory sound bites when he ignores the context and content of the actual Press Release at issue.

Kesselheim went on to suggest that Harkonen’s statement was refuted by a “company-sponsored clinical trial showing that the drug was not effective.” This statement is not only false, but shows a flagrant disregard for statistical analysis and the data in the Harkonen case.  Kesselheim implies that a clinical trial that fails to show treatment efficacy thereby shows that the treatment was not effective.  His statement commits the fundamental error of equating a failure to reject the null hypothesis at a specified level of attained significance with acceptance of the null hypothesis.  This reasoning is fallacious and fundamentally flawed.

To be sure, the prespecified secondary survival endpoint in InterMune’s clinical trial did not meet the 0.05 cutoff (it was 0.08), although the per-protocol analysis for this endpoint came up at 0.055, on a preliminary analysis of the data. When the clinical trial was fully analyzed and written up for publication in the New England Journal of Medicine, the treatment-adherent analysis for survival in the entire clinical trial was 0.02, with a statistically significant hazard ratio for survival, favoring the therapy:

“Analysis of the treatment-adherent cohort of patients showed an absolute reduction in the risk of death of 9 percent in the interferon gamma-1b group, as compared with the placebo group, and a relative reduction in the risk of 66 percent (5 percent of 126 patients in the interferon gamma-1b group and 14 percent of 143 patients in the placebo group died, P=0.02). The hazard ratio for death in the interferon gamma-1b group, as compared with the placebo group, was 0.3 (95 percent confidence interval, 0.1 to 0.9).”

Ganesh Raghu, Kevin K. Brown, Williamson Z. Bradford, Karen Starko, Paul W. Noble, David A. Schwartz, and Talmadge E. King, Jr., for the Idiopathic Pulmonary Fibrosis Study Group, “A Placebo-Controlled Trial of Interferon Gamma-1b in Patients with Idiopathic Pulmonary Fibrosis,” 350 New Engl. Med. J. 125, 129-30 (2004).

Dr. Harkonen, in his Press Release, did focus on what seems like an eminently sensible subgroup, within the survival secondary endpoint, of mild- and moderate-cases, which, a priori, were believed to be the patients mostly likely to benefit from the interferon γ-1b therapy.  (What was not known before the trial was at what point in disease progression might patients no longer respond with greater survival, and hence the difficulty in setting the boundary between moderate and severe cases.)  Kesselheim might argue that the interferon γ-1b clinical trial, standing alone, was inconclusive, but he certainly cannot argue truthfully that the trial showed that the biological product to be ineffective.  Clinical trials do not neatly divide the world of possible results into demonstrations of efficacy and demonstrations of inefficacy.  Not only does the evidence come in degrees, but there is a range of “inconclusiveness” in between the two extremes. Given his background, training, and experience, Kesselheim certainly should know this, and he should apologize for his inaccurate statements.

Kesselheim might well have stopped there, but he went on to acknowledge that the company-sponsored clinical trial at issue did find, in post-hoc analyses, a non-significant trend of benefit in a subset of patients.  Talk of misleading speech!  The p-value at issue was 0.004, uncorrected for multiple comparisons, but no one, not Kesslheim, not the government or anyone else, has offered any appropriate adjustment for multiple comparisons that would inflate that 0.004 to over 0.05.  Kesselheim has no warrant for branding the subgroup finding “non-significant,” until he shows that the p = 0.004, when appropriately modified (if it can be), exceeds 0.05.

Kesselheim mangles other, less technical facts.  He claims that the company saw a ten-fold increase in sales of interferon γ-1b for idiopathic pulmonary fibrosis.  No such fact was ever, or could ever, be established in the Harkonen case.  Kesselheim claims that Dr. Harkonen admitted, in emails, that he did not really believe that the trial “demonstrated” benefit; no such emails were ever adduced at trial, and this seems to be part of a fictional narrative that Dr. Kesselheim has manufactured.  Finally, Kesselheim harrumphs that FDA declined to approve drug.  The company never filed a new drug application for the idiopathic pulmonary fibrosis indication; there was no application to reject.  Perhaps more important is that the Press Release was issued before InterMune had made any formal submission of data to the FDA, an event that did not take place until the following year.

Kesselheim sighs that the Harkonen prosecution will be a difficult act to follow because it requires a case-by-case showing of falsity, with the necessity of expert testimony, and heavy cognitive demands on lay jurors. How ironic that Kesselheim, a lawyer and a physician, and a Harvard Medical School faculty member, buckled under the cognitive demands of his topic. Indeed, Kesselheim’s confusion is a strong argument for why the Supreme Court should put a stop to the practice of asking jurors to second guess whether a scientist has incorrectly inferred causation from accurately presented facts.

Let’s hope Dr. Harkonen gets a fair hearing in the Supreme Court.

Woodside & Davis on the Bradford Hill Considerations

August 23rd, 2013

Dr. Frank Woodside and Allison Davis have published an article on the so-called Bradford Hill criteria.  Frank C. Woodside, III & Allison G. Davis, “The Bradford Hill Criteria: The Forgotten Predicate,” 35 Thomas Jefferson L. Rev. 103 (2013).

Their short paper may be of interest to Rule 702 geeks, and students of how the law parses causal factors in litigation.

The authors argue that a “predicate” to applying the Hill criteria consists of:

  • ascertaining a clear-cut association,
  • determining the studies establishing the association are valid, and
  • satisfying the Daubert [1][sic] requirements.

Id. at 107.  Parties contending for a causal association often try to flyblow the need for statistical significance at any level, and argue that Bradford Hill did not insist upon statistical testing.  Woodside and Davis remind us that Bradford Hill was quite firm in insisting upon the need to rule out random variability as an explanation for an association:

“Our observations reveal an association between two variables, perfectly clear-cut and beyond what we would care to attribute to the play of chance.”

Id. at 105; see Hill, Austin Bradford Hill, “The Environment and Disease: Association or Causation?” 58 Proc. Royal Soc’y Med. 295 (1965).  The authors correctly note that the need for study validity is fairly implied by Bradford Hill’s casual expression about “perfectly clear-cut.”

Woodside and Davis appear to acquiesce in the plaintiffs’ tortured interpretation of Bradford Hill’s speech, on which statistical significance supposedly is unimportant.  Woodside & Davis at 105 & n.7 (suggesting that Bradford Hill “seemingly negates the second [the requirement of statistical significance] when he discounts the value of significance testing, citing Bradford Hill at 299).

Woodside and Davis, however, miss the heavy emphasis that Bradford Hill actually placed upon “tests of significance”:

“No formal tests of significance can answer those questions. Such tests can, and should, remind us of the effects that the play of chance can create, and they will instruct us in the likely magnitude of those effects. Beyond that they contribute nothing to the ‘proof’ of our hypothesis.”

Bradford Hill at 299.  Bradford Hill never says that statistical tests contribute nothing to proving an hypothesis; rather, his emphasis is on the insufficiency of statistical tests alone to establish causality.  Bradford Hill’s “beyond that” language clearly stakes out the preliminary, but necessary importance of ruling out the play of chance before proceeding to consider the causal factors.

Passing beyond their exegetical fumble, Woodside and Davis proceed to discuss the individual Bradford Hill considerations and how they have fared in the crucible of Rule 702.  Their discussion may be helpful to lawyers who want to track the individual considerations, and how they have treated, or dismissed, by trial courts charged with gatekeeping expert witness opinion testimony.

There is another serious problem in the Woodside and Davis paper.  The authors describe risk ratios and the notion of “confidence intervals”:

“A confidence interval provides both the relative risk found in the study and a range (interval) within which the risk would likely fall if the study were repeated numerous times.32 … As such, risk measures used in conjunction with confidence intervals are critical in establishing a perfectly clear-cut association when it comes to examining the results of a single study.35

Woodside & Davis at 110.  The authors cite to the Reference Manual on Scientific Evidence (3d 2011), but they fail to catch important nuances of the definition of a confidence interval.  The obtained interval from a given study is not the interval within which the “risk would likely fall if the study were repeated… .”  Rather it is 95% of the many intervals, from the many repeated studies done on the same population, with the same sample size, which would capture the true risk.  As for the obtained interval, the true risk is either within it, or not, and no probability value attaches to the likelihood that the true value lies within the obtained interval.

It is a mystery why lawyers would bother to define something like the confidence interval, and then do it incorrectly.  Here is how Professors Finkelstein and Levin define the confidence interval in their textbook on statistics:

“A confidence interval for a population proportion P is a range of values around the proportion observed in a sample with the property that no value in the interval would be considered unacceptable as a possible value for P in light of the sample data.”

Michael Finkelstein & Bruce Levin, Statistics for Lawyers 166-67 (2d ed. 2001).   This text explains why and where Woodside and Davis went astray:

“It is the confidence limits PL and PU that are random variables based on the sample data. Thus, a confidence interval (PL, PU) is a random interval, which may or may not contain the population parameter P. The term “confidence” derives from the fundamental property that, whatever the true value of P, the 95% confidence interval will contain P within its limits 95% of the time, or with 95% probability. This statement is made only with reference to the general property of confidence intervals and not to a probabilistic evaluation of its truth in any particular instance with realized values of PL and PU.”

Id. at 167-71.


[1] Surely the time has come to stop referring to the Daubert factors and acknowledge that the Daubert case was just one small step in the maturation of evidence law.  The maturation consisted of three additional Supreme Court cases, many lower court cases, and a statutory revision to Federal Rule of Evidence 702, in 2000.  The Daubert factors hardly give due consideration to the depth and breadth of the law in this area.

Securities Fraud vs Wire Fraud

July 29th, 2013

Pharmaceutical manufacturers are particularly vulnerable to securities fraud claims arising from the manufacturers’ pronouncements about safety or efficacy, the evidence for which is often statistical in nature.  Safety claims may involve complex data sets, both from observational studies and clinical trials.  Efficacy claims are typically based upon clinical trial data.

Publicly traded manufacturers may find themselves caught between competing securities regulations.  In evaluating safety or efficacy data, manufacturers will often consult with an outside science advisory board, or report to regulatory agencies.  Securities regulations specify that any disclosure of confidential inside information to an outsider triggers an obligation of prompt public disclosure of that information.[1]  Companies also routinely seek to keep investors informed of research and marketing developments.  Generally, manufacturers will make their public disclosures through widely circulated press releases.[2]  Not surprisingly, disgruntled investors may challenge the accuracy of the press releases, when the product or drug turns out to be less efficacious or more harmful than represented in the press release.  These challenges, brought under the securities laws, often are maintained in parallel to product liability actions, sometimes in the same multi-district litigation.

Securities laws require accurate disclosure of all material information.[3]  Rule 10b-5 of the Securities Exchange Commission (SEC) prohibits any person from making “any untrue statement of material fact” or from omitting “a material fact necessary in order to make the statements made, in light of the circumstances under which they were made, not misleading.”[4]

A prima facie case of securities fraud requires that plaintiff allege and establish, among other things, a material misrepresentation or omission.[5]  The obligations to speak and to speak accurately have opened manufacturers to second guessing in their analyses of safety and efficacy data.  In most securities fraud cases, courts have given manufacturers a wide berth by rejecting differences in opinions about the proper interpretation of studies as demonstrating fraud under the securities regulations.[6]  This latitude has been given both in judgment of what test procedures to use, as well as in how best to interpret data.[7]   In Padnes v. Scios Nova Inc., the manufacturer was testing a drug for treatment of acute kidney failure.  Scios Nova issued a press release after its phase II trial, to announce a statistically significant reduction in patients’ need for dialysis.  When the early phase III results failed to confirm this result, plaintiffs sued Scios Nova for not disclosing the lack of statistically significant outcomes on other measures of kidney function, as well as for its interpretation of dialysis results as statistically significant.[8]  The trial court dismissed the complaint.[9]

Several securities fraud cases have turned on investor dissatisfaction on how companies interpreted clinical trial subgroup data.  In Noble Asset Management v. Allos Therapeutics, Inc.,[10] the company issued a press release, noting no statistically significant increase overall in survival advantage from a drug for breast cancer, but also noting a statistically significant increased survival in a non-prespecified subgroup of patients with metastatic breast cancer.[11] The plaintiff investors claimed that the company should have disclosed that the FDA was unlikely to approve an indication based upon an ad hoc subgroup analysis, but the trial court rejected this claim because FDA policy on drug approvals is public and well known.[12] The plaintiffs also complained that the press release referred to statistically significant results from a Cox multiple regression analysis rather than the log-rank (non-parametric survival) analysis required by FDA. The trial court rejected this claim as well, opining that the analysis was not misleading when the company correctly reported the raw data and the results of the Cox multiple regression analysis.[13]

Two recent appellate decisions emphasize the courts’ unwillingness to scrutinize the contested statistical methodology that underlies plaintiffs’ claims of misrepresentation.  In In re Rigel Pharmaceuticals, Inc. Securities Litigation, the plaintiff investors were dissatisfied, not with reporting of subgroups, but rather with the failure of the company to report geographic subgroup results, as well as its use of allegedly improper statistical tests and its failure to account for multiple comparisons.[14]

The Ninth Circuit affirmed the dismissal of a complaint.  The appellate court held that allegations of “statistically false p-values” were not sufficient; plaintiffs must allege facts that explain why the difference between two statements “is not merely the difference between two permissible judgments, but rather the results of a falsehood.”[15] Alleging that a company should have used a different statistical method to analyze the data from its clinical trial is not sufficient to raise an issue of factual falsity under the securities fraud statute and regulations.[16]  The Court explained that the burden was on plaintiffs to plead and prove that the difference between two statistical statements “is not merely the difference between two permissible judgments, but rather the result of a falsehood.”[17] The Court characterized the plaintiffs’ allegations to be about judgments of which statistical tests or methods are appropriate, and not about false statements.  Furthermore, the Court emphasized that the company’s statistical method was called for in the trial protocol, and was selected before the data were unblinded and provided to the company.[18]

In Kleinman v. Elan Corporation[19], the Second Circuit affirmed the dismissal of a securities fraud class action against two pharmaceutical joint venturers, which issued a challenged press release on interim phase II clinical trial results for bapineuzumab, a drug for mild- to moderate-Alzheimer’s disease.  The press release at issue announced “top line” findings and promised a full review at an upcoming international conference.[20]  According to the release, the clinical trial data did not show a statistically significant benefit on the primary efficacy end point, but “[p]ost-hoc analyses did show statistically significant and clinically meaningful benefits in important subgroups.”[21]

The plaintiffs in Kleinman complained that the clinical trial had started with crucial imbalances between drug and placebo arms, thus indicating a failure in randomization, and that the positive results had come from impermissible post-hoc subgroup analyses.[22]  The appellate court appeared not to take the randomization issue seriously, and rejected the notion that statements can be false when they represent a defendant company’s reasonable interpretation of the data, even when the interpretation later turns out to be shown to be false[23]:

“At bottom, Kleinman simply has a problem with using post-hoc analyses as a methodology in pharmaceutical studies.  Kleinman cites commentators who liken post-hoc analyses to moving the goal posts or shooting an arrow into the wall and then drawing a target around it. Nonetheless, when it is clear that a post-hoc analysis is being used, it is understood that those results are less significant and should have less impact on investors.  Our job is not to evaluate the use of post-hoc analysis in the scientific community; the FDA has already done so.”

In United States v. Harkonen[24], the government turned the law of statistical analyses in securities fraud on its head, when it prosecuted a pharmaceutical company executive for his role in issuing a press release on clinical trial data. The jury acquitted Dr. Harkonen on a charge of misbranding[25], but convicted on a single count of wire fraud.[26] Dr. Harkonen’s crime?  Bad statistical practice.

The government conceded that the data represented in the press release were accurate, as were the calculated p-values.  The chargeable offense lay in Dr. Harkonen’s describing the clinical trial results as “demonstrating a survival benefit” of the biological product (interferon γ-1b) in a clinical trial subgroup of patients with mild- to moderate-idiopathic pulmonary fibrosis.  The p-value for the subgroup was 0.004, with an effect size of 70% reduction in mortality.  The subgroup, however, was not prespecified, and was not clearly labeled as a post-hoc analysis.  The trial had not achieved statistical significant on its primary end point.

In prosecuting Dr. Harkonen, the government offered no expert witness opinion.  Instead, it relied upon a member of the clinical trial’s data safety monitoring board, who advanced a strict, orthodox view that if the primary end point of a trial “failed,” then the data could not be relied upon to infer any meaningful causal connection within secondary end points, let alone non-prespecified end points.  The prespecified survival secondary end point showed a 40 percent reduction in mortality, p = 0.08 (which shrank to 0.055 on an intent-to-treat analysis). The press release also relied upon a previous small clinical trial that showed a benefit in survival at five years, with the therapy group at 77.8%, compared with 16.7% in the control groups, p = 0.009.

The trial court accepted the government’s claim that p-values less than 0.05 were something of “magic numbers,”[27] and rejected post-trial motions for accquittal. Dr. Harkonen’s use of “demonstrate” to describe a therapeutic benefit was, in the trial court’s view, fraudulent, because of the lack of “statistical significance” on the primary end point, and the multiple testing with respect to the secondary survival end point.  The Ninth Circuit affirmed the judgment of conviction in an unpublished per curiam opinion[28].

In contrast to the criminal wire fraud prosecution, the civil fraud actions against Dr. Harkonen and the company were dismissed.[29] The prosecution and the judgment in United States v. Harkonen is at odds with the latitude afforded companies in securities fraud cases.  Furthermore, the case cannot be fairly squared with the position that the government took as an amicus curiae in Matrixx Initiatives, Inc. v. Siracusano[30], where the Solicitor General’s office, along with counsel for the Food and Drug Division of the Department of Health & Human Services, in their zeal to assist plaintiffs on claims against an over-the-counter pharmaceutical manufacturer, disclaimed the necessity, or even the importance, of statistical significance[31]:

“[w]hile statistical significance provides some indication about the validity of a correlation between a product and a harm, a determination that certain data are not statistically significant … does not refute an inference of causation.”

Suddenly, when prosecuting an unpopular pharmaceutical company executive, the government’s flexibility evaporated. Government duplicity was a much greater problem than statistical multiplicity in Harkonen.[32]


[1] Security Exchange Comm’n Regulation FD, 17 C.F.R. § 243.100 (requiring prompt  public disclosure of any confidential, material inside information after disclosed to non-insiders).

[2] Selective Disclosure and Insider Trading, Securities Act Release No. 7881, Fed. Sec. L. Rep. (CCH) ¶ 86,319 (Aug. 15, 2000) (“As a general matter, acceptable methods of public disclosure for purposes of Regulation FD will include press releases distributed through a widely circulated news or wire service . . . .”).

[3] Section 10(b) of the Exchange Act of 1934 prohibits any person “[t]o use or employ, in connection with the purchase or sale of any security . . . any manipulative or deceptive device or contrivance in contravention of such rules and regulations as the [Securities and Exchange Commission] may prescribe.”  15 U.S.C. § 78j(b).

[4] 17 C.F.R. § 240.10b-5.

[5] Stoneridge Inv. Partners LLC v. Scientific-Atlanta, 552 U.S. 148, 157 (2008) (“(1) a material misrepresentation or omission []; (2) scienter; (3) a connection between the misrepresentation or omission and the purchase or sale of a security; (4) reliance upon the misrepresentation or omission; (5) economic loss; and (6) loss causation.”)

[6] In re Medimmune, Inc. Sec. Litig., 873 F.Supp. 953, 965 (D. Md. 1995).  The biological product at issue in this case was Respivir, a polyclonal antibody product, which “significantly” reduced frequency of hospitalization for respiratory syncytial virus (RSV).  Plaintiffs alleged “flaws” in study design, but the trial court appeared to interpret the statistical significance to mean that Respivir was “unquestionably efficacious.” Id. at 967.

[7] See, e.g., Padnes v. Scios Nova Inc., No. C 95-1693 MHP, 1996 WL 539711, at *5 (N.D. Cal. Sept. 18, 1996) (Patel, J.)[cited herein as Padnes].  See also DeMarco v. DePoTech Corp., 149 F.Supp. 2d 1212, 1225 (S.D. Cal. 2001)(“Although plaintiffs have established a legitimate difference in opinion as to the proper statistical analysis, they have hardly stated a securities fraud claim.”); n re Aldor Corp. Sec. Litig., 616 F.Supp. 2d 551, 568 n.15 (E.D. Pa. 2009) (allegations as to how data should have been analyzed do not support claims for false or misleading statements).

[8] Padnes at *2.

[9] Id. at *10.

[10] 2005 WL 4161977 (D. Colo. Oct. 20, 2005).

[11] Id. at *1.

[12] Id. at *6-7.

[13] Id. at *11.

[14] 2010 WL 8816155 (N.D. Cal. Aug. 24, 2010).

[15] 697 F.3d 869, 877 (9th Cir. 2012) (internal citations omitted), aff’g 2010 WL 8816155 (N.D. Cal. Aug. 24, 2010).

[16] Id. at 877-78.

[17] Id. at 878.

[18] Id. (“Because there are many ways to statistically analyze data, it is necessary to choose the statistical methodology before seeing the data that is collected during the trial; otherwise someone can manipulate the unblinded data to obtain a favorable result.”), citing and attempting to distinguish United States v. Harkonen, 2010 WL 2985257, at *4 (N.D. Cal. July 27, 2010).

[19] 706 F.3d 145 (2d Cir. 2013).

[20] Id. at 149.

[21] Id. at 149-50 (also noting that the press release provided a “preliminary analysis,” which might be less favorable upon further analysis).

[22] Id. at 150.

[23] Id. at 154-55 & 155n.11 (citing and quoting FDA Center for Drug Evaluation and Research:  E9 Statistical Principles for Clinical Trials, 63 Fed. Reg. 49583, 49595 (Sept. 16, 1998), that post-hoc analyses are exploratory and “unlikely” to be accepted as support of efficacy.)

[24] United States v. Harkonen, 2010 WL 2985257 (N.D. Calif. 2010) ((Patel, J.) (denying defendant’s post–trial motions to dismiss the indictment, for acquittal, or for a new trial).  Sometimes judges are looking for bright lines in the wrong places).

[25] 21 U.S.C. §§ 331(k), 333(a)(2), 352(a).

[26] 18 U.S.C. § 1343.

[27] United States v. Harkonen, 2010 WL 2985257, at *5 (N.D. Calif. 2010).

[28] United States v. Harkonen, 2013 WL 782354 (9th Cir. 2013).

[29] In re Actimmune Marketing Litig., 2010 WL 3463491 (N.D. Cal. Sept. 1, 2010), aff’d,  464 Fed.Appx. 651 (9th Cir. 2011).

[30] 131 S. Ct. 1309 (2011).

[31] Brief for the United States as Amicus Curiae Supporting Respondents, in Matrixx Initiatives, Inc. v. Siracusano, 2010 WL 4624148, at *14 (Nov. 12, 2010).

[32] Dr. Harkonen is expected to petition the Supreme Court for certiorari on statutory and constitutional grounds.  See Alex Kozinski & Stuart Banner, “Who’s Afraid of Commercial Speech?” 76 VA. L. REV. 627, 635 (1990) (“[T]here are many varieties of noncommercial speech that are just as objective as paradigmatic commercial speech and yet receive full first amendment protection. Scientific speech is the most obvious; much scientific expression can easily be labeled true or false, but we would be shocked at the suggestion that it is therefore entitled to a lesser degree of protection. If you want, you can proclaim that the sun revolves around the earth, that the earth is flat, that there is no such thing as nitrogen, that flounder smoke cigars, that you have fused atomic nuclei in your bathtub — you can spout any nonsense you want, and the government can’t stop you.”).