Bad pharma How drug companies mislead doctors and harm patients

Ben Goldacre

Sound recording - 2013

Goldacre puts the 600-billion-dollar global pharmaceutical industry under the microscope. What he reveals is a fascinating, terrifying mess.

Saved in:

2nd Floor Show me where

COMPACT DISC/615.1/Goldacre
1 / 1 copies available
Location Call Number   Status
2nd Floor COMPACT DISC/615.1/Goldacre Checked In
Subjects
Genres
Video recordings for the hearing impaired
Published
[Old Saybrook, Ct.] : Tantor Audio p2013.
Language
English
Main Author
Ben Goldacre (-)
Edition
Library ed
Item Description
Unabridged.
Physical Description
10 compact discs (12hr.) : digital ; 4 3/4 in
ISBN
9781452641690
  • Missing data
  • Where do new drugs come from?
  • Bad regulators
  • Bad trials
  • Bigger, simpler trials
  • Marketing
  • Afterword: better data.
Review by New York Times Review

MEDICAL KNOWLEDGE IS evolving at warp speed, spinning out reams of new information into sound bites, articles and more and more new health books to overload your sagging shelves. But if you're tempted to toss out all the old stuff, better think twice - despite the changes in medical science, some constants endure. The human body still sickens and recovers much the way it always has; our dogged, heartfelt efforts to prevent and relieve pain and suffering are no different than they ever were. Even books published decades or in some cases centuries in the past may still speak clearly to today's medical issues. So squeeze all the shiny new books in among the dusty old ones, and let this column, Prescribed Reading, justify the clutter. When health care makes headlines, these occasional essays will take a look back at other volumes on that bulging shelf, for a curated perspective on resources available to the interested reader. No health topic has made more news in this country over the last two decades than the opioid crisis, with heartrending journalism and a handful of comprehensive books. Barry Meier's 2003 "Pain Killer" was the first to probe the deceptive advertising pushing the addictive prescription opioid OxyContin into the American market. Sam Quinones's compulsively readable "Dreamland" linked the popularity of OxyContin to an exploding nationwide traffic in a cheaper alternative, potent Mexican heroin. Beth Macy's recent "Dopesick" outlined the synergistic destruction that legal and illegal narcotics wreak on users and their communities. These authors all covered a similar territory of clueless doctors, rapacious drug sales forces (both the corporate and the criminal varieties), hurting patients and their frantic friends and relatives. Some may wonder what America's doctors could have been thinking all those years, doling out quantities of potent narcotics like so many aspirin. Granted, a few of us were criminals, methodically defrauding Medicaid in pilldispensing "mills." But mostly we were just well-intentioned schlubs with prescription pads, dutifully following then-current practice guidelines. It is impossible to overstate how difficult it was a dozen years ago to ignore the resounding calls for effective pain control by any means possible, narcotic dependency be damned. Our leaders, our teachers, our regulatory agencies and our patients demanded no less. The slick ad campaigns backing OxyContin and all the other addictive prescription opioids assured us all would be well. Dr. Anna Lembke's small but powerful "Drug Dealer, MD" summarizes other seldom-enunciated reasons doctors became complicit. A specialist at Stanford in addiction medicine, Lembke delivers a bottom line as bizarre as it is true: Most doctors hate and fear dealing with pain, and are utterly unequipped to do so. Treating pain properly requires specialized training; it consumes vast amounts of time few doctors have. It requires an ability to set limits and negotiate and to fail and fail again. And, as Lembke points out, doctors are a group chosen for more or less exactly the opposite characteristics. We are habitual pleasers, accustomed - even addicted - to patients' admiration and gratitude. We like to do what we do well, and most of us don't do pain well. So it's no big surprise that when we find ourselves in a pain-filled room we just want to escape fast, and prescribing another round of pills lets us do just that. What about the part played by pharmaceutical companies in this crisis? Was Purdue Pharma, manufacturer of OxyContin, an aberration in its misleading promotion of the drug, with assertions that ultimately resulted in $600 million in fines and penalties? Or is this just how Big Pharma always operates, dodging the law and relevant moral principles in single-minded pursuit of Big Profit? Back in 2004, Dr. Marcia Angelí, a former editor of The New England Journal of Medicine, wrote in "The Truth About Drug Companies" that the pharmaceutical industry "has moved very far from its original high purpose of discovering and producing useful new drugs." Her book presents instead a vision of an industry motivated largely by greed. Pretty much every specific outrage critics have identified in Purdue's behavior appears in Angell's analysis, including the "jaw-dropping" gifts doctors may receive for prescribing a company's drugs, the subtle advertisement masquerading as education and the lukewarm, often ineffective protestations by regulatory agencies. Writing from Britain, Dr. Ben Goldacre echoes Angell's concerns, charging in " Bad Pharma" that the worst misbehavior in the pharmaceutical industry actually occurs not when prescription drugs are being burnished for market, but far earlier in their development. Lackluster compounds are evaluated in ways guaranteed to make them look good, Goldacre writes, then prettied up even further with distorted claims of efficacy. While "bad behavior in marketing departments is unpleasant," he concludes, the real scientific outrage and the big public danger lie in uniformly dubious practices of drug development. The immersion journalist Barry Werth provided a more sympathetic view of the pharmaceutical industry after taking two deep dives into the workings of a young biotech company. In exchange for many years of fly-on-the-wall privileges at Vertex Pharmaceuticals, Werth allowed his manuscripts to be vetted by corporate executives, and presumably some of their sharper edges were sanded down. What remains, though, is an impressively detailed, dense epic in two volumes, "The Billion-Dollar Molecule" and "The Antidote," describing the truly herculean labors needed to birth a panel of marketable drugs. Vertex developed early H.I.V. and hepatitis C drugs, and many of the executives Werth trails are clearly committed to the humanitarian aspects of their work. But they are also committed to their own proprietary molecules, sometimes with a passion bordering on worship. As the head of Vertex once announced at an industry conference, "It's a tremendous responsibility to live up to a molecule like VX-950." A regulator for the F.D.A. put it a little more bluntly, screaming over the phone to another Vertex executive, "You guys believe in your drug too much." Opioid-dependent patients may be trapped in a web of pain and addiction for months, years or a lifetime. One of the most eloquent descriptions of that hard fate comes to us from more than 200 years ago, back in the days when opium and laudanum (a solution of opium in alcohol) were perfectly legal in England, and as widely available as, yes, aspirin is today. It was in the fall of 1804 that Thomas De Quincey, a 19-year-old wannabe intellectual, decided to try a little opium for a bad toothache. When De Quincey published his "Confessions of an English Opium-Eater" 17 years later, his poetic depictions of the wild hallucinations that punctuated his years with the drug transfixed his contemporaries. To us, though, his story is revelatory mostly for its eerie familiarity. As a writer, De Quincey was dominated and arguably ruined by his habit. He produced quantities of hack prose, but nothing as enduring as the "Confessions." He lovingly embroidered the original version of that classic into sequels and revisions, all of which were increasingly wordy and decreasingly interesting. De Quincey had wonderful ideas he could never remember, often scribbled down on pieces of paper he could never find. What he might have accomplished sober is anyone's guess. ABIGAIL ZUGER is a physician and a frequent contributor to Science Times. Rx Pad DRUG DEALER, MD: How Doctors Were Duped, Patients Got Hooked, and Why It's So Hard to Stop, by Dr. Anna Lembke. The best summary around of the perils of owning a prescription pad. THE TRUTH ABOUT THE DRUG COMPANIES: How They Deceive Us and What to Do About it, by Dr. Marcia Angelí. An eloquent classic. No one has despised Big Pharma for as long as Dr. Angelí has. BAD PHARMA: How Drug Companies Mislead Doctors and Harm Patients, by Dr. Ben Goldacre. Read it and you may never put another pill in your mouth. THE BILLION-DOLLAR MOLECULE: The Quest for the Perfect Drug, by Barry Werth. THE ANTIDOTE: Inside the World of New Pharma, by Barry Werth. The drug business presented meeting by meeting. Tedious, but you won't find this level of detail anywhere else. confessions of an English opium-eater, by Thomas De Quincey. Bear with the antique prose for a mind-blowingly modern story line.

Copyright (c) The New York Times Company [June 30, 2019]
Review by Booklist Review

In the follow-up to his popular Bad Science (2010), British medical doctor Goldacre reveals how pharmaceutical companies mislead doctors and hurt patients. They sponsor trials, which tend to yield favorable results, while negative results often remain unreported. He also reports that drug companies spend twice as much on marketing and advertising as on researching and developing new drugs. Unfortunately for U.S. readers, he focuses largely on the UK, but ghost authorship of studies and continuing medical education boondoggle trips for doctors are problematic everywhere, and he does refer to the U.S. Food and Drug Administration on multiple occasions. And everyone, everywhere should feel unsettled by his discovery that pharmaceutical companies funnel $10 million to $20 million a year to such major medical journals as the New England Journal of Medicine and the Journal of the American Medical Association. Not surprisingly, he notes, studies funded by the pharmaceutical industry are that much more likely to get published in these influential journals. Goldacre's essential expose will prompt readers to ask more questions before automatically popping a doctor-prescribed pill.--Springen, Karen Copyright 2010 Booklist

From Booklist, Copyright (c) American Library Association. Used with permission.
Review by Publisher's Weekly Review

In his latest, British physician and author Goldacre tackles the misdeeds of the pharmaceutical industry. As Goldacre presents a laundry list of flawed research projects, narrator Jonathan Cowley handles the author's academic style with ease, never missing a beat. His precise annunciation matches the seriousness of the issues at hand, and he successfully balances the shifting tones of the narrative. Yet, as Goldacre recounts his adventures uncovering greed and corruption, Cowley ably takes on the author's populist persona. Cowley especially entertains in sections devoted to industry schmoozing and networking, providing doses of humor to help bring home the author's underlying messages. And if technical and scientific sections of the book make for a sometimes-demanding listening experience, Cowley's winning reading helps broaden the appeal. A Faber & Faber hardcover. (Feb.) (c) Copyright PWxyz, LLC. All rights reserved.

(c) Copyright PWxyz, LLC. All rights reserved
Review by Library Journal Review

British epidemiologist Goldacre (Bad Science: Quacks, Hacks, and Big Pharma Flacks), who writes the popular Guardian column "Bad Science," is unabashedly polemical in his latest book. He addresses biases against publishing negative results, the limitations of current clinical trials, and the growing (but still limited) use of evidence-based medicine and comparative effectiveness research, along with the challenges and uncertainties of medicine. Published originally in the UK, the book also includes a number of examples from the United States. Building upon work such as Iain Chambers's Testing Treatments: Better Research for Better Healthcare (for which Goldacre wrote the foreword), Jerry Avorn's Powerful Medicines, and Marcia Angell's The Truth About Drug Companies, Goldacre's in-your-face rhetoric will capture attention and should spark important conversations about ways to improve medicine, clinician education, drug safety monitoring, and drug regulation and marketing. Verdict Goldacre's recommendations for much larger, simpler trials and for more access to clinical trial data, as well as educating people about risk assessment, clinical trial design, and statistical literacy make this much more than a condemnation of the pharmaceutical industry. Recommended for academic, health industry, and lay audiences.-Mary Chitty, Cambridge Healthtech, Neeedham, MA (c) Copyright 2013. Library Journals LLC, a wholly owned subsidiary of Media Source, Inc. No redistribution permitted.

(c) Copyright Library Journals LLC, a wholly owned subsidiary of Media Source, Inc. No redistribution permitted.
Review by Kirkus Book Review

An explanation of why pharmaceutical companies have come under increasing scrutiny in recent years. As both a physician and medical correspondent (Bad Science, 2010), Goldacre has been in the catbird seat in regard to the detection of medical fraud. Here, he discusses the gray area in which drug companies can legally manipulate data in order to package experimental results in the most attractive way. Although Goldacre practices medicine in the U.K., his book is applicable to the U.S. pharmaceutical industry. He explains what should be an obvious flaw in the system: The majority of experiments validating the safety and benefits of new drugs are funded by the pharmaceutical industry, either directly or indirectly, through grants to universities and other methods. Because of this, writes Goldacre, "industry-sponsored studies are more likely to produce results that flatter the sponsor's drug." The author cites the conclusion by three researchers from Harvard and Toronto, who conducted a meta-analysis of data on antidepressants and other drugs in 2010. They found major discrepancies in results related to reported effectiveness between research conducted by industry-funded and government-funded trials. Goldacre believes that such reviews should be conducted routinely, backdated and made easily available to the public. Currently, pharmaceutical companies are not obligated to publish results that are unfavorable to their product. Another shady tactic is touting the short-term performance of drugs that become less effective over time. The author supports enforceable government regulation. A useful guide for policymakers, doctors and the patients who need protection against deliberate disinformation.]] Copyright Kirkus Reviews, used with permission.

Copyright (c) Kirkus Reviews, used with permission.

1 Missing Data Sponsors get the answer they want Before we get going, we need to establish one thing beyond any doubt: industry-funded trials are more likely to produce a positive, flattering result than independently funded trials. This is our core premise, and you're about to read a very short chapter, because this is one of the most well-documented phenomena in the growing field of 'research about research'. It has also become much easier to study in recent years, because the rules on declaring industry funding have become a little clearer. We can begin with some recent work: in 2010, three researchers from Harvard and Toronto found all the trials looking at five major classes of drug - antidepressants, ulcer drugs and so on - then measured two key features: were they positive, and were they funded by industry?1 They found over five hundred trials in total: 85 per cent of the industry-funded studies were positive, but only 50 per cent of the government-funded trials were. That's a very significant difference. In 2007, researchers looked at every published trial that set out to explore the benefit of a statin.2 These are cholesterol-lowering drugs which reduce your risk of having a heart attack, they are prescribed in very large quantities, and they will loom large in this book. This study found 192 trials in total, either comparing one statin against another, or comparing a statin against a different kind of treatment. Once the researchers controlled for other factors (we'll delve into what this means later), they found that industry-funded trials were twenty times more likely to give results favouring the test drug. Again, that's a very big difference. We'll do one more. In 2006, researchers looked into every trial of psychiatric drugs in four academic journals over a ten-year period, finding 542 trial outcomes in total. Industry sponsors got favourable outcomes for their own drug 78 per cent of the time, while independently funded trials only gave a positive result in 48 per cent of cases. If you were a competing drug put up against the sponsor's drug in a trial, you were in for a pretty rough ride: you would only win a measly 28 per cent of the time.3 These are dismal, frightening results, but they come from individual studies. When there has been lots of research in a field, it's always possible that someone - like me, for example - could cherry-pick the results, and give a partial view. I could, in essence, be doing exactly what I accuse the pharmaceutical industry of doing, and only telling you about the studies that support my case, while hiding the reassuring ones from you. To guard against this risk, researchers invented the systematic review. We'll explore this in more detail soon (here), since it's at the core of modern medicine, but in essence a systematic review is simple: instead of just mooching through the research literature, consciously or unconsciously picking out papers here and there that support your pre-existing beliefs, you take a scientific, systematic approach to the very process of looking for scientific evidence, ensuring that your evidence is as complete and representative as possible of all the research that has ever been done. Systematic reviews are very, very onerous. In 2003, by coincidence, two were published, both looking specifically at the question we're interested in. They took all the studies ever published that looked at whether industry funding is associated with pro-industry results. Each took a slightly different approach to finding research papers, and both found that industry-funded trials were, overall, about four times more likely to report positive results.4 A further review in 2007 looked at the new studies that had been published in the four years after these two earlier reviews: it found twenty more pieces of work, and all but two showed that industry-sponsored trials were more likely to report flattering results.5 I am setting out this evidence at length because I want to be absolutely clear that there is no doubt on the issue. Industry-sponsored trials give favourable results, and that is not my opinion, or a hunch from the occasional passing study. This is a very well-documented problem, and it has been researched extensively, without anybody stepping out to take effective action, as we shall see. There is one last study I'd like to tell you about. It turns out that this pattern of industry-funded trials being vastly more likely to give positive results persists even when you move away from published academic papers, and look instead at trial reports from academic conferences, where data often appears for the first time (in fact, as we shall see, sometimes trial results only appear at an academic conference, with very little information on how the study was conducted). Fries and Krishnan studied all the research abstracts presented at the 2001 American College of Rheumatology meetings which reported any kind of trial, and acknowledged industry sponsorship, in order to find out what proportion had results that favoured the sponsor's drug. There is a small punchline coming, and to understand it we need to cover a little of what an academic paper looks like. In general, the results section is extensive: the raw numbers are given for each outcome, and for each possible causal factor, but not just as raw figures. The 'ranges' are given, subgroups are perhaps explored, statistical tests are conducted, and each detail of the result is described in table form, and in shorter narrative form in the text, explaining the most important results. This lengthy process is usually spread over several pages. In Fries and Krishnan [2004] this level of detail was unnecessary. The results section is a single, simple, and - I like to imagine - fairly passive-aggressive sentence: The results from every RCT (45 out of 45) favored the drug of the sponsor. This extreme finding has a very interesting side effect, for those interested in time-saving shortcuts. Since every industry-sponsored trial had a positive result, that's all you'd need to know about a piece of work to predict its outcome: if it was funded by industry, you could know with absolute certainty that the trial found the drug was great. How does this happen? How do industry-sponsored trials almost always manage to get a positive result? It is, as far as anyone can be certain, a combination of factors. It may be that companies are more likely to run trials when they're most confident their treatment is going to 'win': this sounds reasonable, although even it conflicts with the ethical principle that you should only do a trial when there's genuine uncertainty about which treatment is best (otherwise you're exposing half of your participants to a treatment you already know to be inferior). Sometimes the chances of one treatment winning can be increased by outright design flaws. You can compare your new drug with something you know to be rubbish - an existing drug at an inadequate dose, perhaps, or a placebo sugar pill that does almost nothing. You can choose your patients very carefully, so they are more likely to get better on your treatment. You can peek at the results halfway through, and stop your trial early if they look good (which is - for interesting reasons we shall discuss - statistical poison). And so on. But before we get to these fascinating methodological twists and quirks, these nudges and bumps that stop a trial from being a fair test of whether a treatment works or not, there is something very much simpler at hand. Sometimes drug companies conduct lots of trials, and when they see that the results are unflattering, they simply fail to publish them. This is not a new problem, and it's not limited to medicine. In fact, this issue of negative results that go missing in action cuts into almost every corner of science. It distorts findings in fields as diverse as brain imaging and economics, it makes a mockery of all our efforts to exclude bias from our studies, and despite everything that regulators, drug companies and even some academics will tell you, it is a problem that has been left unfixed for decades. In fact, it is so deep-rooted that even if we fixed it today - right now, for good, forever, without any flaws or loopholes in our legislation - that still wouldn't help, because we would still be practising medicine, cheerfully making decisions about which treatment is best, on the basis of decades of medical evidence which is - as you've now seen - fundamentally distorted. But there is a way ahead. Why missing data matters Reboxetine is a drug I myself have prescribed. Other drugs had done nothing for this particular patient, so we wanted to try something new. I'd read the trial data before I wrote the prescription, and found only well-designed, fair tests, with overwhelmingly positive results. Reboxetine was better than placebo, and as good as any other antidepressant in head-to-head comparisons. It's approved for use by the Medicines and Healthcare products Regulatory Agency (the MHRA) in the UK, but wisely, the US FDA chose not to approve it. (This is no proof of the FDA being any smarter; there are plenty of drugs available in the US that the UK never approved.) Reboxetine was clearly a safe and and effective treatment. The patient and I discussed the evidence briefly, and agreed it was the right treatment to try next. I signed a prescription saying I wanted my patient to have this drug. But we had both been misled. In October 2010 a group of researchers were finally able to bring together all the trials that had ever been conducted on reboxetine.6 Through a long process of investigation - searching in academic journals, but also arduously requesting data from the manufacturers and gathering documents from regulators - they were able to assemble all the data, both from trials that were published, and from those that had never appeared in academic papers. When all this trial data was put together it produced a shocking picture. Seven trials had been conducted comparing reboxetine against placebo. Only one, conducted in 254 patients, had a neat, positive result, and that one was published in an academic journal, for doctors and researchers to read. But six more trials were conducted, in almost ten times as many patients. All of them showed that reboxetine was no better than a dummy sugar pill. None of these trials was published. I had no idea they existed. It got worse. The trials comparing reboxetine against other drugs showed exactly the same picture: three small studies, 507 patients in total, showed that reboxetine was just as good as any other drug. They were all published. But 1,657 patients' worth of data was left unpublished, and this unpublished data showed that patients on reboxetine did worse than those on other drugs. If all this wasn't bad enough, there was also the side-effects data. The drug looked fine in the trials which appeared in the academic literature: but when we saw the unpublished studies, it turned out that patients were more likely to have side effects, more likely to drop out of taking the drug, and more likely to withdraw from the trial because of side effects, if they were taking reboxetine rather than one of its competitors. If you're ever in any doubt about whether the stories in this book make me angry - and I promise you, whatever happens, I will keep to the data, and strive to give a fair picture of everything we know - you need only look at this story. I did everything a doctor is supposed to do. I read all the papers, I critically appraised them, I understood them, I discussed them with the patient, and we made a decision together, based on the evidence. In the published data, reboxetine was a safe and effective drug. In reality, it was no better than a sugar pill, and worse, it does more harm than good. As a doctor I did something which, on the balance of all the evidence, harmed my patient, simply because unflattering data was left unpublished. If you find that amazing, or outrageous, your journey is just beginning. Because nobody broke any law in that situation, reboxetine is still on the market, and the system that allowed all this to happen is still in play, for all drugs, in all countries in the world. Negative data goes missing, for all treatments, in all areas of science. The regulators and professional bodies we would reasonably expect to stamp out such practices have failed us. In a few pages, we will walk through the literature that demonstrates all of this beyond any doubt, showing that 'publication bias' - the process whereby negative results go unpublished - is endemic throughout the whole of medicine and academia; and that regulators have failed to do anything about it, despite decades of data showing the size of the problem. But before we get to that research, I need you to feel its implications, so we need to think about why missing data matters. Evidence is the only way we can possibly know if something works - or doesn't work - in medicine. We proceed by testing things, as cautiously as we can, in head-to-head trials, and gathering together all of the evidence. This last step is crucial: if I withhold half the data from you, it's very easy for me to convince you of something that isn't true. If I toss a coin a hundred times, for example, but only tell you about the results when it lands heads-up, I can convince you that this is a two-headed coin. But that doesn't mean I really do have a two-headed coin: it means I'm misleading you, and you're a fool for letting me get away with it. This is exactly the situation we tolerate in medicine, and always have. Researchers are free to do as many trials as they wish, and then choose which ones to publish. The repercussions of this go way beyond simply misleading doctors about the benefits and harms of interventions for patients, and way beyond trials. Medical research isn't an abstract academic pursuit: it's about people, so every time we fail to publish a piece of research we expose real, living people to unnecessary, avoidable suffering. TGN1412 In March 2006, six volunteers arrived at a London hospital to take part in a trial. It was the first time a new drug called TGN1412 had ever been given to humans, and they were paid £2,000 each.7 Within an hour these six men developed headaches, muscle aches, and a feeling of unease. Then things got worse: high temperatures, restlessness, periods of forgetting who and where they were. Soon they were shivering, flushed, their pulses racing, their blood pressure falling. Then, a cliff: one went into respiratory failure, the oxygen levels in his blood falling rapidly as his lungs filled with fluid. Nobody knew why. Another dropped his blood pressure to just 65/40, stopped breathing properly, and was rushed to an intensive care unit, knocked out, intubated, mechanically ventilated. Within a day all six were disastrously unwell: fluid on their lungs, struggling to breathe, their kidneys failing, their blood clotting uncontrollably throughout their bodies, and their white blood cells disappearing. Doctors threw everything they could at them: steroids, antihistamines, immune-system receptor blockers. All six were ventilated on intensive care. They stopped producing urine; they were all put on dialysis; their blood was replaced, first slowly, then rapidly; they needed plasma, red cells, platelets. The fevers continued. One developed pneumonia. And then the blood stopped getting to their peripheries. Their fingers and toes went flushed, then brown, then black, and then began to rot and die. With heroic effort, all escaped, at least, with their lives. The Department of Health convened an Expert Scientific Group to try to understand what had happened, and from this two concerns were raised.8 First: can we stop things like this from happening again? It's plainly foolish, for example, to give a new experimental treatment to all six participants in a 'first-in-man' trial at the same time, if that treatment is a completely unknown quantity. New drugs should be given to participants in a staggered process, slowly, over a day. This idea received considerable attention from regulators and the media. Less noted was a second concern: could we have foreseen this disaster? TGN1412 is a molecule that attaches to a receptor called CD28 on the white blood cells of the immune system. It was a new and experimental treatment, and it interfered with the immune system in ways that are poorly understood, and hard to model in animals (unlike, say, blood pressure, because immune systems are very variable between different species). But as the final report found, there was experience with a similar intervention: it had simply not been published. One researcher presented the inquiry with unpublished data on a study he had conducted in a single human subject a full ten years earlier, using an antibody that attached to the CD3, CD2 and CD28 receptors. The effects of this antibody had parallels with those of TGN1412, and the subject on whom it was tested had become unwell. But nobody could possibly have known that, because these results were never shared with the scientific community. They sat unpublished, unknown, when they could have helped save six men from a terrifying, destructive, avoidable ordeal. That original researcher could not foresee the specific harm he contributed to, and it's hard to blame him as an individual, because he operated in an academic culture where leaving data unpublished was regarded as completely normal. The same culture exists today. The final report on TGN1412 concluded that sharing the results of all first-in-man studies was essential: they should be published, every last one, as a matter of routine. But phase 1 trial results weren't published then, and they're still not published now. In 2009, for the first time, a study was published looking specifically at how many of these first-in-man trials get published, and how many remain hidden.9 They took all such trials approved by one ethics committee over a year. After four years, nine out of ten remained unpublished; after eight years, four out of five were still unpublished. In medicine, as we shall see time and again, research is not abstract: it relates directly to life, death, suffering and pain. With every one of these unpublished studies we are potentially exposed, quite unnecessarily, to another TGN1412. Even a huge international news story, with horrific images of young men brandishing blackened feet and hands from hospital beds, wasn't enough to get movement, because the issue of missing data is too complicated to fit in one sentence. When we don't share the results of basic research, such as a small first-in-man study, we expose people to unnecessary risks in the future. Was this an extreme case? Is the problem limited to early, experimental, new drugs, in small groups of trial participants? No. In the 1980s, US doctors began giving anti-arrhythmic drugs to all patients who'd had a heart attack. This practice made perfect sense on paper: we knew that anti-arrhythmic drugs helped prevent abnormal heart rhythms; we also knew that people who've had a heart attack are quite likely to have abnormal heart rhythms; we also knew that often these went unnoticed, undiagnosed and untreated. Giving anti-arrhythmic drugs to everyone who'd had a heart attack was a simple, sensible preventive measure. Unfortunately, it turned out that we were wrong. This prescribing practice, with the best of intentions, on the best of principles, actually killed people. And because heart attacks are very common, it killed them in very large numbers: well over 100,000 people died unnecessarily before it was realised that the fine balance between benefit and risk was completely different for patients without a proven abnormal heart rhythm. Could anyone have predicted this? Sadly, yes, they could have. A trial in 1980 tested a new anti-arrhythmic drug, lorcainide, in a small number of men who'd had a heart attack - less than a hundred - to see if it was any use. Nine out of forty-eight men on lorcainide died, compared with one out of forty-seven on placebo. The drug was early in its development cycle, and not long after this study it was dropped for commercial reasons. Because it wasn't on the market, nobody even thought to publish the trial. The researchers assumed it was an idiosyncrasy of their molecule, and gave it no further thought. If they had published, we would have been much more cautious about trying other anti-arrhythmic drugs on people with heart attacks, and the phenomenal death toll - over 100,000 people in their graves prematurely - might have been stopped sooner. More than a decade later, the researchers finally did publish their results, with a mea culpa , recognising the harm they had done by not sharing them earlier: When we carried out our study in 1980, we thought that the increased death rate that occurred in the lorcainide group was an effect of chance. The development of lorcainide was abandoned for commercial reasons, and this study was therefore never published; it is now a good example of 'publication bias'. The results described here might have provided an early warning of trouble ahead.10 As we shall shortly see, this problem of unpublished data is widespread throughout medicine, and indeed the whole of academia, even though the scale of the problem, and the harm it causes, have been documented beyond any doubt. We will see stories on basic cancer research, Tamiflu, cholesterol blockbusters, obesity drugs, antidepressants and more, with evidence that goes from the dawn of medicine to the present day, and data that is still being withheld, right now, as I write, on widely used drugs which many of you reading this book will have taken this morning. We will also see how regulators and academic bodies have repeatedly failed to address the problem. Because researchers are free to bury any result they please, patients are exposed to harm on a staggering scale throughout the whole of medicine, from research to practice. Doctors can have no idea about the true effects of the treatments they give. Does this drug really work best, or have I simply been deprived of half the data? Nobody can tell. Is this expensive drug worth the money, or have the data simply been massaged? No one can tell. Will this drug kill patients? Is there any evidence that it's dangerous? No one can tell. This is a bizarre situation to arise in medicine, a discipline where everything is supposed to be based on evidence, and where everyday practice is bound up in medico-legal anxiety. In one of the most regulated corners of human conduct we've taken our eyes off the ball, and allowed the evidence driving practice to be polluted and distorted. It seems unimaginable. We will now see how deep this problem goes. Copyright © 2012, 2013 by Ben Goldacre Excerpted from Bad Pharma: How Drug Companies Mislead Doctors and Harm Patients by Ben Goldacre All rights reserved by the original copyright owners. Excerpts are provided for display purposes only and may not be reproduced, reprinted or distributed without the written permission of the publisher.