Meltdown Why our systems fail and what we can do about it

Chris Clearfield

Book - 2018

"Weaving together cutting-edge social science with riveting stories that take us from the frontlines of the Volkswagen scandal to backstage at the Oscars, and from deep beneath the Gulf of Mexico to the top of Mount Everest, Chris Clearfield and András Tilcsik explain how the increasing complexity of our systems creates conditions ripe for failure and why our brains and teams can't keep up. They highlight the paradox of progress: Though modern systems have given us new capabilities, they've become vulnerable to surprising meltdowns--and even to corruption and misconduct. But Meltdown isn't just about failure; it's about solutions--whether you're managing a team or the chaos of your family's morning routin...e. It reveals why ugly designs make us safer, how a five-minute exercise can prevent billion-dollar catastrophes, why teams with fewer experts are better at managing risk, and why diversity is one of our best safeguards against failure. The result is an eye-opening, empowering, and entirely original book--one that will change the way you see our complex world and your own place in it."--Jacket flap.

Saved in:

2nd Floor Show me where

158.1/Clearfield
1 / 1 copies available
Location Call Number   Status
2nd Floor 158.1/Clearfield Checked In
Subjects
Published
New York : Penguin Press 2018.
Language
English
Main Author
Chris Clearfield (author)
Other Authors
András Tilcsik (author)
Physical Description
294 pages : illustrations ; 25 cm
Bibliography
Includes bibliographical references (pages 251-286) and index.
ISBN
9780735222632
  • Prologue: A Day Like Any Other
  • Part 1. Failure All Around Us
  • 1. The Danger Zone
  • 2. Deep Waters, New Horizons
  • 3. Hacking, Fraud, and All the News That's Unfit to Print
  • Part 2. Conquering Complexity
  • 4. Out of the Danger Zone
  • 5. Complex Systems, Simple Tools
  • 6. Reading the Writing on the Wall
  • 7. The Anatomy of Dissent
  • 8. The Speed Bump Effect
  • 9. Strangers in a Strange Land
  • 10. Surprise!
  • Epilogue: The Golden Age of Meltdowns
  • Acknowledgments
  • Notes
  • Index
Review by Publisher's Weekly Review

Clearfield, a former derivatives trader, and Tilcsik, a management academic, share cautionary tales of disaster resulting from small vulnerabilities in large systems. Their analysis is enlightening but they flounder in translating their insights into usable takeaways. Clearfield and Tilcsik's approach centers on sociologist Charles Perrow's 1984 theory that as a system's complexity and "tight coupling" (a lack of slack between different parts) increase, the "danger zone" does as well. Clear, well-paced storytelling around diverse events, including a fatal D.C. Metro train accident, Three Mile Island, the collapse of Enron, Volkswagen's emissions cheating, the Flint water crisis, the 2017 Oscars mix-up, and the commandeering of a Starbucks hashtag by the company's critics, will keep readers interested, whether or not they are invested in the organizational lessons. The solutions offered, however, tend toward the less-than-revolutionary: keeping decision-making parameters clear, increasing workforce diversity, and building organizational cultures in which dissent is genuinely encouraged. Clearfield and Tilcsik's most important warning is about the "normalization of deviance," when people come to redefine commonly encountered risks as acceptable, as can occur when automated warnings constantly cry wolf in hospitals. This manual articulates the ubiquitous nature of system failure well, but its approaches to "reducing complexity and adding slack" are too vague to be practically implementable. (Mar.) © Copyright PWxyz, LLC. All rights reserved.

(c) Copyright PWxyz, LLC. All rights reserved
Review by Kirkus Book Review

If it can be built, it can fall apart: a cautionary study in how complex systems can easily go awry.As systems become more complex, guided by artificial intelligence and algorithms as well as human experience, they become more likely to fail. The result, write one-time derivatives trader and commercial pilot Clearfield and Tilcsik (Univ. of Toronto Rotman School of Management), is that we are now "in the golden age of meltdowns," confronted on all sides by things that fall apart, whether the financial fortunes of entrepreneurs, the release valves of dam plumbing, or the ailerons of jetliners. The authors examine numerous case studies of how miscommunications and failed checklists figure into disaster, as with one notorious air crash where improperly handled oxygen canisters produced a fatal in-flight fire: "The investigation," they write, "revealed a morass of mistakes, coincidences, and everyday confusions." Against this, Clearfield and Tilcsik helpfully propose ways in which the likelihood of disaster or unintended consequences can be lessened: cross-training, for instance, so that members of a team know something of one another's jobs and responsibilities, and iterative processes of checking and cross-checking. At times, the authors venture into matters of controversy, as when they observe that mandatory diversity training yields more rather than less racist behavior and suggest that "targeted recruitment" of underrepresented groups sends a more positive message: "Help us find a greater variety of promising employees!" Though the underlying argument isn't newthe authors draw heavily on the work of social scientist Charles Perrow, particularly his 1984 book Normal Accidentsthe authors' body of examples is relatively fresh, if sometimes not so well remembered todaye.g., the journalistic crimes of Jayson Blair, made possible by a complex accounting system that just begged to be gamed.Programmers, social engineers, and management consultants are among the many audiences for this useful, thought-provoking book. Copyright Kirkus Reviews, used with permission.

Copyright (c) Kirkus Reviews, used with permission.

I.     It was a warm Monday in late June, just before rush hour. Ann and David Wherley boarded the first car of Metro Train 112, bound for Washington, DC, on their way home from an orientation for hospital volunteers. A young woman gave up her seat near the front of the car, and the Wherleys sat together, inseparable as they had been since high school. David, sixty-two, had retired recently, and the couple was looking forward to their fortieth wedding anniversary and a trip to Europe.     David had been a decorated fighter pilot and Air Force officer. In fact, during the 9/ 11 attacks, he was the general who scrambled fighter jets over Washington and ordered pilots to use their discretion to shoot down any passenger plane that threatened the city. But even as a commanding general, he refused to be chauffeured around. He loved taking the Metro.     At 4:58 p.m., a screech interrupted the rhythmic click-clack of the wheels as the driver slammed on the emergency brake. Then came a cacophony of broken glass, bending metal, and screams as Train 112 slammed into something: a train inexplicably stopped on the tracks. The impact drove a thirteen-foot-thick wall of debris--a mass of crushed seats, ceiling panels, and metal posts--into Train 112 and killed David, Ann, and seven others.     Such a collision should have been impossible. The entire Washington Metro system, made up of over one hundred miles of track, was wired to detect and control trains. When trains got too close to each other, they would automatically slow down. But that day, as Train 112 rounded a curve, another train sat stopped on the tracks ahead--present in the real world, but somehow invisible to the track sensors. Train 112 automatically accelerated; after all, the sensors showed that the track was clear. By the time the driver saw the stopped train and hit the emergency brake, the collision was inevitable.      As rescue workers pulled injured riders from the wreckage, Metro engineers got to work. They needed to make sure that other passengers weren't at risk. And to do that, they had to solve a mystery: How does a train twice the length of a football field just disappear?   II. Alarming failures like the crash of Train 112 happen all the time. Take a look at this list of headlines, all from a single week:   CATASTROPHIC MINING DISASTER IN BRAZIL   ANOTHER DAY, ANOTHER HACK: CREDIT CARD STEALING MALWARE HITS HOTEL CHAIN   HYUNDAI CARS ARE RECALLED OVER FAULTY BRAKE SWITCH   STORY OF FLINT WATER CRISIS, "FAILURE OF GOVERNMENT," UNFOLDS IN WASHINGTON   "MASSIVE INTELLIGENCE FAILURE" LED TO THE PARIS TERROR ATTACKS   VANCOUVER SETTLES LAWSUIT WITH MAN WRONGFULLY IMPRISONED FOR NEARLY THREE DECADES   EBOLA RESPONSE: SCIENTISTS BLAST "DANGEROUSLY FRAGILE GLOBAL SYSTEM"   INQUEST INTO MURDER OF SEVEN- YEAR-OLD HAS BECOME SAGA OF THE SYSTEM'S FAILURE TO PROTECT HER   FIRES TO CLEAR LAND SPARK VAST WILDFIRES AND CAUSE ECOLOGICAL DISASTER IN INDONESIA   FDA INVESTIGATES E. COLI OUTBREAK AT CHIPOTLE RESTAURANTS IN WASHINGTON AND OREGON   It might sound like an exceptionally bad week, but there was nothing special about it. Hardly a week goes by without a handful of meltdowns. One week it's an industrial accident, another it's a bankruptcy, and another it's an awful medical error. Even small issues can wreak great havoc. In recent years, for example, several airlines have grounded their entire fleets of planes because of glitches in their technology systems, stranding passengers for days. These problems may make us angry, but they don't surprise us anymore. To be alive in the twenty-first century is to rely on countless complex systems that profoundly affect our lives--from the electrical grid and water treatment plants to transportation systems and communication networks to healthcare and the law. But sometimes our systems fail us.     These failures--and even large-scale meltdowns like BP's oil spill in the Gulf of Mexico, the Fukushima nuclear disaster, and the global financial crisis--seem to stem from very different problems. But their underlying causes turn out to be surprisingly similar. These events have a shared DNA, one that researchers are just beginning to understand. That shared DNA means that failures in one industry can provide lessons for people in other fields: dentists can learn from pilots, and marketing teams from SWAT teams. Understanding the deep causes of failure in high-stakes, exotic domains like deepwater drilling and high-altitude mountaineering can teach us lessons about failure in our more ordinary systems, too. It turns out that everyday meltdowns--failed projects, bad hiring decisions, and even disastrous dinner parties--have a lot in common with oil spills and mountaineering accidents. Fortunately, over the past few decades, researchers around the world have found solutions that can transform how we make decisions, build our teams, design our systems, and prevent the kinds of meltdowns that have become all too common.     This book has two parts. The first explores why our systems fail. It reveals that the same reasons lie behind what appear to be very different events: a social media disaster at Starbucks, the Three Mile Island nuclear accident, a meltdown on Wall Street, and a strange scandal in small-town post offices in the United Kingdom. Part One also explores the paradox of progress: as our systems have become more capable, they have also become more complex and less forgiving, creating an environment where small mistakes can turn into massive failures. Systems that were once innocuous can now accidentally kill people, bankrupt companies, and jail the innocent. And Part One shows that the changes that made our systems vulnerable to accidental failures also provide fertile ground for intentional wrongdoing, like hacking and fraud.     The second part--the bulk of the book--looks at solutions that we can all use. It shows how people can learn from small errors to find out where bigger threats are brewing, how a receptionist saved a life by speaking up to her boss, and how a training program that pilots initially dismissed as "charm school" became one of the reasons flying is safer than ever. It examines why diversity helps us avoid big mistakes and what Everest climbers and Boeing engineers can teach us about the power of simplicity. We'll learn how film crews and ER teams manage surprises--and how their approach could have saved the mismanaged Facebook IPO and Target's failed Canadian expansion. And we'll revisit the puzzle of the disappearing Metro train and see how close engineers were to averting that tragedy. Excerpted from Meltdown: Why Our Systems Fail and What We Can Do about It by Chris Clearfield, András Tilcsik All rights reserved by the original copyright owners. Excerpts are provided for display purposes only and may not be reproduced, reprinted or distributed without the written permission of the publisher.