• Tweet

  • Post

  • Share

  • Save

  • Get PDF

  • Buy Copies

The wisdom of learning from failure is incontrovertible. However organizations that do information technology well are extraordinarily rare. This gap is not due to a lack of commitment to learning. Managers in the vast majority of enterprises that I have studied over the past twenty years—pharmaceutical, fiscal services, production design, telecommunications, and construction companies; hospitals; and NASA'south space shuttle programme, among others—genuinely wanted to help their organizations acquire from failures to amend time to come functioning. In some cases they and their teams had devoted many hours to subsequently-action reviews, postmortems, and the like. Just fourth dimension after time I saw that these painstaking efforts led to no existent change. The reason: Those managers were thinking virtually failure the wrong way.

Most executives I've talked to believe that failure is bad (of grade!). They besides believe that learning from it is pretty straightforward: Ask people to reflect on what they did wrong and exhort them to avoid similar mistakes in the future—or, better nevertheless, assign a team to review and write a report on what happened and then distribute it throughout the organization.

These widely held beliefs are misguided. First, failure is not always bad. In organizational life it is sometimes bad, sometimes inevitable, and sometimes even good. Second, learning from organizational failures is anything but straightforward. The attitudes and activities required to effectively discover and analyze failures are in brusque supply in virtually companies, and the need for context-specific learning strategies is underappreciated. Organizations need new and better ways to go beyond lessons that are superficial ("Procedures weren't followed") or self-serving ("The market simply wasn't ready for our corking new production"). That means jettisoning old cultural beliefs and stereotypical notions of success and embracing failure'due south lessons. Leaders tin can begin by understanding how the blame game gets in the style.

The Blame Game

Failure and fault are virtually inseparable in about households, organizations, and cultures. Every kid learns at some point that admitting failure ways taking the blame. That is why and so few organizations have shifted to a civilisation of psychological safe in which the rewards of learning from failure can be fully realized.

Executives I've interviewed in organizations as unlike as hospitals and investment banks admit to being torn: How can they respond constructively to failures without giving ascension to an anything-goes mental attitude? If people aren't blamed for failures, what will ensure that they endeavour every bit difficult as possible to do their best work?

This concern is based on a false dichotomy. In actuality, a civilisation that makes it safety to acknowledge and report on failure can—and in some organizational contexts must—coexist with loftier standards for operation. To sympathise why, expect at the exhibit "A Spectrum of Reasons for Failure," which lists causes ranging from deliberate difference to thoughtful experimentation.

Which of these causes involve blameworthy actions? Deliberate deviance, beginning on the list, obviously warrants blame. But inattention might not. If it results from a lack of effort, perhaps it's blameworthy. But if it results from fatigue well-nigh the end of an overly long shift, the director who assigned the shift is more at mistake than the employee. Every bit we go downwards the list, it gets more than and more than hard to notice blameworthy acts. In fact, a failure resulting from thoughtful experimentation that generates valuable information may really be praiseworthy.

When I ask executives to consider this spectrum then to estimate how many of the failures in their organizations are truly blameworthy, their answers are usually in single digits—perchance ii% to five%. Simply when I enquire how many are treated as blameworthy, they say (after a pause or a laugh) seventy% to 90%. The unfortunate event is that many failures get unreported and their lessons are lost.

Not All Failures Are Created Equal

A sophisticated understanding of failure'south causes and contexts volition aid to avert the blame game and institute an effective strategy for learning from failure. Although an space number of things can go wrong in organizations, mistakes fall into three broad categories: preventable, complexity-related, and intelligent.

Preventable failures in predictable operations.

Most failures in this category tin can indeed be considered "bad." They ordinarily involve deviations from spec in the closely defined processes of high-volume or routine operations in manufacturing and services. With proper training and support, employees tin follow those processes consistently. When they don't, deviance, inattention, or lack of ability is commonly the reason. Just in such cases, the causes can be readily identified and solutions developed. Checklists (as in the Harvard surgeon Atul Gawande'southward contempo best seller The Checklist Manifesto) are one solution. Another is the vaunted Toyota Production Organisation, which builds continual learning from tiny failures (small procedure deviations) into its arroyo to improvement. As most students of operations know well, a squad member on a Toyota assembly line who spots a problem or fifty-fifty a potential problem is encouraged to pull a rope called the andon cord, which immediately initiates a diagnostic and trouble-solving procedure. Production continues unimpeded if the problem tin can be remedied in less than a minute. Otherwise, production is halted—despite the loss of acquirement entailed—until the failure is understood and resolved.

Unavoidable failures in circuitous systems.

A large number of organizational failures are due to the inherent uncertainty of work: A particular combination of needs, people, and bug may have never occurred before. Triaging patients in a infirmary emergency room, responding to enemy deportment on the battlefield, and running a fast-growing start-up all occur in unpredictable situations. And in circuitous organizations like shipping carriers and nuclear power plants, system failure is a perpetual risk.

Although serious failures tin can exist averted by following all-time practices for safety and risk management, including a thorough assay of any such events that do occur, minor process failures are inevitable. To consider them bad is not just a misunderstanding of how complex systems piece of work; it is counterproductive. Avoiding consequential failures means apace identifying and correcting modest failures. Most accidents in hospitals result from a serial of small failures that went unnoticed and unfortunately lined upwards in simply the incorrect way.

Intelligent failures at the frontier.

Failures in this category tin can rightly be considered "good," because they provide valuable new noesis that tin can help an organisation leap ahead of the competition and ensure its future growth—which is why the Knuckles University professor of management Sim Sitkin calls them intelligent failures. They occur when experimentation is necessary: when answers are non knowable in advance considering this exact state of affairs hasn't been encountered earlier and perhaps never will be again. Discovering new drugs, creating a radically new concern, designing an innovative production, and testing client reactions in a brand-new market are tasks that crave intelligent failures. "Trial and error" is a common term for the kind of experimentation needed in these settings, but it is a misnomer, because "error" implies that there was a "right" event in the first identify. At the frontier, the right kind of experimentation produces good failures quickly. Managers who practice it can avoid the unintelligent failure of conducting experiments at a larger scale than necessary.

Leaders of the product design firm IDEO understood this when they launched a new innovation-strategy service. Rather than help clients design new products inside their existing lines—a process IDEO had all but perfected—the service would assistance them create new lines that would take them in novel strategic directions. Knowing that it hadn't still figured out how to deliver the service effectively, the visitor started a pocket-sized project with a mattress company and didn't publicly announce the launch of a new business.

Although the project failed—the client did not change its product strategy—IDEO learned from it and figured out what had to be washed differently. For example, information technology hired squad members with MBAs who could better assistance clients create new businesses and made some of the clients' managers function of the squad. Today strategic innovation services account for more than a third of IDEO's revenues.

Tolerating unavoidable process failures in complex systems and intelligent failures at the frontiers of cognition won't promote mediocrity. Indeed, tolerance is essential for any organization that wishes to excerpt the knowledge such failures provide. But failure is still inherently emotionally charged; getting an organization to accept it takes leadership.

Building a Learning Culture

But leaders tin can create and reinforce a culture that counteracts the arraign game and makes people feel both comfortable with and responsible for surfacing and learning from failures. (See the sidebar "How Leaders Can Build a Psychologically Safe Environment.") They should insist that their organizations develop a articulate understanding of what happened—not of "who did information technology"—when things become wrong. This requires consistently reporting failures, pocket-size and large; systematically analyzing them; and proactively searching for opportunities to experiment.

Leaders should as well send the correct message most the nature of the work, such as reminding people in R&D, "We're in the discovery business, and the faster nosotros fail, the faster we'll succeed." I have found that managers often don't empathise or appreciate this subtle but crucial signal. They likewise may arroyo failure in a way that is inappropriate for the context. For example, statistical process command, which uses data analysis to assess unwarranted variances, is not good for catching and correcting random invisible glitches such as software bugs. Nor does it aid in the development of creative new products. Conversely, though great scientists intuitively adhere to IDEO'southward slogan, "Fail often in order to succeed sooner," information technology would hardly promote success in a manufacturing establish.

The slogan "Fail frequently in social club to succeed sooner" would hardly promote success in a manufacturing plant.

Oft i context or i kind of piece of work dominates the culture of an enterprise and shapes how information technology treats failure. For instance, automotive companies, with their anticipated, high-volume operations, understandably tend to view failure as something that tin can and should be prevented. But about organizations engage in all three kinds of work discussed above—routine, circuitous, and frontier. Leaders must ensure that the correct arroyo to learning from failure is applied in each. All organizations acquire from failure through three essential activities: detection, analysis, and experimentation.

Detecting Failure

Spotting large, painful, expensive failures is easy. But in many organizations any failure that can be hidden is hidden as long as information technology's unlikely to crusade immediate or obvious harm. The goal should be to surface it early, before information technology has mushroomed into disaster.

Shortly after arriving from Boeing to take the reins at Ford, in September 2006, Alan Mulally instituted a new system for detecting failures. He asked managers to color code their reports green for good, xanthous for caution, or red for problems—a common management technique. According to a 2009 story in Fortune, at his commencement few meetings all the managers coded their operations green, to Mulally's frustration. Reminding them that the company had lost several billion dollars the previous twelvemonth, he asked straight out, "Isn't anything not going well?" After ane tentative yellow written report was fabricated nigh a serious product defect that would probably delay a launch, Mulally responded to the deathly silence that ensued with applause. After that, the weekly staff meetings were total of color.

That story illustrates a pervasive and cardinal problem: Although many methods of surfacing electric current and pending failures be, they are grossly underutilized. Total Quality Management and soliciting feedback from customers are well-known techniques for bringing to light failures in routine operations. Loftier-reliability-arrangement (HRO) practices help forestall catastrophic failures in circuitous systems like nuclear ability plants through early detection. Electricité de French republic, which operates 58 nuclear ability plants, has been an exemplar in this surface area: It goes beyond regulatory requirements and religiously tracks each plant for anything even slightly out of the ordinary, immediately investigates whatever turns up, and informs all its other plants of any anomalies.

Such methods are not more widely employed considering all besides many messengers—even the most senior executives—remain reluctant to convey bad news to bosses and colleagues. One senior executive I know in a large consumer products visitor had grave reservations well-nigh a takeover that was already in the works when he joined the management squad. Only, overly conscious of his newcomer status, he was silent during discussions in which all the other executives seemed enthusiastic about the plan. Many months subsequently, when the takeover had clearly failed, the squad gathered to review what had happened. Aided past a consultant, each executive considered what he or she might take done to contribute to the failure. The newcomer, openly apologetic about his by silence, explained that others' enthusiasm had made him unwilling to be "the skunk at the picnic."

In researching errors and other failures in hospitals, I discovered substantial differences across patient-intendance units in nurses' willingness to speak up almost them. It turned out that the beliefs of midlevel managers—how they responded to failures and whether they encouraged open discussion of them, welcomed questions, and displayed humility and curiosity—was the crusade. I have seen the same blueprint in a wide range of organizations.

A horrific case in signal, which I studied for more than 2 years, is the 2003 explosion of the Columbia space shuttle, which killed seven astronauts (see "Facing Cryptic Threats," by Michael A. Roberto, Richard M.J. Bohmer, and Amy C. Edmondson, HBR November 2006). NASA managers spent some 2 weeks downplaying the seriousness of a slice of foam's having cleaved off the left side of the shuttle at launch. They rejected engineers' requests to resolve the ambiguity (which could have been done past having a satellite photograph the shuttle or request the astronauts to conduct a space walk to inspect the area in question), and the major failure went largely undetected until its fatal consequences 16 days later. Ironically, a shared merely unsubstantiated belief amongst program managers that there was little they could do contributed to their inability to notice the failure. Postevent analyses suggested that they might indeed have taken fruitful action. Simply conspicuously leaders hadn't established the necessary civilisation, systems, and procedures.

One challenge is educational activity people in an organization when to declare defeat in an experimental form of action. The human being tendency to hope for the best and try to avert failure at all costs gets in the fashion, and organizational hierarchies exacerbate it. As a result, failing R&D projects are often kept going much longer than is scientifically rational or economically prudent. We throw skilful money after bad, praying that we'll pull a rabbit out of a hat. Intuition may tell engineers or scientists that a project has fatal flaws, only the formal conclusion to telephone call it a failure may be delayed for months.

Over again, the remedy—which does not necessarily involve much time and expense—is to reduce the stigma of failure. Eli Lilly has done this since the early 1990s by holding "failure parties" to honor intelligent, high-quality scientific experiments that fail to reach the desired results. The parties don't toll much, and redeploying valuable resources—particularly scientists—to new projects before rather than subsequently tin can save hundreds of thousands of dollars, not to mention kickstart potential new discoveries.

Analyzing Failure

One time a failure has been detected, it's essential to go beyond the obvious and superficial reasons for it to understand the root causes. This requires the discipline—better yet, the enthusiasm—to use sophisticated assay to ensure that the right lessons are learned and the right remedies are employed. The job of leaders is to see that their organizations don't just move on after a failure but stop to dig in and find the wisdom contained in it.

Why is failure analysis ofttimes shortchanged? Because examining our failures in depth is emotionally unpleasant and can chip away at our cocky-esteem. Left to our own devices, near of us volition speed through or avoid failure analysis birthday. Another reason is that analyzing organizational failures requires research and openness, patience, and a tolerance for causal ambivalence. Yet managers typically admire and are rewarded for decisiveness, efficiency, and action—not thoughtful reflection. That is why the right culture is so important.

The challenge is more than emotional; it's cognitive, also. Even without significant to, we all favor show that supports our existing beliefs rather than alternative explanations. Nosotros besides tend to downplay our responsibility and place undue blame on external or situational factors when nosotros fail, only to practice the reverse when assessing the failures of others—a psychological trap known as fundamental attribution error.

My enquiry has shown that failure assay is often express and ineffective—fifty-fifty in complex organizations like hospitals, where human being lives are at pale. Few hospitals systematically analyze medical errors or process flaws in order to capture failure'due south lessons. Recent research in North Carolina hospitals, published in November 2010 in the New England Journal of Medicine, establish that despite a dozen years of heightened awareness that medical errors consequence in thousands of deaths each year, hospitals take not become safer.

Fortunately, at that place are shining exceptions to this pattern, which continue to provide hope that organizational learning is possible. At Intermountain Healthcare, a system of 23 hospitals that serves Utah and southeastern Idaho, physicians' deviations from medical protocols are routinely analyzed for opportunities to better the protocols. Allowing deviations and sharing the data on whether they actually produce a better event encourages physicians to buy into this program. (Come across "Fixing Health Intendance on the Front Lines," past Richard M.J. Bohmer, HBR April 2010.)

Motivating people to become beyond first-order reasons (procedures weren't followed) to agreement the second- and third-order reasons tin exist a major challenge. 1 mode to do this is to utilise interdisciplinary teams with diverse skills and perspectives. Complex failures in detail are the event of multiple events that occurred in dissimilar departments or disciplines or at unlike levels of the organisation. Agreement what happened and how to prevent it from happening once more requires detailed, team-based discussion and analysis.

A team of leading physicists, engineers, aviation experts, naval leaders, and even astronauts devoted months to an assay of the Columbia disaster. They conclusively established not merely the first-order crusade—a slice of foam had striking the shuttle'southward leading edge during launch—but also 2nd-gild causes: A rigid hierarchy and schedule-obsessed culture at NASA made information technology specially difficult for engineers to speak upward about annihilation just the most stone-solid concerns.

Promoting Experimentation

The 3rd disquisitional activity for effective learning is strategically producing failures—in the right places, at the right times—through systematic experimentation. Researchers in basic science know that although the experiments they conduct will occasionally result in a spectacular success, a large percentage of them (seventy% or higher in some fields) will fail. How practise these people become out of bed in the morning? Kickoff, they know that failure is not optional in their piece of work; information technology'due south part of being at the leading edge of scientific discovery. Second, far more than than about of us, they understand that every failure conveys valuable information, and they're eager to get it before the competition does.

In contrast, managers in charge of piloting a new product or service—a classic instance of experimentation in business—typically practice any they tin can to make sure that the airplane pilot is perfect right out of the starting gate. Ironically, this hunger to succeed tin can later inhibit the success of the official launch. Likewise often, managers in charge of pilots design optimal conditions rather than representative ones. Thus the airplane pilot doesn't produce knowledge about what won't work.

Too oftentimes, pilots are conducted nether optimal conditions rather than representative ones. Thus they can't show what won't piece of work.

In the very early days of DSL, a major telecommunications company I'll telephone call Telco did a full-scale launch of that loftier-speed technology to consumer households in a major urban market. It was an unmitigated customer-service disaster. The company missed 75% of its commitments and found itself confronted with a staggering 12,000 late orders. Customers were frustrated and upset, and service reps couldn't even begin to respond all their calls. Employee morale suffered. How could this happen to a leading company with high satisfaction ratings and a brand that had long stood for excellence?

A small and extremely successful suburban airplane pilot had lulled Telco executives into a misguided conviction. The problem was that the pilot did not resemble existent service weather condition: It was staffed with unusually personable, skillful service reps and took identify in a customs of educated, tech-savvy customers. But DSL was a brand-new applied science and, dissimilar traditional telephony, had to interface with customers' highly variable home computers and technical skills. This added complication and unpredictability to the service-delivery challenge in ways that Telco had not fully appreciated before the launch.

A more useful pilot at Telco would have tested the engineering science with limited back up, unsophisticated customers, and old computers. It would accept been designed to discover everything that could become wrong—instead of proving that under the best of weather condition everything would go right. (Meet the sidebar "Designing Successful Failures.") Of grade, the managers in charge would have to have understood that they were going to be rewarded non for success but, rather, for producing intelligent failures as apace as possible.

In short, exceptional organizations are those that go beyond detecting and analyzing failures and try to generate intelligent ones for the express purpose of learning and innovating. Information technology'due south not that managers in these organizations enjoy failure. But they recognize it as a necessary past-product of experimentation. They also realize that they don't have to do dramatic experiments with large budgets. Frequently a small pilot, a dry out run of a new technique, or a simulation volition suffice.

The courage to confront our own and others' imperfections is crucial to solving the apparent contradiction of wanting neither to discourage the reporting of bug nor to create an environment in which anything goes. This means that managers must ask employees to be brave and speak up—and must not respond by expressing anger or strong disapproval of what may at showtime appear to be incompetence. More often than we realize, complex systems are at piece of work backside organizational failures, and their lessons and improvement opportunities are lost when conversation is stifled.

Savvy managers sympathise the risks of unbridled toughness. They know that their ability to find out about and help resolve problems depends on their power to learn about them. Only virtually managers I've encountered in my research, teaching, and consulting work are far more than sensitive to a different run a risk—that an agreement response to failures will simply create a lax work environment in which mistakes multiply.

This common worry should be replaced by a new image—ane that recognizes the inevitability of failure in today'southward circuitous piece of work organizations. Those that catch, right, and learn from failure before others do will succeed. Those that wallow in the blame game will not.

A version of this article appeared in the April 2011 issue of Harvard Business Review.