Barry Shore Of Global Project Strategy
Instructions:
Write a page with resources in APA format. Think back to a project that you have been involved with at work (preferably one that followed a structured process). If you have not been involved with a project at work, select a project from this list to evaluate:
Barry Shore of Global Project Strategy. (2009).attached below:
Evaluate the project based on accepted effectiveness theories and describe:
Goal/Team/Resources/Timeline
Challenges in implementation
Successes in Implementation
Overall success (include how you would measure this)
Lessons learned/what you do differently?
Why Do Projects Fail? Project failure rates are certainly cause for concern, but consider that more and more organizations are adopting a project-based model of organization, called PBO, and it is not surprising to find that addressing failures and learning from them has become increasingly important (Eden, Ackermann, & Williams, 2005; Gray & Larson, 2006; Hyvari, 2006; Robertson & Williams, 2006; Thiry & Deguire, 2007).
Failures occur despite the fact that we have significantly improved the process of planning, executing, and controlling projects. Two contributions would include the Project Management Institute’s (PMI’s) A Guide to the Project Management Body of Knowledge (PMBOK® Guide) (2004) and the literature on critical success factors (CSFs) (Cooke-Davies, 2002; Fortune & White, 2006; Hyvari, 2006; Pinto & Slevin, 1987; Sutterfield, Friday-Stroud, & Shivers-Blackwell, 2006).
To help us understand how projects fail, it may be useful to classify the approaches represented by the PMBOK® Guide, Capability Maturity Model Integration (CMMI), Earned Value Management (EVM), Critical Chain Project Management (CCPM), and CSFs as the Rational Expectation view of project management. They assume that project leaders follow a rational and consistent approach to project management and strive to achieve spe- cific organizational goals (Bazerman, 1994; Beach & Connolly, 2005). It is a view that emphasizes what “should” be done. Argyris (1999) referred to this as the “espoused” theory of individuals and organizations.
There is, however, another view, and it focuses on the way in which indi- viduals within an organization actually behave and make decisions. Borrowing from the work of Simon (1955) and Tversky and Kahneman (1974, 1981), it can be classified as the “behavioral” view of project management. It emphasizes what individuals and groups “actually” do and how managers make decisions involving values and risk preferences (Bazerman, 1994). Argyris (1999) called this the “theory-in-practice.”
This article focuses on the behavioral view of project management and how an understanding of systematic biases—those common to the human decision-making process—can prove useful in diagnosing project failure. By studying these systematic biases, we can learn how decision makers respond to ambiguity, complexity, and uncertainty, as well as how their own particu- lar psychological processes influence project decision making (Schwenk, 1984). From this behavioral view we can learn more about why management approves an overly ambitious scope, why communications between teams is limited, why a manager might ignore signs that the project is going badly, or why a manager discourages the participation of a wider constituency in the project management process.
The article begins with a framework for analyzing project outcomes, introduces the systematic biases commonly associated with decision processes, briefly summarizes eight project failures, uses these biases as a diagnostic tool in understanding how these projects failed, and develops an approach that links these biases to the project culture of failed projects. The
Systematic Biases and Culture in Project Failures Barry Shore, Whittemore School of Business and Economics, University of New Hampshire, Durham, NH, USA
ABSTRACT �
Project success rates have improved, and much of the credit can be given to the knowledge, practices, and standards that have contributed to the professionalization of the field. Unfortunately, too many failures still occur. Because many of them can be traced to man- agement and decision-making practices, it might be useful at this stage to explore a set of systematic biases to determine if understand- ing them can help diagnose and perhaps even prevent failures from occurring. This article begins with a framework identifying the influ- ences on project outcomes, defines the systematic biases that may derail projects, summarizes eight project failures, uses the framework to diagnose those failures, and con- cludes by suggesting how organizational and project culture may contribute to these very common and natural biases.
KEYWORDS: project failure; project cul- ture; systematic biases; project success
Project Management Journal, Vol. 39, No. 4, 5–16
© 2008 by the Project Management Institute
Published online in Wiley InterScience
(www.interscience.wiley.com)
DOI: 10.1002/pmj.20082
P A
P E
R S
6 December 2008 � Project Management Journal � DOI: 10.1002/pmj
Systematic Biases and Culture in Project Failures P
A P
E R
S
article concludes with two examples of how organizations have limited the damage from systematic biases.
The Interaction of Cultural, Leadership, Project, Management, and Behavioral Factors on Project Outcomes The outcome of a project can be related to the influence of cultural, leadership, project, management, and behavioral factors. These relationships are summa- rized in Figure 1. National culture can be defined as the values and belief sys- tems held by a group of individuals, learned early in life, and difficult to change (Hofstede, 1997). Given the international reach of an increasing number of projects, a contemporary view of project management must acknowledge the influence of national culture on the management of projects (Shore & Cross, 2005; Wang & Liu, 2007).
Organizational culture develops within the context of national culture and executive leadership. It can be defined as the shared perceptions of organizational work practices within organizational units (Hofstede, 1999). It also represents the particular ways of conducting organizational business and is instrumental in establishing the com- petence of the organization (Belassi, Kondra, & Tukel, 2007; Schein, 1985; van den Berg & Wilderom, 2004; van Marrewijk, 2007). While executive lead- ership shapes the culture of the organi- zation, project leadership shapes project culture (Turner & Müller, 2006).
Project culture is then the shared perceptions of project work practices, influenced by both the project leader and organizational culture. It is charac- terized by the way in which project planning, execution, and control are exercised.
The systematic biases, common to human decision processes and to be addressed in this article, influence management and team decisions, which in turn influence the planning, execution, and control of the project process.
Methodology The first step in this study was to iden- tify and define the systematic biases that have been studied in the decision literature (Bazerman, 1994; Beach & Connolly, 2005; Hammond, Keeney, & Raiffa, 2006; Keil, Depledge, & Rai, 2007; Tversky & Kahneman, 1974). These biases, defined in the next sec- tion, include the following: • available data, • conservatism, • escalation of commitment, • groupthink, • illusion of control, • overconfidence, • recency, • selective perception, and • sunk cost.
Establishing a clear distinction between them is difficult. For example, Keil et al. (2007) contended that selec- tive perception plays an important role in escalation of commitment. Langer’s (1975) illusion of control overlaps with overconfidence (Russo & Schoemaker, 1989). While the apparent overlap in the definition of these biases is prob- lematic, they have still proven useful in studying failures (Keil et al., 2007; Roberto, 2002).
In the second step of this study, the following project failures are briefly summarized: • Airbus 380, • Coast Guard Maritime Domain
Awareness Project, • Columbia Shuttle, • Denver Baggage Handling, • Mars Climate Orbiter and Mars Polar
Lander, • Merck Vioxx, • Microsoft Xbox 360, and • NYC Police Communications System.
Project Culture
Organizational Culture
Project Outcome
Project Planning Execution and
Control Processes
National Culture
Systematic Biases
Project Goals
Budget Schedule
Complexity
Management and Team Decision
Processes
Project Standards PMBOK ® Guide
Executive Leadership
Project Leadership
Figure 1: Influence of cultural, leadership, project, management, and behavioral factors on project outcome.
Data for these projects was obtained from public and government sources. Case studies were written for each failure (Siggelkow, 2007). Twenty- two business professionals attending a graduate program in “Management of Technology” discussed the nine sys- tematic biases. The participants were then divided into five groups and pre- sented with summaries of the eight cases. None of these professionals was employed by the organizations includ- ed in the study. Using a modified Delphi Method, each group was asked to read the cases and reach consensus on the systematic biases that could best explain why the projects failed (Skulmoski, Hartman, & Krahn, 2007). Finally, each of the five groups present- ed their results and a discussion fol- lowed, during which consensus for the group as a whole was reached.
Systematic Biases Systematic biases represent common distortions in the human decision- making process. They reflect a particu- lar point of view that may be contrary to rational thought. Further, they are systematic in contrast to random errors that, on average, cancel each other out (Bazerman, 1994; Beach & Connolly, 2005). These biases are summarized in Table 1.
Project Failures This section briefly summarizes the full version of the eight case studies pre- sented to the 22 participants. At the end of each case is a summary of the con- sensus reached by the entire group.
Airbus A380 Airbus was founded in 1970 as a loose consortium of 16 independent aero- space companies with facilities in France, Germany, Britain, and Spain. In 2000, Airbus started the A380 project, the goal of which was to design and manufacture a superjumbo jet capable of carrying up to 800 passengers. The aircraft was to usher in a new era of travel.
In the fall of 2006, when the aircraft was in the assembly stage at Toulouse,
France, a preassembled wiring harness produced in the Hamburg, Germany, plant failed to fit into the airframe. The problem, according to several press reports, was that the wiring har- ness had been designed in Hamburg using an older version of CATIA, soft- ware commonly used in aircraft design. The assembly plant in Toulouse, how- ever, used the most recent version of the software. Unfortunately, the ver- sions were incompatible, and the ability to share design specifications between these two plants was compromised. As
a result, hundreds of miles of cabin wiring failed to fit. There was no choice but to halt production, postpone deliv- eries of the aircraft for two years, and redesign the wiring system. Not only was the cost expected to exceed $6 billion, but it also placed the program two years behind schedule. When this delay was announced, the stock lost one-third of its value. Worse, the copresident of the company was accused in June 2008 of selling his stock before the problems were made public.
December 2008 � Project Management Journal � DOI: 10.1002/pmj 7
Table 1: Summary of systematic biases.
Systematic Bias Definition
Available data A data-collection process that is restricted to data that is readily or conveniently available (Bazerman, 1994)
Conservatism Failure to consider new information or negative feedback (Beach & Connolly, 2005)
Escalation of commitment to Additional resources allocated to a project that a failing course of action is increasingly unlikely to succeed. (Keil &
Montealegre, 2000; Keil et al., 2007; Schwenk, 1984; Staw, 1981)
Groupthink Members of a group under pressure to think alike, and to resist evidence that may threaten their view (Haslam, 2004; Haslam et al., 2006; Janis, 1971)
Illusion of control When decision makers conclude that they have more control over a situation than an objective evaluation of the situation would suggest (Langer, 1975; Martz, Neil, & Biscaccianti, 2003)
Overconfidence Level of expressed confidence that is unsupported by the evidence (Bazerman, 1994; Fischoff, Slovic, & Lichtenstein, 1977; Russo & Schoemaker, 1989; Schwenk, 1984)
Recency Disproportionate degree of emphasis placed on the most recent data (Beach & Connolly, 2005; Chan, 1995)
Selective perception The situation where several people perceive the same circumstances differently; varies with the ambiguity of the problem or task (Dearborn & Simon, 1958; Russo & Schoemaker, 1989)
Sunk cost The inability to accept that costs incurred earlier can no longer be recovered and should not be considered a factor in future decisions (Beach & Connolly, 2005; Staw & Ross, 1987)
8 December 2008 � Project Management Journal � DOI: 10.1002/pmj
Systematic Biases and Culture in Project Failures
What had been apparent for a long time, and confirmed in Business Week and the Wall Street Journal, was that Airbus had failed to transform itself from a balkanized organization into an integrated company, and as a result suf- fered from a convoluted management structure. Managers, moreover, acted to protect their former constituency and made political rather than economic decisions (Gauthier-Villars & Michaels, 2007; Matlack, 2006).
When the research groups were presented with this case, they conclud- ed that several systematic biases could be identified, including selective per- ception, groupthink, illusion of control, and availability bias.
Selective perception occurs when a problem is viewed from a limited or narrow frame. This, the group conclud- ed, explained why the organization was unable to move toward an integrated enterprise and why, as a practical example, top managers acted to protect their former constituents, unaware that incompatible software would jeopard- ize the project.
Second, the research groups sug- gested that the insular nature of the separate organizations created a proj- ect culture that limited communication among these units. It created, they maintained, a breeding ground for groupthink, where each group was iso- lated from others in the organization. As a result, many of the practices and procedures within these organizational units went unchallenged.
Third, while management at the central facilities in Toulouse initially envisioned a centralized organization with some control over its divisions, the inability of these divisions to use the same version of CATIA suggested that this transition was far from effective. Senior management, the research groups concluded, suffered from the illusion of control.
The fourth explanation was attrib- uted to availability bias, in which management at each of the plants was limited to the data that was available to
them, and thereby concluded that the project was meeting its local objectives. From their perspective, they were on schedule and within budget.
Coast Guard Maritime Domain Awareness Project In 2001, the U.S. Department of Homeland Security and the U.S. Coast Guard undertook a project that prom- ised to create the maritime equivalent of an air traffic control system. It was a project that would combine the use of long-range surveillance cameras, radar systems, and information technology to automatically identify vessels as they entered U.S. ports. It was also a project that would span 24 federal agencies responsible for the protection of water- ways and coasts.
In the first phase of the project, a complete surveillance system, called Project Hawkeye, was to be developed and implemented for the port of Miami. It would track larger vessels with radar, and smaller vessels, even jet skis, with infrared cameras. Finally, a software system would process the data to deter- mine which vessels posed a security threat and deserved closer scrutiny.
The first test of the system was declared a failure (Lipton, 2006). The cameras were ineffective in tracking the small boats, the radar system proved unreliable when it incorrectly identi- fied waves as boats, the Automated Identification System used for large boats failed to meet its objectives, and the software systems needed to make sense of the data had yet to be installed. Although some data from the system was available to the Coast Guard, they were unable to use it. Because the test failed, the implementation of this sys- tem in 35 ports was delayed until at least 2014.
When the research groups present- ed their conclusions, they focused on the complexity of working with 24 agencies. They concluded that since so many components of the system had failed, project leaders succumbed to the illusion of control bias; these lead-
ers assumed that they had control over the agencies and subcontractors, when in fact they did not. Control issues, the research groups continued, should have been resolved before the project was undertaken, not after.
The groups also suggested that selective perception contributed to the failure. Each separate agency focused only on its immediate task, with appar- ently little effort directed at integrating their role with that of others; there was no suggestion of a strategic relationship among vendors and agencies.
Columbia Shuttle On February 1, 2003, seven astronauts perished when their Columbia Shuttle disintegrated as it re-entered the earth’s atmosphere. During launch, a piece of foam insulation, similar in composition to a Styrofoam cup and about the size of a briefcase, broke away from the main propellant tank. The foam struck the left wing, seriously breaching the protective panels on its leading edge (Gehman, 2003).
It was not the first time that a sec- tion of foam had broken away during launch. In fact, it had happened on every previous flight. But on each of these flights, the spacecraft reentered the earth’s atmosphere without incident and safely returned home. Management assumed that it was a problem of minor significance and that it did not increase the risk level of the flight (Starbuck & Farjoun, 2005).
Many concluded, certainly just after the 2003 tragedy occurred, that technology was to blame. But a more thorough and comprehensive investi- gation, undertaken by the Columbia Accident Investigation Board (CAIB), concluded differently. It maintained that management was as much to blame for the failure as was the foam strike. The Board described a culture where, at every juncture, program man- agers were resistant to new informa- tion. It was a culture where people were unwilling to speak up, or if they did speak up, they were convinced they
P A
P E
R S
would not be heard. They also conclud- ed that the organizational failure was a product of NASA’s history, culture, and politics (Columbia Accident Investiga- tion Board, 2003).
The study participants concluded that NASA had created a culture in which systematic biases went unchecked. First, the participants identified the recency effect. Foam insulation had broken away on previous flights and caused no harm. To them this was confirmation that these recent events distorted the real danger presented by this problem.
Conservatism was also suggested, because the data from these previous flights was largely ignored by senior managers; they failed to revise their prior belief that the system was operat- ing properly. There was also evidence of overconfidence. During the flight, engi- neers, concerned that the foam strike may have caused a problem, asked a manager of the Mission Management Team (MMT) to request satellite imagery of the spacecraft. Management, howev- er, was apparently confident that there was no safety issue, and a decision was made against imagery. Had the imagery been authorized, and the damage dis- covered, the conjecture is that a rescue attempt would have had a reasonable chance of success.
Selective perception was suggested, since management of the shuttle pro- gram had shifted from an engineering focus to a managerial focus. This mor- phed the organization in such a way that engineering problems were less likely to be recognized and more likely to be dominated by schedules and budgets (Gehman, 2003).
Denver Baggage Handling The new airport in Denver, with a budget of $4.9 billion and originally scheduled for completion in October 1993, would be one of the nation’s largest public works projects of the 1990s (Brooke, 1995). It would cover 53 square miles and include five runways, with future expansion to 12 runways. Due to its size and the necessity to
move baggage quickly between flights, the airport would feature a completely automated baggage-handling system.
In April 1995, after many delays, the baggage system project was completed. Reporters were invited to attend a demonstration but instead witnessed a public disaster. Delivery carts were derailed, luggage was torn, and piles of clothes and other personal items were strewn beneath the tracks (Myerson, 1994). After scaling back the scope of the baggage system and making the necessary design changes, the airport finally opened, 16 months behind schedule and almost $2 billion over budget (Keil & Montealegre, 2000).
The baggage-handling project at Denver was more complex than any- thing that had been attempted before at any airport. Luggage was to be first loaded onto conveyor belts, much as it is in conventional baggage-handling sys- tems. These moving conveyors would then deposit the luggage into moving computer-controlled carts at precisely the right moment. The luggage would then travel at 17 miles per hour to its des- tination, as much as one mile away. This underground rail system would be com- pletely automated and would include 4,000 baggage carts traveling throughout the airport and under the control of 100 computers. It would be capable of pro- cessing up to 1,400 bags per minute.
After the system failed its public test, and after design changes were implemented, the system still had problems. Only United Airlines used it, and then only for outgoing flights. Other carriers turned to a hastily constructed manual system, since no contingent plans had been made should the automated system fail. Finally, in 2005, after a decade of frus- trating attempts to solve its problems, the system was abandoned. Under the lease agreement, United Airlines, one of the major stakeholders in the proj- ect, would still be liable for $60 million per year for 25 years.
In two papers that raised concerns about the scope and feasibility of the
project, deNeufville (1994, 2000) contended that this baggage-handling system represented an enormous technological leap over current practices. He concluded that the prob- lem of accommodating the variable demand made on the system, charac- terized in the literature as a classic line- balancing problem, would be difficult to solve.
After discussing the case, the research group identified overconfi- dence as a major factor in the failure. They referenced a quote in the case summary taken from the New York Times: “While the airport was being designed, United insisted that the air- port have the fancier baggage handling system, which it contended would sharply reduce delays” (Johnson, 1994). Overconfidence was also suggested in another quote from the New York Times (Myerson, 1994), in which Gene Di Fonso, president of BAE, the prime contractor for the project, declared, “Who would turn down a $193 million contract? You’d expect to have a little trouble for that kind of money.” With widespread support, the group con- cluded that no one questioned whether it could be done.
They also identified the sunk cost trap. In spite of years of disappoint- ments, when all the airlines, with the exception of United, opted out and used a manual backup baggage- handling system, the project continued. Both the City of Denver and United Airlines had already incurred high costs and were unwilling to disregard these past expenditures, even as their prob- lems persisted and grew worse.
The illusion of control was also identified for its role in keeping the project alive too long. It helped explain why, after evidence was presented at the beginning, a line-balancing prob- lem of this magnitude was very difficult to solve, and why, after an embarrass- ing preview of the system to reporters, management was still confident that it could fix the problems and control the outcome.
December 2008 � Project Management Journal � DOI: 10.1002/pmj 9
10 December 2008 � Project Management Journal � DOI: 10.1002/pmj
Systematic Biases and Culture in Project Failures
Mars Climate Orbiter and Mars Polar Lander As part of the NASA Mars Surveyor Program, the Mars Climate Orbiter was to orbit Mars and collect environmen- tal and weather data. But as the spacecraft approached its destination, telemetry signals fell silent, and a $125 million mission failed.
The root cause identified by NASA was the failure to convert between met- ric and English units. When the fatal error was detected, Noel Hinners, vice- president for flight systems at Lockheed, the company that built the spacecraft, said in disbelief, “It can’t be something that simple that could cause this to hap- pen” (Pollack, 1999). But it was.
Apparently, Lockheed had used pounds during the design of the engines, while NASA scientists, respon- sible for the operation and flight, thought the data was in metric units.
There were early signs during its flight that something was wrong with the craft’s trajectory, and an internal review later confirmed that it may have been off course for months (Oberg, 1999; Pollack, 1999). Project culture, however, required that engineers prove that something was wrong rather than “prove that everything was right.” This difference in perspective prevented the team from looking into the problem. Edward Weiler, NASA associate adminis- trator for space science, said, “The prob- lem here was not the error; it was the fail- ure of NASA’s systems engineering, and the checks and balances in our processes to detect the error” (Oberg, 1999, p. 34).
The Mars Investigation Panel report identified several contributing factors to the failure: the system engineering process did not adequately address the transition from development (Lockheed) to operations (NASA); inadequate com- munications between project elements; and inadequate staffing and training.
Within a few months of the Orbiter failure, the Mars Polar Lander, a related NASA project with a price tag of $165 million, suffered the same fate. Its flight was uneventful until it began its landing
approach. Then, during its descent to the rough terrain of the polar cap, telemetry signals fell silent. With no data to pinpoint the precise cause of failure, the teams investigating the accident speculated that the vehicle’s descent engines prematurely shut down. Unable to slow the descent, the speculation was that the engines quit when the Lander was 130 feet high, crashing into the sur- face of Mars at about 50 miles per hour. The inappropriate response of its engines was attributed to software glitches (Leary, 2000).
The prevailing culture at NASA of “Better, Faster, and Cheaper,” which defined the period when these projects were in development, has been high- lighted many times as the contributing factor behind these failures. Thomas Young, a former NASA official, said that they were trying “to do too much with too little.” He continued, “No one had a sense of how much trouble they were actually in” (Broad, 1999).
The prevailing culture was best expressed in an internal memo written by a laboratory official at the Jet Propulsion Lab: “There might have been some overconfidence, inadequate robustness in our processes, designs or operations, inadequate modeling and simulation of operations, and failure to heed early warnings” (Oberg, 1999, p. 35).
While the trajectory problem asso- ciated with the Orbiter and the engine ignition problem associated with the Lander could be characterized as tech- nical, the Mars Climate Orbiter Failure Board Report (2000) said that manage- ment failures were also to blame. They found that these projects suffered from a lack of senior management involve- ment and too much reliance on inexpe- rienced project managers. The Board also criticized the strategy where proj- ect managers in one organization (Lockheed) were responsible for devel- opment and a separate organization (NASA) was responsible for operations after launch.
The study group first identified the sunk cost trap. If the orbiter did not
launch on schedule, it would have to wait several months before its next opportunity to launch. With launch windows far apart, and with budgets unable to tolerate a substantial delay, managers were under pressure to meet the deadline; it was important not to “waste” the effort put into the project to that point.
Selective perception bias was iden- tified and used to explain why the engi- neers at the Jet Propulsion Lab, the design team, failed to coordinate with the operational team at NASA. In large- scale complex projects such as the Orbiter and Lander, with countless activities, contractors, and suppliers, it is very possible that teams may take a narrow view of their own activities. The risk is that the work of one team may be incompatible with the work of another.
Conservatism, the group contin- ued, explained why engineers failed to take action when they noticed that the trajectory of the spacecraft was off. They even held a meeting in Denver to address the issue, but it was never resolved. Even as the spacecraft approached its destination and data showed that it was drifting off course, controllers largely ignored the real data and assumed it was on course (Oberg, 1999).
Merck Vioxx In 2000, the New England Journal of Medicine published an article suggest- ing that Merck misrepresented clinical trial data on the risks of Vioxx, a drug used to treat arthritis pain. Suspicions were raised again when the Journal of the American Medical Association, pub- lished a paper in 2001 finding that those who took Vioxx were more than five times more likely to experience a cardiac event than those taking a commonly used over-the-counter anti- inflammatory drug, Naproxen. Merck denied these claims, insisting the find- ings were “flawed” (Topol, 2004). Then, under increasing pressure, they revised the Vioxx label in 2002 to reflect these added risks.
P A
P E
R S
During this period, Merck had undertaken a separate study focusing on the use of the drug in treating colon polyps. New data from this trial simply confirmed the risks that had been raised earlier.
Shortly thereafter, on September 30, 2004, five years after it had been introduced to the market, after 84 mil- lion people had taken the drug, and after three years of denying the drug could induce heart attacks and strokes, it was pulled from the shelves (Topol, 2004).
The legal consequences were signif- icant. Over 27,000 claims were filed contending that certain incriminating data were withheld during the FDA review process. In an early trial, a New Jersey jury ruled unanimously in March 2007 that Merck committed consumer fraud by intentionally suppressing, concealing, or omitting information on the risks of Vioxx. Eventually, Merck proposed an out-of-court settlement to the remaining complainants at a cost of over $5 billion (Berenson, 2007).
The study group identified organi- zation and project culture as important contributors to the Vioxx project fail- ure. They cited a Business Week article contending that Richard Clark, CEO, had watched the company degenerate into a “collection of fiefdoms” more focused on their own agendas than on the company’s agenda (Weintraub, 2007).
Financial pressures, the group con- tended, also shaped the culture. Drug discovery is a costly and lengthy process, fraught with risk. The average cost to bring a drug to market exceeds $1 billion. As drug trials proceeded from animal to human trials and even- tually to FDA review, it was not unrea- sonable to conclude that the pressure to continue with the project increased as investment increased.
These cultural problems, together with financial pressures, continued, creating a breeding ground for system- atic biases to emerge. The sunk cost trap was identified as the dominant
bias. After incurring nearly $1 billion to develop the drug, after generating $2.5 billion in sales during 2003, it was not difficult to understand why the compa- ny resisted pressure to remove Vioxx from the market.
While the sunk cost trap dominat- ed, conservatism was also identified as contributing to the failure, because Merck suppressed early data suggesting that the drug could have serious and sometimes tragic side effects.
Microsoft Xbox 360 When Microsoft rushed its Xbox video game console to market in November 2005, it had a one-year advantage over Sony and Nintendo. By 2007, Microsoft had sold over 11.6 million units at $279 to $479, depending on configuration.
Unresolved issues plagued the proj- ect from the beginning. When journal- ists and reviewers were invited to try the Xbox 360 in 2005, before it became available on store shelves, they encountered problems connecting it to the Internet (Croal, 2007). Shortly after it was introduced to the public, users complained that that the console dam- aged game disks, so much that they could no longer be used (Cliff, 2007). In 2005, Microsoft recalled the power cords, concerned that they posed a fire hazard (Wolverton & Takahashi, 2007). Then, in December 2006, in an appar- ent response to these and other issues, Microsoft extended the warranty from 90 days to one year.
But problems persisted. Blogs and forums complained about the “Red Ring of Death,” referring to a string of three lights that illuminate on the con- sole when a serious malfunction occurs. One survey found that the return rate was 33% (Cliff, 2007).
Then, in July 2007, Robbie Bach, president of Microsoft’s Entertainment and Devices Division, said that there had been an “unacceptable high num- ber of repairs” (Taub, 2007). Shortly thereafter, Microsoft announced an extension of the warranty from one to three years at an expected cost of
$1 billion. This represented about $100 for every Xbox sold since its introduc- tion in 2005.
Later in the same month Microsoft announced that its top gaming execu- tive, Peter Moore, was leaving the com- pany, but denied that his departure was related to the Xbox’s engineering prob- lems (Wingfield, 2007).
Three systematic biases were iden- tified by the group. The first was con- servatism. In the face of a continuous stream of product returns and cus- tomer complaints, those who were responsible for the project were unwill- ing to acknowledge that the problem was serious, that customer satisfac- tion and loyalty were deteriorating rapidly, that the product needed to be redesigned, and that customer satisfac- tion needed to be addressed.
It was also suggested that manage- ment fell prey to the sunk cost trap. Considerable investment in the prod- uct had already been made, sales were strong, and since the division had yet to turn a profit, there was pressure to con- tinue at any cost. Returning to earlier stages of design, issuing a recall for the defective units, and replacing them with new units were apparently not realistic options.
Because Microsoft declined to com- ment on the exact cause of the prob- lem, which many suspected was tied to either a power cord or component that was overheating, it was concluded that groupthink was also an issue. The only public comment, made by Robbie Bach, president of the Entertainment and Devices Division responsible for the Xbox, was that the company made manufacturing and production changes that should reduce hardware lockups (Taub, 2007). It was suggested by the group that this could be interpreted as protecting the company to prevent exposing its failures.
New York City Subway Communications System In New York City, police officers who worked underground in the city’s
December 2008 � Project Management Journal � DOI: 10.1002/pmj 11
12 December 2008 � Project Management Journal � DOI: 10.1002/pmj
Systematic Biases and Culture in Project Failures
extensive rapid transit subway system were routinely unable to communicate with officers working the streets above- ground. Incompatible systems were at fault. Not that this was a new problem to law enforcement and emergency organizations in New York City. On September 11, 2001, for example, it was not possible for police to communicate with firefighters and warn them that the World Trade Center towers were in jeopardy of collapsing.
As early as the 1990s, preliminary plans for an integrated communication system had been proposed. In 1999, a contract was signed with two firms. The project was scheduled for completion in 2004, with an approved budget of $115 million.
In 2001, a report warned of an inter- ference problem that could jeopardize the ability of the systems to work together. Rather than return to the design stage and study the validity of this concern, subcontractors continued with the project. It was completed in October 2007, but during the trial of the system it became apparent that inter- ference did indeed create serious communication problems. As a result, implementation was halted. Fixing the problem was expected to increase the cost of the project to $210 million.
The group linked the failure of the New York City Subway Communi- cations Project to conservatism, over- confidence, and illusion of control. Conservatism was suggested when the project managers failed to take the interference warning seriously enough to change their plans early in the proj- ect. Overconfidence was also suggested to explain why they ignored the warn- ing: project managers were apparently convinced that the proposed design would work or that all problems could eventually be solved.
Illusion of control was also identi- fied. The project management team presumably believed that they could fix the interference problem later, that somehow they had enough control to assure a successful outcome. Another
explanation is that they believed that the vendor would take responsibility to solve the problem.
Discussion A summary of the biases identified by the 22 participants can be found in Table 2. Four biases were mentioned more frequently than the others. Conservatism, or the failure to consid- er new information, was mentioned for the Columbia, Merck, Microsoft, and New York City Subway projects. Illusion of control was mentioned for the Airbus, Coast Guard, Denver Baggage, and New York City Subway projects. Selective perception was mentioned for the Airbus, Coast Guard, Mars, and Merck projects. Sunk cost was mentioned for the Denver Baggage, Mars, Merck, and Microsoft projects. Both groupthink and over- confidence were mentioned some- what less. Two biases, recency and available data, were mentioned only once, while escalation of commitment was not mentioned at all.
The results from this small sample prevent making conclusive statements about the dominant biases in project management, but the data begins to suggest that conservatism, illusion of control, selective perception, and sunk cost may be more common than the other biases. Whether they were iden- tified more frequently in this study because they were more easily under- stood by the participants or whether they actually contribute more than the others to project failure is difficult to conclude at this juncture. At the other extreme, the study suggests that esca- lation of commitment, available data, and recency are more difficult to iden- tify and may not contribute signifi- cantly to project failure.
It is rather surprising that escalation was not mentioned at all, because the Denver Baggage, Coast Guard, and the New York City Subway projects required additional funding after evidence became available that these projects were in trouble. One possible explana-
P A
P E
R S
Ta bl
e 2:
Su m
m ar
y of
b ia
se s
af fe
ct in
g ea
ch c
as e
st ud
y.
Av ai
la bl
e Es
ca la
tio n
of Ill
us io
n of
Se le
ct iv
e Su
nk Da
ta Co
ns er
va tis
m Co
m m
itm en
t Gr
ou pt
hi nk
Co nt
ro l
Ov er
co nf
id en
ce Re
ce nc
y Pe
rc ep
tio n
Co st
Ai rb
us A
38 0
X X
X X
Co as
t G ua
rd M
ar in
e X
X
Co lu
m bi
a Sh
ut tle
X X
X
De nv
er B
ag ga
ge X
X X
M ar
s Or
bi te
r X
X
M er
ck V
io xx
X X
X
M ic
ro so
ft X
bo x
36 0
X X
X
Ne w
Y or
k Ci
ty S
ub w
ay X
X
December 2008 � Project Management Journal � DOI: 10.1002/pmj 13
literature that focuses on the study of systematic biases. There is also a long tradition in the project management lit- erature that focuses on project failures. The objective of this article has been to determine if bringing these two tradi- tions together could prove useful in learning more about project failures and then in understanding how culture may provide the environment within which these biases may emerge.
Twenty-two professionals partici- pated in the study. They were intro- duced to systematic biases and then asked to determine which of these bias- es could help explain eight failed proj- ects. Their responses suggest that the vocabulary of systematic biases could prove very useful in understanding how the rational processes of project man- agement can be derailed by the human decision-making process.
What this result underscores is that the skills and techniques expressed in the rational view of project manage- ment, regardless of how aggressively they are pursued, may be insufficient to
Several of the dimensions used in van den Berg and Wilderom overlap with the Competing Values Model.
The nine systematic biases used in this article were mapped onto the Competing Values Model. The results are summarized in Table 3. For exam- ple, the available data bias suggests an organization and project culture char- acterized by an internal focus and a concern that external data may lead to unwelcomed changes. Airbus is an example of a case study in which this bias was observed. It can be concluded that the Airbus project culture, at the very least, could be characterized as having a preference for an internal focus and stability. One can hypothe- size from Table 3 that failed projects, in general, can be associated with organi- zational and project cultures character- ized by an internal focus and a prefer- ence for stability, not change.
Summary There is a long tradition in the organiza- tional psychology and decision-making
Cognitive Bias Competing Case in Which Cognitive Values Model Bias Was Observed
Available data Internal focus, stability Airbus
Conservatism Internal focus, stability Columbia, Merck, Microsoft, New York City Subway
Escalation of commitment Internal focus, stability
Groupthink Internal focus, stability Airbus, Microsoft
Illusion of control Internal focus, stability Airbus, Coast Guard, Denver Baggage, New York City Subway
Overconfidence Internal focus, stability Columbia, Denver Baggage
Recency Internal focus, stability Columbia
Selective perception Internal focus, stability Airbus, Coast Guard, Mars, Merck
Sunk cost Stability Denver Baggage, Mars, Merck, Microsoft
Note. The results suggest that the failed projects studied in this article reflect a project culture that can be characterized as having a preference for an internal focus and stability.
Table 3: Cognitive biases mapped onto the competing values model.