Loading...

Messages

Proposals

Stuck in your homework and missing deadline? Get urgent help in $10/Page with 24 hours deadline

Get Urgent Writing Help In Your Essays, Assignments, Homeworks, Dissertation, Thesis Or Coursework & Achieve A+ Grades.

Privacy Guaranteed - 100% Plagiarism Free Writing - Free Turnitin Report - Professional And Experienced Writers - 24/7 Online Support

Airbus a380 project failure reasons

03/12/2021 Client: muhammad11 Deadline: 2 Day

P A

P E

R S

December 2008 � Project Management Journal � DOI: 10.1002/pmj 5

Why Do Projects Fail? Project failure rates are certainly cause for concern, but consider that more and more organizations are adopting a project-based model of organization, called PBO, and it is not surprising to find that addressing failures and learning from them has become increasingly important (Eden, Ackermann, & Williams, 2005; Gray & Larson, 2006; Hyvari, 2006; Robertson & Williams, 2006; Thiry & Deguire, 2007).

Failures occur despite the fact that we have significantly improved the process of planning, executing, and controlling projects. Two contributions would include the Project Management Institute’s (PMI’s) A Guide to the Project Management Body of Knowledge (PMBOK® Guide) (2004) and the literature on critical success factors (CSFs) (Cooke-Davies, 2002; Fortune & White, 2006; Hyvari, 2006; Pinto & Slevin, 1987; Sutterfield, Friday-Stroud, & Shivers-Blackwell, 2006).

To help us understand how projects fail, it may be useful to classify the approaches represented by the PMBOK® Guide, Capability Maturity Model Integration (CMMI), Earned Value Management (EVM), Critical Chain Project Management (CCPM), and CSFs as the Rational Expectation view of project management. They assume that project leaders follow a rational and consistent approach to project management and strive to achieve spe- cific organizational goals (Bazerman, 1994; Beach & Connolly, 2005). It is a view that emphasizes what “should” be done. Argyris (1999) referred to this as the “espoused” theory of individuals and organizations.

There is, however, another view, and it focuses on the way in which indi- viduals within an organization actually behave and make decisions. Borrowing from the work of Simon (1955) and Tversky and Kahneman (1974, 1981), it can be classified as the “behavioral” view of project management. It emphasizes what individuals and groups “actually” do and how managers make decisions involving values and risk preferences (Bazerman, 1994). Argyris (1999) called this the “theory-in-practice.”

This article focuses on the behavioral view of project management and how an understanding of systematic biases—those common to the human decision-making process—can prove useful in diagnosing project failure. By studying these systematic biases, we can learn how decision makers respond to ambiguity, complexity, and uncertainty, as well as how their own particu- lar psychological processes influence project decision making (Schwenk, 1984). From this behavioral view we can learn more about why management approves an overly ambitious scope, why communications between teams is limited, why a manager might ignore signs that the project is going badly, or why a manager discourages the participation of a wider constituency in the project management process.

The article begins with a framework for analyzing project outcomes, introduces the systematic biases commonly associated with decision processes, briefly summarizes eight project failures, uses these biases as a diagnostic tool in understanding how these projects failed, and develops an approach that links these biases to the project culture of failed projects. The

Systematic Biases and Culture in Project Failures Barry Shore, Whittemore School of Business and Economics, University of New Hampshire, Durham, NH, USA

ABSTRACT �

Project success rates have improved, and much of the credit can be given to the knowledge, practices, and standards that have contributed to the professionalization of the field. Unfortunately, too many failures still occur. Because many of them can be traced to man- agement and decision-making practices, it might be useful at this stage to explore a set of systematic biases to determine if understand- ing them can help diagnose and perhaps even prevent failures from occurring. This article begins with a framework identifying the influ- ences on project outcomes, defines the systematic biases that may derail projects, summarizes eight project failures, uses the framework to diagnose those failures, and con- cludes by suggesting how organizational and project culture may contribute to these very common and natural biases.

KEYWORDS: project failure; project cul- ture; systematic biases; project success

Project Management Journal, Vol. 39, No. 4, 5–16

© 2008 by the Project Management Institute

Published online in Wiley InterScience

(www.interscience.wiley.com)

DOI: 10.1002/pmj.20082

P A

P E

R S

6 December 2008 � Project Management Journal � DOI: 10.1002/pmj

Systematic Biases and Culture in Project Failures P

A P

E R

S

article concludes with two examples of how organizations have limited the damage from systematic biases.

The Interaction of Cultural, Leadership, Project, Management, and Behavioral Factors on Project Outcomes The outcome of a project can be related to the influence of cultural, leadership, project, management, and behavioral factors. These relationships are summa- rized in Figure 1. National culture can be defined as the values and belief sys- tems held by a group of individuals, learned early in life, and difficult to change (Hofstede, 1997). Given the international reach of an increasing number of projects, a contemporary view of project management must acknowledge the influence of national culture on the management of projects (Shore & Cross, 2005; Wang & Liu, 2007).

Organizational culture develops within the context of national culture and executive leadership. It can be defined as the shared perceptions of organizational work practices within organizational units (Hofstede, 1999). It also represents the particular ways of conducting organizational business and is instrumental in establishing the com- petence of the organization (Belassi, Kondra, & Tukel, 2007; Schein, 1985; van den Berg & Wilderom, 2004; van Marrewijk, 2007). While executive lead- ership shapes the culture of the organi- zation, project leadership shapes project culture (Turner & Müller, 2006).

Project culture is then the shared perceptions of project work practices, influenced by both the project leader and organizational culture. It is charac- terized by the way in which project planning, execution, and control are exercised.

The systematic biases, common to human decision processes and to be addressed in this article, influence management and team decisions, which in turn influence the planning, execution, and control of the project process.

Methodology The first step in this study was to iden- tify and define the systematic biases that have been studied in the decision literature (Bazerman, 1994; Beach & Connolly, 2005; Hammond, Keeney, & Raiffa, 2006; Keil, Depledge, & Rai, 2007; Tversky & Kahneman, 1974). These biases, defined in the next sec- tion, include the following: • available data, • conservatism, • escalation of commitment, • groupthink, • illusion of control, • overconfidence, • recency, • selective perception, and • sunk cost.

Establishing a clear distinction between them is difficult. For example, Keil et al. (2007) contended that selec- tive perception plays an important role in escalation of commitment. Langer’s (1975) illusion of control overlaps with overconfidence (Russo & Schoemaker, 1989). While the apparent overlap in the definition of these biases is prob- lematic, they have still proven useful in studying failures (Keil et al., 2007; Roberto, 2002).

In the second step of this study, the following project failures are briefly summarized: • Airbus 380, • Coast Guard Maritime Domain

Awareness Project, • Columbia Shuttle, • Denver Baggage Handling, • Mars Climate Orbiter and Mars Polar

Lander, • Merck Vioxx, • Microsoft Xbox 360, and • NYC Police Communications System.

Project Culture

Organizational Culture

Project Outcome

Project Planning Execution and

Control Processes

National Culture

Systematic Biases

Project Goals

Budget Schedule

Complexity

Management and Team Decision

Processes

Project Standards PMBOK ® Guide

Executive Leadership

Project Leadership

Figure 1: Influence of cultural, leadership, project, management, and behavioral factors on project outcome.

Data for these projects was obtained from public and government sources. Case studies were written for each failure (Siggelkow, 2007). Twenty- two business professionals attending a graduate program in “Management of Technology” discussed the nine sys- tematic biases. The participants were then divided into five groups and pre- sented with summaries of the eight cases. None of these professionals was employed by the organizations includ- ed in the study. Using a modified Delphi Method, each group was asked to read the cases and reach consensus on the systematic biases that could best explain why the projects failed (Skulmoski, Hartman, & Krahn, 2007). Finally, each of the five groups present- ed their results and a discussion fol- lowed, during which consensus for the group as a whole was reached.

Systematic Biases Systematic biases represent common distortions in the human decision- making process. They reflect a particu- lar point of view that may be contrary to rational thought. Further, they are systematic in contrast to random errors that, on average, cancel each other out (Bazerman, 1994; Beach & Connolly, 2005). These biases are summarized in Table 1.

Project Failures This section briefly summarizes the full version of the eight case studies pre- sented to the 22 participants. At the end of each case is a summary of the con- sensus reached by the entire group.

Airbus A380 Airbus was founded in 1970 as a loose consortium of 16 independent aero- space companies with facilities in France, Germany, Britain, and Spain. In 2000, Airbus started the A380 project, the goal of which was to design and manufacture a superjumbo jet capable of carrying up to 800 passengers. The aircraft was to usher in a new era of travel.

In the fall of 2006, when the aircraft was in the assembly stage at Toulouse,

France, a preassembled wiring harness produced in the Hamburg, Germany, plant failed to fit into the airframe. The problem, according to several press reports, was that the wiring har- ness had been designed in Hamburg using an older version of CATIA, soft- ware commonly used in aircraft design. The assembly plant in Toulouse, how- ever, used the most recent version of the software. Unfortunately, the ver- sions were incompatible, and the ability to share design specifications between these two plants was compromised. As

a result, hundreds of miles of cabin wiring failed to fit. There was no choice but to halt production, postpone deliv- eries of the aircraft for two years, and redesign the wiring system. Not only was the cost expected to exceed $6 billion, but it also placed the program two years behind schedule. When this delay was announced, the stock lost one-third of its value. Worse, the copresident of the company was accused in June 2008 of selling his stock before the problems were made public.

December 2008 � Project Management Journal � DOI: 10.1002/pmj 7

Table 1: Summary of systematic biases.

Systematic Bias Definition

Available data A data-collection process that is restricted to data that is readily or conveniently available (Bazerman, 1994)

Conservatism Failure to consider new information or negative feedback (Beach & Connolly, 2005)

Escalation of commitment to Additional resources allocated to a project that a failing course of action is increasingly unlikely to succeed. (Keil &

Montealegre, 2000; Keil et al., 2007; Schwenk, 1984; Staw, 1981)

Groupthink Members of a group under pressure to think alike, and to resist evidence that may threaten their view (Haslam, 2004; Haslam et al., 2006; Janis, 1971)

Illusion of control When decision makers conclude that they have more control over a situation than an objective evaluation of the situation would suggest (Langer, 1975; Martz, Neil, & Biscaccianti, 2003)

Overconfidence Level of expressed confidence that is unsupported by the evidence (Bazerman, 1994; Fischoff, Slovic, & Lichtenstein, 1977; Russo & Schoemaker, 1989; Schwenk, 1984)

Recency Disproportionate degree of emphasis placed on the most recent data (Beach & Connolly, 2005; Chan, 1995)

Selective perception The situation where several people perceive the same circumstances differently; varies with the ambiguity of the problem or task (Dearborn & Simon, 1958; Russo & Schoemaker, 1989)

Sunk cost The inability to accept that costs incurred earlier can no longer be recovered and should not be considered a factor in future decisions (Beach & Connolly, 2005; Staw & Ross, 1987)

8 December 2008 � Project Management Journal � DOI: 10.1002/pmj

Systematic Biases and Culture in Project Failures

What had been apparent for a long time, and confirmed in Business Week and the Wall Street Journal, was that Airbus had failed to transform itself from a balkanized organization into an integrated company, and as a result suf- fered from a convoluted management structure. Managers, moreover, acted to protect their former constituency and made political rather than economic decisions (Gauthier-Villars & Michaels, 2007; Matlack, 2006).

When the research groups were presented with this case, they conclud- ed that several systematic biases could be identified, including selective per- ception, groupthink, illusion of control, and availability bias.

Selective perception occurs when a problem is viewed from a limited or narrow frame. This, the group conclud- ed, explained why the organization was unable to move toward an integrated enterprise and why, as a practical example, top managers acted to protect their former constituents, unaware that incompatible software would jeopard- ize the project.

Second, the research groups sug- gested that the insular nature of the separate organizations created a proj- ect culture that limited communication among these units. It created, they maintained, a breeding ground for groupthink, where each group was iso- lated from others in the organization. As a result, many of the practices and procedures within these organizational units went unchallenged.

Third, while management at the central facilities in Toulouse initially envisioned a centralized organization with some control over its divisions, the inability of these divisions to use the same version of CATIA suggested that this transition was far from effective. Senior management, the research groups concluded, suffered from the illusion of control.

The fourth explanation was attrib- uted to availability bias, in which management at each of the plants was limited to the data that was available to

them, and thereby concluded that the project was meeting its local objectives. From their perspective, they were on schedule and within budget.

Coast Guard Maritime Domain Awareness Project In 2001, the U.S. Department of Homeland Security and the U.S. Coast Guard undertook a project that prom- ised to create the maritime equivalent of an air traffic control system. It was a project that would combine the use of long-range surveillance cameras, radar systems, and information technology to automatically identify vessels as they entered U.S. ports. It was also a project that would span 24 federal agencies responsible for the protection of water- ways and coasts.

In the first phase of the project, a complete surveillance system, called Project Hawkeye, was to be developed and implemented for the port of Miami. It would track larger vessels with radar, and smaller vessels, even jet skis, with infrared cameras. Finally, a software system would process the data to deter- mine which vessels posed a security threat and deserved closer scrutiny.

The first test of the system was declared a failure (Lipton, 2006). The cameras were ineffective in tracking the small boats, the radar system proved unreliable when it incorrectly identi- fied waves as boats, the Automated Identification System used for large boats failed to meet its objectives, and the software systems needed to make sense of the data had yet to be installed. Although some data from the system was available to the Coast Guard, they were unable to use it. Because the test failed, the implementation of this sys- tem in 35 ports was delayed until at least 2014.

When the research groups present- ed their conclusions, they focused on the complexity of working with 24 agencies. They concluded that since so many components of the system had failed, project leaders succumbed to the illusion of control bias; these lead-

ers assumed that they had control over the agencies and subcontractors, when in fact they did not. Control issues, the research groups continued, should have been resolved before the project was undertaken, not after.

The groups also suggested that selective perception contributed to the failure. Each separate agency focused only on its immediate task, with appar- ently little effort directed at integrating their role with that of others; there was no suggestion of a strategic relationship among vendors and agencies.

Columbia Shuttle On February 1, 2003, seven astronauts perished when their Columbia Shuttle disintegrated as it re-entered the earth’s atmosphere. During launch, a piece of foam insulation, similar in composition to a Styrofoam cup and about the size of a briefcase, broke away from the main propellant tank. The foam struck the left wing, seriously breaching the protective panels on its leading edge (Gehman, 2003).

It was not the first time that a sec- tion of foam had broken away during launch. In fact, it had happened on every previous flight. But on each of these flights, the spacecraft reentered the earth’s atmosphere without incident and safely returned home. Management assumed that it was a problem of minor significance and that it did not increase the risk level of the flight (Starbuck & Farjoun, 2005).

Many concluded, certainly just after the 2003 tragedy occurred, that technology was to blame. But a more thorough and comprehensive investi- gation, undertaken by the Columbia Accident Investigation Board (CAIB), concluded differently. It maintained that management was as much to blame for the failure as was the foam strike. The Board described a culture where, at every juncture, program man- agers were resistant to new informa- tion. It was a culture where people were unwilling to speak up, or if they did speak up, they were convinced they

P A

P E

R S

would not be heard. They also conclud- ed that the organizational failure was a product of NASA’s history, culture, and politics (Columbia Accident Investiga- tion Board, 2003).

The study participants concluded that NASA had created a culture in which systematic biases went unchecked. First, the participants identified the recency effect. Foam insulation had broken away on previous flights and caused no harm. To them this was confirmation that these recent events distorted the real danger presented by this problem.

Conservatism was also suggested, because the data from these previous flights was largely ignored by senior managers; they failed to revise their prior belief that the system was operat- ing properly. There was also evidence of overconfidence. During the flight, engi- neers, concerned that the foam strike may have caused a problem, asked a manager of the Mission Management Team (MMT) to request satellite imagery of the spacecraft. Management, howev- er, was apparently confident that there was no safety issue, and a decision was made against imagery. Had the imagery been authorized, and the damage dis- covered, the conjecture is that a rescue attempt would have had a reasonable chance of success.

Selective perception was suggested, since management of the shuttle pro- gram had shifted from an engineering focus to a managerial focus. This mor- phed the organization in such a way that engineering problems were less likely to be recognized and more likely to be dominated by schedules and budgets (Gehman, 2003).

Denver Baggage Handling The new airport in Denver, with a budget of $4.9 billion and originally scheduled for completion in October 1993, would be one of the nation’s largest public works projects of the 1990s (Brooke, 1995). It would cover 53 square miles and include five runways, with future expansion to 12 runways. Due to its size and the necessity to

move baggage quickly between flights, the airport would feature a completely automated baggage-handling system.

In April 1995, after many delays, the baggage system project was completed. Reporters were invited to attend a demonstration but instead witnessed a public disaster. Delivery carts were derailed, luggage was torn, and piles of clothes and other personal items were strewn beneath the tracks (Myerson, 1994). After scaling back the scope of the baggage system and making the necessary design changes, the airport finally opened, 16 months behind schedule and almost $2 billion over budget (Keil & Montealegre, 2000).

The baggage-handling project at Denver was more complex than any- thing that had been attempted before at any airport. Luggage was to be first loaded onto conveyor belts, much as it is in conventional baggage-handling sys- tems. These moving conveyors would then deposit the luggage into moving computer-controlled carts at precisely the right moment. The luggage would then travel at 17 miles per hour to its des- tination, as much as one mile away. This underground rail system would be com- pletely automated and would include 4,000 baggage carts traveling throughout the airport and under the control of 100 computers. It would be capable of pro- cessing up to 1,400 bags per minute.

After the system failed its public test, and after design changes were implemented, the system still had problems. Only United Airlines used it, and then only for outgoing flights. Other carriers turned to a hastily constructed manual system, since no contingent plans had been made should the automated system fail. Finally, in 2005, after a decade of frus- trating attempts to solve its problems, the system was abandoned. Under the lease agreement, United Airlines, one of the major stakeholders in the proj- ect, would still be liable for $60 million per year for 25 years.

In two papers that raised concerns about the scope and feasibility of the

project, deNeufville (1994, 2000) contended that this baggage-handling system represented an enormous technological leap over current practices. He concluded that the prob- lem of accommodating the variable demand made on the system, charac- terized in the literature as a classic line- balancing problem, would be difficult to solve.

After discussing the case, the research group identified overconfi- dence as a major factor in the failure. They referenced a quote in the case summary taken from the New York Times: “While the airport was being designed, United insisted that the air- port have the fancier baggage handling system, which it contended would sharply reduce delays” (Johnson, 1994). Overconfidence was also suggested in another quote from the New York Times (Myerson, 1994), in which Gene Di Fonso, president of BAE, the prime contractor for the project, declared, “Who would turn down a $193 million contract? You’d expect to have a little trouble for that kind of money.” With widespread support, the group con- cluded that no one questioned whether it could be done.

They also identified the sunk cost trap. In spite of years of disappoint- ments, when all the airlines, with the exception of United, opted out and used a manual backup baggage- handling system, the project continued. Both the City of Denver and United Airlines had already incurred high costs and were unwilling to disregard these past expenditures, even as their prob- lems persisted and grew worse.

The illusion of control was also identified for its role in keeping the project alive too long. It helped explain why, after evidence was presented at the beginning, a line-balancing prob- lem of this magnitude was very difficult to solve, and why, after an embarrass- ing preview of the system to reporters, management was still confident that it could fix the problems and control the outcome.

December 2008 � Project Management Journal � DOI: 10.1002/pmj 9

10 December 2008 � Project Management Journal � DOI: 10.1002/pmj

Systematic Biases and Culture in Project Failures

Mars Climate Orbiter and Mars Polar Lander As part of the NASA Mars Surveyor Program, the Mars Climate Orbiter was to orbit Mars and collect environmen- tal and weather data. But as the spacecraft approached its destination, telemetry signals fell silent, and a $125 million mission failed.

The root cause identified by NASA was the failure to convert between met- ric and English units. When the fatal error was detected, Noel Hinners, vice- president for flight systems at Lockheed, the company that built the spacecraft, said in disbelief, “It can’t be something that simple that could cause this to hap- pen” (Pollack, 1999). But it was.

Apparently, Lockheed had used pounds during the design of the engines, while NASA scientists, respon- sible for the operation and flight, thought the data was in metric units.

There were early signs during its flight that something was wrong with the craft’s trajectory, and an internal review later confirmed that it may have been off course for months (Oberg, 1999; Pollack, 1999). Project culture, however, required that engineers prove that something was wrong rather than “prove that everything was right.” This difference in perspective prevented the team from looking into the problem. Edward Weiler, NASA associate adminis- trator for space science, said, “The prob- lem here was not the error; it was the fail- ure of NASA’s systems engineering, and the checks and balances in our processes to detect the error” (Oberg, 1999, p. 34).

The Mars Investigation Panel report identified several contributing factors to the failure: the system engineering process did not adequately address the transition from development (Lockheed) to operations (NASA); inadequate com- munications between project elements; and inadequate staffing and training.

Within a few months of the Orbiter failure, the Mars Polar Lander, a related NASA project with a price tag of $165 million, suffered the same fate. Its flight was uneventful until it began its landing

approach. Then, during its descent to the rough terrain of the polar cap, telemetry signals fell silent. With no data to pinpoint the precise cause of failure, the teams investigating the accident speculated that the vehicle’s descent engines prematurely shut down. Unable to slow the descent, the speculation was that the engines quit when the Lander was 130 feet high, crashing into the sur- face of Mars at about 50 miles per hour. The inappropriate response of its engines was attributed to software glitches (Leary, 2000).

The prevailing culture at NASA of “Better, Faster, and Cheaper,” which defined the period when these projects were in development, has been high- lighted many times as the contributing factor behind these failures. Thomas Young, a former NASA official, said that they were trying “to do too much with too little.” He continued, “No one had a sense of how much trouble they were actually in” (Broad, 1999).

The prevailing culture was best expressed in an internal memo written by a laboratory official at the Jet Propulsion Lab: “There might have been some overconfidence, inadequate robustness in our processes, designs or operations, inadequate modeling and simulation of operations, and failure to heed early warnings” (Oberg, 1999, p. 35).

While the trajectory problem asso- ciated with the Orbiter and the engine ignition problem associated with the Lander could be characterized as tech- nical, the Mars Climate Orbiter Failure Board Report (2000) said that manage- ment failures were also to blame. They found that these projects suffered from a lack of senior management involve- ment and too much reliance on inexpe- rienced project managers. The Board also criticized the strategy where proj- ect managers in one organization (Lockheed) were responsible for devel- opment and a separate organization (NASA) was responsible for operations after launch.

The study group first identified the sunk cost trap. If the orbiter did not

launch on schedule, it would have to wait several months before its next opportunity to launch. With launch windows far apart, and with budgets unable to tolerate a substantial delay, managers were under pressure to meet the deadline; it was important not to “waste” the effort put into the project to that point.

Selective perception bias was iden- tified and used to explain why the engi- neers at the Jet Propulsion Lab, the design team, failed to coordinate with the operational team at NASA. In large- scale complex projects such as the Orbiter and Lander, with countless activities, contractors, and suppliers, it is very possible that teams may take a narrow view of their own activities. The risk is that the work of one team may be incompatible with the work of another.

Conservatism, the group contin- ued, explained why engineers failed to take action when they noticed that the trajectory of the spacecraft was off. They even held a meeting in Denver to address the issue, but it was never resolved. Even as the spacecraft approached its destination and data showed that it was drifting off course, controllers largely ignored the real data and assumed it was on course (Oberg, 1999).

Merck Vioxx In 2000, the New England Journal of Medicine published an article suggest- ing that Merck misrepresented clinical trial data on the risks of Vioxx, a drug used to treat arthritis pain. Suspicions were raised again when the Journal of the American Medical Association, pub- lished a paper in 2001 finding that those who took Vioxx were more than five times more likely to experience a cardiac event than those taking a commonly used over-the-counter anti- inflammatory drug, Naproxen. Merck denied these claims, insisting the find- ings were “flawed” (Topol, 2004). Then, under increasing pressure, they revised the Vioxx label in 2002 to reflect these added risks.

P A

P E

R S

During this period, Merck had undertaken a separate study focusing on the use of the drug in treating colon polyps. New data from this trial simply confirmed the risks that had been raised earlier.

Shortly thereafter, on September 30, 2004, five years after it had been introduced to the market, after 84 mil- lion people had taken the drug, and after three years of denying the drug could induce heart attacks and strokes, it was pulled from the shelves (Topol, 2004).

The legal consequences were signif- icant. Over 27,000 claims were filed contending that certain incriminating data were withheld during the FDA review process. In an early trial, a New Jersey jury ruled unanimously in March 2007 that Merck committed consumer fraud by intentionally suppressing, concealing, or omitting information on the risks of Vioxx. Eventually, Merck proposed an out-of-court settlement to the remaining complainants at a cost of over $5 billion (Berenson, 2007).

The study group identified organi- zation and project culture as important contributors to the Vioxx project fail- ure. They cited a Business Week article contending that Richard Clark, CEO, had watched the company degenerate into a “collection of fiefdoms” more focused on their own agendas than on the company’s agenda (Weintraub, 2007).

Financial pressures, the group con- tended, also shaped the culture. Drug discovery is a costly and lengthy process, fraught with risk. The average cost to bring a drug to market exceeds $1 billion. As drug trials proceeded from animal to human trials and even- tually to FDA review, it was not unrea- sonable to conclude that the pressure to continue with the project increased as investment increased.

These cultural problems, together with financial pressures, continued, creating a breeding ground for system- atic biases to emerge. The sunk cost trap was identified as the dominant

bias. After incurring nearly $1 billion to develop the drug, after generating $2.5 billion in sales during 2003, it was not difficult to understand why the compa- ny resisted pressure to remove Vioxx from the market.

While the sunk cost trap dominat- ed, conservatism was also identified as contributing to the failure, because Merck suppressed early data suggesting that the drug could have serious and sometimes tragic side effects.

Microsoft Xbox 360 When Microsoft rushed its Xbox video game console to market in November 2005, it had a one-year advantage over Sony and Nintendo. By 2007, Microsoft had sold over 11.6 million units at $279 to $479, depending on configuration.

Unresolved issues plagued the proj- ect from the beginning. When journal- ists and reviewers were invited to try the Xbox 360 in 2005, before it became available on store shelves, they encountered problems connecting it to the Internet (Croal, 2007). Shortly after it was introduced to the public, users complained that that the console dam- aged game disks, so much that they could no longer be used (Cliff, 2007). In 2005, Microsoft recalled the power cords, concerned that they posed a fire hazard (Wolverton & Takahashi, 2007). Then, in December 2006, in an appar- ent response to these and other issues, Microsoft extended the warranty from 90 days to one year.

But problems persisted. Blogs and forums complained about the “Red Ring of Death,” referring to a string of three lights that illuminate on the con- sole when a serious malfunction occurs. One survey found that the return rate was 33% (Cliff, 2007).

Then, in July 2007, Robbie Bach, president of Microsoft’s Entertainment and Devices Division, said that there had been an “unacceptable high num- ber of repairs” (Taub, 2007). Shortly thereafter, Microsoft announced an extension of the warranty from one to three years at an expected cost of

$1 billion. This represented about $100 for every Xbox sold since its introduc- tion in 2005.

Later in the same month Microsoft announced that its top gaming execu- tive, Peter Moore, was leaving the com- pany, but denied that his departure was related to the Xbox’s engineering prob- lems (Wingfield, 2007).

Three systematic biases were iden- tified by the group. The first was con- servatism. In the face of a continuous stream of product returns and cus- tomer complaints, those who were responsible for the project were unwill- ing to acknowledge that the problem was serious, that customer satisfac- tion and loyalty were deteriorating rapidly, that the product needed to be redesigned, and that customer satisfac- tion needed to be addressed.

It was also suggested that manage- ment fell prey to the sunk cost trap. Considerable investment in the prod- uct had already been made, sales were strong, and since the division had yet to turn a profit, there was pressure to con- tinue at any cost. Returning to earlier stages of design, issuing a recall for the defective units, and replacing them with new units were apparently not realistic options.

Because Microsoft declined to com- ment on the exact cause of the prob- lem, which many suspected was tied to either a power cord or component that was overheating, it was concluded that groupthink was also an issue. The only public comment, made by Robbie Bach, president of the Entertainment and Devices Division responsible for the Xbox, was that the company made manufacturing and production changes that should reduce hardware lockups (Taub, 2007). It was suggested by the group that this could be interpreted as protecting the company to prevent exposing its failures.

New York City Subway Communications System In New York City, police officers who worked underground in the city’s

December 2008 � Project Management Journal � DOI: 10.1002/pmj 11

12 December 2008 � Project Management Journal � DOI: 10.1002/pmj

Systematic Biases and Culture in Project Failures

extensive rapid transit subway system were routinely unable to communicate with officers working the streets above- ground. Incompatible systems were at fault. Not that this was a new problem to law enforcement and emergency organizations in New York City. On September 11, 2001, for example, it was not possible for police to communicate with firefighters and warn them that the World Trade Center towers were in jeopardy of collapsing.

As early as the 1990s, preliminary plans for an integrated communication system had been proposed. In 1999, a contract was signed with two firms. The project was scheduled for completion in 2004, with an approved budget of $115 million.

In 2001, a report warned of an inter- ference problem that could jeopardize the ability of the systems to work together. Rather than return to the design stage and study the validity of this concern, subcontractors continued with the project. It was completed in October 2007, but during the trial of the system it became apparent that inter- ference did indeed create serious communication problems. As a result, implementation was halted. Fixing the problem was expected to increase the cost of the project to $210 million.

The group linked the failure of the New York City Subway Communi- cations Project to conservatism, over- confidence, and illusion of control. Conservatism was suggested when the project managers failed to take the interference warning seriously enough to change their plans early in the proj- ect. Overconfidence was also suggested to explain why they ignored the warn- ing: project managers were apparently convinced that the proposed design would work or that all problems could eventually be solved.

Homework is Completed By:

Writer Writer Name Amount Client Comments & Rating
Instant Homework Helper

ONLINE

Instant Homework Helper

$36

She helped me in last minute in a very reasonable price. She is a lifesaver, I got A+ grade in my homework, I will surely hire her again for my next assignments, Thumbs Up!

Order & Get This Solution Within 3 Hours in $25/Page

Custom Original Solution And Get A+ Grades

  • 100% Plagiarism Free
  • Proper APA/MLA/Harvard Referencing
  • Delivery in 3 Hours After Placing Order
  • Free Turnitin Report
  • Unlimited Revisions
  • Privacy Guaranteed

Order & Get This Solution Within 6 Hours in $20/Page

Custom Original Solution And Get A+ Grades

  • 100% Plagiarism Free
  • Proper APA/MLA/Harvard Referencing
  • Delivery in 6 Hours After Placing Order
  • Free Turnitin Report
  • Unlimited Revisions
  • Privacy Guaranteed

Order & Get This Solution Within 12 Hours in $15/Page

Custom Original Solution And Get A+ Grades

  • 100% Plagiarism Free
  • Proper APA/MLA/Harvard Referencing
  • Delivery in 12 Hours After Placing Order
  • Free Turnitin Report
  • Unlimited Revisions
  • Privacy Guaranteed

6 writers have sent their proposals to do this homework:

Finance Master
Ideas & Innovations
Unique Academic Solutions
Premium Solutions
Quick N Quality
Maths Master
Writer Writer Name Offer Chat
Finance Master

ONLINE

Finance Master

I have assisted scholars, business persons, startups, entrepreneurs, marketers, managers etc in their, pitches, presentations, market research, business plans etc.

$50 Chat With Writer
Ideas & Innovations

ONLINE

Ideas & Innovations

As an experienced writer, I have extensive experience in business writing, report writing, business profile writing, writing business reports and business plans for my clients.

$19 Chat With Writer
Unique Academic Solutions

ONLINE

Unique Academic Solutions

I have assisted scholars, business persons, startups, entrepreneurs, marketers, managers etc in their, pitches, presentations, market research, business plans etc.

$31 Chat With Writer
Premium Solutions

ONLINE

Premium Solutions

I will provide you with the well organized and well research papers from different primary and secondary sources will write the content that will support your points.

$50 Chat With Writer
Quick N Quality

ONLINE

Quick N Quality

As per my knowledge I can assist you in writing a perfect Planning, Marketing Research, Business Pitches, Business Proposals, Business Feasibility Reports and Content within your given deadline and budget.

$49 Chat With Writer
Maths Master

ONLINE

Maths Master

I will be delighted to work on your project. As an experienced writer, I can provide you top quality, well researched, concise and error-free work within your provided deadline at very reasonable prices.

$27 Chat With Writer

Let our expert academic writers to help you in achieving a+ grades in your homework, assignment, quiz or exam.

Similar Homework Questions

Tandara motor inn case - Lead and lag indicators - Waste land documentary essay - Characteristics of accounting information pdf - Current density of copper wire - Threats to trader joe's competitive advantage - Criminal justice - Ucf vpn cisco - Columbia 400 vs cirrus sr22 - +971561686603 Abortion pills in Dubai/Abu Dhabi-mifepristone & misoprostol in DUBAI - Bus 308 week 1 assignment - Business Continuity and Disaster Recovery planning - Housekeeping essay - Rufus fry watsons go to birmingham - Discussion - Using excel solver to maximize profits - What errand did ruth run - Malware domain list txt - Https www citiprogram org members index cfm pageid 125 view - Assume the following information about the market and jumpmasters stock - A christmas carol essay - Microscopy microbiology lab report - Butchery course tafe nsw - English - Assignment - Which of the following is a tool that can be used to check for modified or trojanzied files? - Stephensons estate agents dalton in furness - Infrared thermometer best used for servsafe - The talk funny girl - Denton company manufactures and sells - ECON - Research paper - Gentra puregene blood kit - Assignment 1 - Alicia has been asked to approve a marketing campaign that - 05.05 analyzing 20th century poetry - How to select cells a3 through f3 in excel - Homework: three questions (250 Words each) (urgent in 6 hours) - Strategic positioning attempts to achieve sustainable competitive advantage by - From ancient grudge to new mutiny meaning - Melrose primary school teachers - Example of anecdotal record of a child - Technology and the good life - Lawrie voutier demolition shop - Owner builder permit qld - Marketing james l burrow answers - Experimental research - Fixed and flexible expenses - Critical Reasoning - Tata nano the people's car case study - Chapter 9 natural laws and car control answer key - Excel chapter 9 grader project school of information sciences - Save smog city 2 from ozone worksheet answers - The dulac box plant produces 500 - Zhang ailing sealed off text - Electrolysis of sodium sulphate using carbon electrodes - Paint net lesson plans - What landforms are created at divergent boundaries - Quality Improvement Project in Nursing - Discussion post - Cliff program rockville correctional facility - What does macduff do in macbeth - How customers think gerald zaltman - Khyber kennels & cattery - Cormier and hackney counseling model - Closest airport to harvard - Project - Animal farm discussion questions chapter - Led to the use of scrims and whirling stage platforms - Pop and mentos rocket - Nursing - Music - Senge's five disciplines and their applications - Lección 4 grammar quiz completar - Discussion Question - Homework 9.57 - Poem friend by hone tuwhare - King ethelbert school teachers - Perception errors in organisational behaviour - Research proposal on cloud computing pdf - The guest by albert camus questions and answers - Www dolphinhousesurgery co uk - I'll pour this pestilence in his ear - A10 - Interprofess 1 - Nursing - Departure and up the coolly essay - Sir george cayley facts - The morality of migration by seyla benhabib summary - Windshield wiper mechanism design - Human resources managment legal ethical. deliverable 03 - Building a catapult with popsicle sticks - Leadership enhancing the lessons of experience - Accounting clerk education requirements - Activity based costing acca - PCI Compliance - Waves and vibrations lab answers - 7 11 case study supply chain - Utopia definition and characteristics - Discussion 3