Username
 
  Password
 
 



Avoiding Human Error - Creating the Right Environment

 

Summary
The high percentage of Cost of Quality (COQ) related to sales (15-40%) and the high percentage of errors attributed to humans (35-80%), provide opportunities for significant improvement to an organization’s bottom line. While everyone in an organization wants to perform well and be error free, the reality is as Cicero stated in 200 BC, “to err is human.” Individual awareness, leadership roles, and organizational culture must be considered in creating the right environment for avoiding human error. From an individual perspective, avoiding human error includes the review of specific task demands, unique capabilities of individuals, and our general human nature. Examples of leaders helping their team members in avoiding human error include ensuring open communication, promoting teamwork, and reinforcing the desired jobsite behaviors. There are a number of activities that an organization can do to foster a culture that creates the right environment, such as valuing the avoidance of errors and strengthening the integrity of defenses to mitigate consequences of an error. Because of their influence, members of the Quality Organization must be familiar with the implementation of error avoidance concepts and techniques. These concepts and techniques are valuable tools for creating the right environment to achieve a reduction in human error and cost of quality. These actions can result in significant improvements by reductions in cost of quality and improving the bottom line and customer image.

Introduction to Avoiding Human Error
While many factors can lead to errors, the human factor is by far the most significant contributor, with 35–70% of all errors attributed directly to human beings. In fact, the transportation industry reports that its human error rate is 80%. The Cost of Quality (COQ) ranges from 15–45% and a large portion of COQ can be directly related to human errors within an organization. As stated by the noted psychologist James Reason, “human factors cannot be ignored.”

The human factor is not limited to individuals. It is actually a complex situation that involves individuals, leadership, process, organizational structure, and culture. Recent publications and key reports also recognize the contribution of these elements to errors. The Shuttle Columbia accident report noted the organization’s impact on human error by stating “people’s actions are influenced by the organizations in which they work, shaping their choices in directions that they many not even recognize.” In its landmark publication “To err is human,” the Institute of Medicine (2000) reported that human error results in up to 98,000 deaths per year The report stated that 80% of error in medicine administration was directly attributed to human error A key report investigating the death of 14 firefighters in the 1994 Storm King Mountain fire outside Glenwood Spring, CO found that the firefighting community places too much emphasis on technology and not enough emphasis on human factors (Putnam 1995). In addition, The National Law Enforcement Officers Memorial Foundation identifies attitude as the number one in the list of ten fatal errors that have killed experienced officers.

Although the industries referenced earlier highlight more severe examples of human error, all organizations have a great need to create the right environment at all levels to detect and avoid human errors. Also, investigating and understanding where breakdowns have or may occur is essential to mitigate and preclude recurrence. When looking at errors, the tendency is to focus more on technical issues than human factors because technical issues are measurable, more easily understood, and more readily changed. When human factors are considered, solutions for nonconformance generally include statements like the root cause was human error; corrective action is to retrain and/or discipline the individual; and anomaly report completed. Unfortunately, this does not address the systemics or total environment related to the error, such as the process, organizational structure, or culture that may have set up or contributed to the situation and allowed the error to occur. Without looking for systemic issues, the true root cause may not have been determined and the adverse environment with the potential for the error may still exist to trap the next unsuspecting individual.

Cicero statement in 200 BC “to err is human,” still applies today. It is essential for an organization to have a proactive approach to avoiding and managing human error and creating the environment necessary to minimize the potential for error at all levels. The aviation industry’s highly successful Crew Resource Management (CRM) program addresses the human factors and interactions in the cockpit, which has significantly reduced the number and severity of accidents/incidents since its implementation (Helmreich, Merritt, Wilhelm, (1999). Similar proactive program are needed in all industries. Recently, an aerospace executive that requested and received specialized error avoidance training after a major incident stated “why does something have to happen for us to do something we should have done before it happened.”

Background on Ongoing Activities
The aviation industry initiated the CRM training program during the 1970s in response to the high accident rate, and it is the first proactive program designed to create the environment necessary to avoid human error. The Federal Aviation Administration (FAA) performed analysis of flight voice recorder data, which revealed that an unacceptable working environment between crew members in the cockpit caused most of the accidents. Very few accidents were due solely to technical conditions. In addition, the analysis showed that two primary factors leading to the accidents existed in the working environment: lack of leadership by the command pilot and lack of assertiveness by the crew members. This analysis was summarized in the Public Broadcasting Service Nova series program entitled “Why Planes Crash.” Combined, these tendencies lead to catastrophic results. Crew members often knew that a problem existed, but they did not feel comfortable demanding the captain’s attention or the captain did not pay adequate attention to the concerns of the crew. A prime example is the crash of United Flight 173 en route to Portland, OR on December 28, 1978. The co-pilot and first officer knew the plane would run out of fuel and asked the captain to check the fuel. When the captain did not respond, the crew did not assertively alert the captain of the fuel status and the plane crashed (National Transportation Safety Board Report, 1979).

A National Aeronautics and Space Administration workshop in 1979 on aircraft accidents is usually referred to the initiating event in the effort to improve air safety. Using research performed by the Department of Psychology at the University of Texas, the FAA established specialized cockpit member training in six error management areas: Situational Awareness, Communication, Attitude, Risk, Workload, and Group Dynamics. The training evolved over time with various names and is now referred to as Crew Resource Management

In 2001, the University of Texas, Department of Psychology suggested that the basic principles of CRM could be applied in other domains, specifically identifying the medical and maritime industries as prime candidates, (Helmreich, Wilhelm, Klinect, and Merritt, 2001). The Institute of Medicine recognized this need in 2000 and recommended “incorporating proven methods of managing work teams as exemplified in aviation industry.” The Fire Service also recognized the value of CRM and is adapting these same techniques. Gary Briese, Executive Director of International Association of Fire Chiefs, put the importance of their efforts in perspective when he said “In the 10 years it will take CRM to be introduced nationally, we will attend 1,000 firefighter funerals…I can’t get that out of my mind.”

Other industries have initiated adaptations of CRM, including Nuclear Power, Aerospace, Coast Guard, and Military Aviation industries. In fact, an Air Force Regulation establishes a CRM program for the Air Force. In its publication, the Air Force stated that although this type of training was “historically geared toward the operational flight environment, the potential exists to adapt the fundamental program principles to any tasks or functional areas requiring cooperation or time critical efforts.”

Pioneering efforts by Nancy Leveson at MIT analyzing major aerospace accidents showed that leaders and organizations can also create adverse environments leading to human error (Leveson, 2001). She referred to this adverse environment as “systemic conditions because the system which includes the individual as well as the organization broke down and allowed the error to occur” (Leveson, 2001). She recommended that accident investigations include the examination of the technical, human, organizational, management, and societal systemic conditions. Dr. Leveson’s approach included the usual accident/incident description of events and conditions and expands to include the review of systemic factors as the indicators for true root cause. This basic approach was used in the Shuttle Columbia accident investigation to assess the conditions, events, and systemic factors that caused the accident to occur. The approach pioneered by Dr. Leveson is being considered by other industries to determine the true root cause of an accident or incident.

Other tools to improve organization efficiency and effectiveness are Lean Thinking and Cycle Time Reduction. The application of these tools, which is often used in conjunction with Six Sigma to reduce steps in a process, includes reducing waste and close examination of the methods used to reduce human error (e.g., reducing workmanship errors is a major element of waste).

The aviation industry forged the path by creating a comprehensive human error program. Comparatively, very little has been done in other industries to establish a total system approach to avoiding human errors. Some industries are just starting robust activities, and they are beginning to understand the difficulty in creating the necessary environment in avoiding errors and obtaining acceptance on the importance of human factors to avoid errors. The medical and fire service communities have cultural similarities to that of the aviation industry in the 1970s and are finding difficulties with implementation. The aviation industry again provides valuable insight for those who research and adapt the lessons learned to their specific culture.

Although there are limited examples of a comprehensive, total system approach to human error outside of aviation, there are some exceptional examples on the consideration of human factors as a key component in avoiding errors. One example is the pioneering work in the Doctor Quality Incident Reporting System developed and implemented for employees at the Baylor Medical Center at Grapevine Texas (Atherton, 2002). The most significant aspect of this incident reporting system is the development of a new organizational culture regarding errors and error reporting that includes both employees and hospital leadership. Another example is found in the Fire Service’s landmark study on the Storm King fire where it recognized that continuing a primary focus on the technical aspects of fire fighting without considering human factors would not reduce the number of firefighter deaths in similar situations (Putnam, 1995).

When establishing robust error avoidance activities, industries and organizations must recognize and accept human fallibility and the potential for error and expand beyond industry adaptations to include the unique cultures within their organizations. One type of training does not fit the needs of every organization. The book Crew Resource Management for the Fire Service states that training must be tailored for every individual fire station (Okray and Lubnau, 2004-2011). Based on the author’s seven years of experience analyzing organizations and conducting training, an error avoidance culture can only develop by considering the existing leadership and workforce culture, adapting training appropriately, planning for strategic and tactical implementation, and developing a follow-up strategy to incorporate changes.


Figure 1 Error Management Areas - Traps and Tools

New Activities to Create the Right Environment
The starting point for implementing new error avoidance activities or methods is accepting that although the basic principles inherited from the successes of CRM activities must be considered, training must be specifically adapted and tailored for each unique organizational culture. Standardized training courses cannot fit every industry or every organization within an industry. Because errors are powerful and insidious, training and other activities must be developed and adapted for each organization to help individuals and groups raise their level of awareness of the potential for an error and the impact on activities. James Reason summarized it very well when he stated that we need to “create an environment of chronic unease…”

Pearl Buck, the Pulitzer Prize author, gave us hope in avoiding errors when she stated “Every great mistake has a half-way moment, a split second when it can be recalled and perhaps remedied.” As part of our efforts to create the right environment and avoid errors, we can learn to recognize when we are about to make a mistake and can be trained to take advantage of the split second before we make an error.

Training of individuals and teams to avoid errors in each error management area (e.g., Situational Awareness) focuses on understanding the Mind Traps that can prevent us from seeing that an error may occur. Through this understanding we are better able to recognize when we are about to make a mistake by sensing that something isn’t right. Training in error avoidance Tools assists us in avoiding the potential for error. Figure 1 gives examples of Traps and Tools associated with the six error management areas described earlier. One of the most powerful tools is intuition or the gut feeling that senses something is not right and by verbalizing it to ourselves or others it becomes real. In his book Blink, Gladwell, 2005 vividly describes this powerful tool, referring to it as “adaptive unconscious.” Evidence suggests that this powerful tool can be traced back to our primitive human nature when survival required its use and the integration of all our senses to determine and assess danger. Ideally, an organization and its leadership create an environment where it is expected for individuals to feel free to speak up without the fear of retribution.

Error avoidance training can use an adaptation of the Swiss Cheese Model, Figure 2, developed by James Reason to remind the individual, the last line of defense, to look at the adequacy of technical, organization, and culture defenses (Reason, 1997). To fill holes in these defenses, every person in the organization should be looking upstream to determine what could go wrong in any of the defenses they use or create. Therefore, individuals and team members need to be trained to look for what could go wrong.

Figure 2 Adapted Swiss Cheese Model

AIn an error avoidance culture, effective leaders are learning how they can adversely influence the behavior of members of their organization (Banda, Associated Press). Under certain circumstances, encouraging a “can do” attitude can be interpreted by team members as permission to take unnecessary risks. Leaders also understand that they must create and accept assertive followers (Abeyta, 1997) in order to receive the information needed to fully understand risks in making informed decisions. They need to instill a culture of encouraging bad news to travel at least six times faster than good news if they want their employees to really keep them informed on current situations and the risks associated with various actions. Additionally, they find it better to determine what happened rather than following the usual pattern of blame, discipline, and/or retrain.

Studies also show that leaders are working to improve their interpersonal skills, since leadership failures stem more from a lack of these skills than a lack of technical or business knowledge. From an error avoidance perspective, leaders are trained that sharing about errors made during their own careers demonstrates their humanity and proves beneficial to both the team members and the organization. It illustrates the leader’s expectation that errors are to be avoided but also recognizes human fallibility and the importance that everyone in the organization understands where the system broke down to cause an error, thereby avoiding future recurrence.

Both individuals and teams are using error avoidance techniques to carefully examine established processes and determine what could go wrong. Error avoidance Tools are used to examine the specifics of the task, equipment to be used, barriers to completing the task, and the ability of individuals and the team to perform the activity. Too often individuals and teams think they cannot make a mistake because of an approved or established process. They are forgetting that processes cannot evaluate themselves. It is people who care about processes and who can make errors by following a faulty process. As James Reason suggested, those organizations are considered to be high reliability organizations, such as Navy aircraft carriers or the nuclear power industry, which have created an environment of “chronic unease” about what could go wrong.

Due to the inherent diversity in individual personalities and the possibility of an adverse environment in their own organization, many individuals are reluctant to bring attention to errors and near misses or analyze errors within their organization. To overcome this situation, some organizations find it useful to start by analyzing incidents from other organizations as case studies and discussing how the incidents could occur in their own environment and how they can prevent it from happening.

One of the most promising practices for avoiding future errors is the report, analysis, and distribution of information about near misses. The aviation industry pioneered the creation of an anonymous reporting system called the Aviation Safety Reporting System (ASRS) used by both flight crew members and maintenance personnel. Errors are submitted anonymously to an outside organization, the information is analyzed, names and identifying flight information is removed, and the analysis of the event and suggestions for preventing the incident in the future are distributed throughout the aviation industry. The individual submitting the information is provided an acknowledgement of its receipt and is protected from retribution. It is more important to understand all aspects of the problem than to find someone to blame. The system does, however, include methods to deal with individuals abusing the system. As it should be, unsatisfactory behavior and abuse of using the system is not tolerated. Public access to samples and resulting analysis from the ASRS can be found at asrs.arc.nasa.gov/main_nf.htm.

In 2004-2011 an aerospace company created a propriety system to receive and thoroughly analyze near misses and distribute the results. The organization recognizes that near misses are a source of valuable information to avoid future errors. Information about the near misses is distributed throughout the enterprise, and individuals submit feedback on how the information helped them avoid the same or similar errors. The net result of these activities was a 26% increase in reporting of near misses and a 78% reduction of incidents. In 1999 Westrum identified the importance of “looking for and listening to faint signals” that may indicate the potential for a problem or error. Another approach looking for what could go wrong by a high reliability organization is described in Managing the Unexpected (Weick and Sutcliffe, 2001).

To encourage error reporting, some organizations have created a program where individuals identifying areas where they have or may have made an error are recognized with an award. By identifying these possible errors, an organization can review whether an error actually occurred and determine whether or not it can be corrected. Additionally, it may highlight the necessity of requiring additional training for the individual or teams in the specific area where the error occurred. This approach is widely accepted by the organizational members and provides another example where creating the right environment to determine cause is more important than individual blame. It also makes individuals more willing to accept responsibility for actions in their work activities.

Arguments for the Proposed Activities
The biggest argument for the implementation of error avoidance activities is the acceptance by numerous industries and organizations that human factors must be considered. This was specifically emphasized in the report on the Storm King Mountain fire. In addition, and based on the diversity of industries trained by The Center for Error Management, many industries are adapting broad-based error management techniques to avoid human error. For example, law enforcement agencies are being trained using basic CRM principles with special emphasis on recognizing the Mind Trap of a “risky attitude” and how to counter this Trap by the use error avoidance Tools.

Organizations are beginning to realize that they have a tendency to blame an individual rather than taking time to thoroughly analyze the situation to determine a systemic breakdown in the organization that allowed the incident to happen. To demonstrate a resolution to a non-conformance, it is easy to identify the individual who committed the error and define corrective action as discipline and/or retrain the individual. After seeing the repetition of the same or similar incident, customers ask more questions about true root cause and the corrective action plans to prevent future occurrence. As a result, more organizations take time to find the true root cause and established techniques to more easily identify it in future incidents. A good example of this is the NASA root cause analysis approach (Bradley, O’Connor, 2003) that follows many of the concepts advocated by Nancy Leveson.

Error avoidance must be considered a series of techniques and methodologies that can be incorporated throughout an organization rather than a new stand-alone program. In addition to flowing down these techniques to various programs, one aerospace company embracing error management recognizes the importance of creating an atmosphere for individuals to speak up and instituted a proactive two-way communication program encouraging leaders to understand the error and all its causes versus placing blame.
.
Error and near miss reporting is the next logical evolution of CRM-based programs. As discussed earlier, Baylor University Hospital has reported extraordinary results after establishing an anonymous reporting system for errors and near misses (Atherton, 2002). The reporting system achieved up to a 500% increase in responses and because of the data being readily available, the time required to track and make improvements was reduced up to 50%. The concentrated efforts and commitments of the leadership team created a culture that accepts and encourages reporting of errors and near misses. Many features of this reporting system were modeled after the highly successful FAA Aviation Safety Reporting System (ASRS).

An aerospace organization with specialized and detailed tasks by machinists, tracked a specific area they called “mental lapse”, or loss of situational awareness. After training team leadership and individuals by The Center for Error Management, this mental lapse category was improved by 30%, from a “red” level of concern to a level of “green” (CEM training 2004-2011). The primary focus of the training was to help facilitate open discussions with the machinists, leaders, and internal customers to bring up potential problems on received work orders and in process concerns. In addition, training was provided to facilitate smooth transitions during shift changes and to encouraging joint discussions on assignment of cause and corrective action when an anomaly occurred.

Justification of Why Proposed Activities Can Create the Right Environment
Implementing error avoidance methods and techniques as part of the organizational culture yields outstanding results. The suggested activities are proactive and are designed to detect the potential for and the avoidance of errors. Organizations must recognize that some errors will occur despite all efforts and when an error does occur, they must have a solid reactive effort to identify the error; mitigate it, if possible; and perform error analysis to determine systemics to prevent recurrence. Additionally, the activities must be specifically tailored and adapted for the industry and individual organizations. The bottom line is that a single approach cannot fit every organization.

Creating an environment where individuals and team members are owners of organizational success has proven to be very successful. A foreign object elimination (FOE) program demonstrated that when the operators/technicians/laborers were the suspected cause of a foreign object induced failure, the team took undirected action and cleaned up a shop beyond expectations, demonstrating a significant pride of ownership. An assembly shop that had been an eye-sore became an organizational showpiece for customers and provided a great sense of pride for employees.

When leaders are trained in error management techniques, they become sensitive to the many ways that what they say and how they say it, including the subtleties of body language, can influence their teams to cut corners, take chances, bypass procedures and not feel free to speak up about concerns. If the leader understand these sensitivities, they can become a better leader by encouraging Assertive Followership and creating expectations that team members need to speak up to identify things that could go wrong.

Thorough systemic analysis of major anomalies has demonstrated the need for more human error awareness, proactive understanding, and sensitivity for maintaining situational awareness. In addition, it confirmed the need for a “no blame” organizational environment that rewards near miss and active error reporting. The FAA’s success with ASRS verifies that a reporting culture reduces the cost of quality and establishes a more productive work force through fewer missed days of work by employees. As an example of effectiveness of error reporting by the another organization, the FAA has data showing that after implementing near miss reporting and analysis, the organization achieved a 40% reduction in accidents and a 37% decrease in lost time (Chris Hart, undated). Another organization was recognized within an incentive fee for their incorporation of error reporting
Because some errors can result in injury or even death, efforts to avoid errors must be a prime objective for every organization. In addition, having a solid and proven error avoidance program in place sends a strong message about the organization’s commitment to current and future customers. Successful reduction in errors, accidents, and injuries combined with a decrease in workmanship may be used to assist in negotiations with insurance organizations when determining rates.

Considering the high percentages of human error and the significant influence these errors have on the cost of quality, creating the right environment to avoiding error has a remarkable positive impact on the organization. Creating the right environment for avoiding human errors reduces the cost of quality and injuries and improves the bottom line, and customers like it. The development of an organizational understanding of human fallibility results in involvement by leadership, improved team communication, and team building.

Anticipated Directions of Future Activities
Envisioning the ideal environment provides a roadmap of the activities needed to avoid human error. Many activities, although fragmented throughout many industries and organizations, are starting the process of creating the desired end result. For those industries with a fairly robust program, future activities will focus on extensive specialized training to introduce and sustain greater error avoidance awareness. This emphasis will assist all members of the organization to develop an environment of “chronic unease” and better able to effectively maintain situational awareness. The overall concept of error management and lessons learned from CRM will expand into other industries, and industries that have already started implementation will continue to increase their activities and scope. As CRM concepts become more widely understood, other industries will seek out training based on the CRM principles, and include methods for adapting the principles to their particular organization. Many organizations will learn and apply the experiences from other successful organizations and intensify their activities to focus on what could go wrong with ongoing/planned activities and processes. In addition, increased individual and team awareness will expand the well recognized medical Hippocratic Oath “Do No Harm” to include an assertive dedication of “Allow No Harm” by actions of others.

When looking at the desired end result, the individual feels free to speak up when things seem wrong and shares information with others on errors they made or techniques used to avoid an error. In this ideal environment, the leader encourages high productivity and error-free performance as the expectation, but the leader creates an environment where individuals feel free to talk about anything that doesn’t seem right without the fear of retribution. The leader creates a team environment where anyone feels free to be a courageous follower and stop an operation they feel may cause an accident or injury. The leader rewards individuals for avoiding human error, and the leader’s actions are fully supported by a senior organization leadership structure. The organization openly shares information about accidents, incidents, and near misses so individuals can use the information to avoid future errors. Insurance rates decrease as reports show that incorporating a near miss reporting and analysis system with proactive error avoidance activities significantly decreases the number of injuries, lost time, and payments of workman’s compensation. The organization recognizes and takes preventive actions so organizational culture, structure, and communications do not inadvertently encourage individuals to take unnecessary risks.

In addition, processes are written by individuals who understand the working level and who are implementing the process. This person works proactively with the individuals performing the job to thoroughly review the process and assess what can go wrong. A part of this examination includes how to use error avoidance tools to detect the potential for an error and avoid the error when implementing the process.

At the organizational and culture level, the management and leadership teams are trained in their roles in creating the right environment for error avoidance. They understand that the environment created by their actions, words, and deeds influence the behavior of the members of their organization. They are committed to understand what happened in an incident rather than looking for someone to blame. They support identifying what broke down in the system and what created the environment for the error and recognizes the value of the expertise in their own organization to avoid future errors. Leadership follows the proven practice of creating peer review teams. These teams operate independently and make periodic visits to functional areas similar to their own, evaluate the effectiveness of their error avoidance methods and techniques, and provide personal feedback to the individual team being evaluated. Management is not involved in these self assessments and is only informed on recommendations that need their attention for avoiding human error. To support internal activities on a routine basis, key individuals in the organization from various disciplines are trained as Error Avoidance Coaches. These coaches are called upon in special situations to conduct refresher training, e.g., a team that is about to perform a specialized task. They can also be called upon to be present during critical activities, and provide on the spot assistance for implementing error avoidance methods and techniques. This is an expansion of another option the organization can consider by establishing an in-house training team who receives specialized training in a Train the Trainer program for error avoidance.

In conjunction with creating and sustaining an error reporting and analysis system for near misses, the organization leadership encourages discussion about mistakes and sharing this information with other members. These activities result in the establishment of a Learning Organization where knowledge is openly shared for the benefit of everyone.

Implementation of the techniques and methods discussed above reduce human error and the overall cost of quality, while increasing the organization’s bottom line. The quality organization is able to focus more on prevention/avoidance activities and less on internal rework before shipment or correction of failures that occur after the product is delivered to the customer. Furthermore, the quality organization takes the lead to ensure that error avoidance techniques and methods are implemented throughout the organization.

Conclusion
When considering the high percentages of human error and the significant influence of error on the cost of quality, creating the right environment to avoid error offers a tremendous potential for a positive impact on the organization. Creating the right environment for avoiding human errors at the individual, leadership, and organizational levels, can reduce the cost of quality, avoid injuries, improve the bottom line, and is viewed by customers as proactive corrective action. To create the right environment for avoiding errors, there is a need for development of an organizational acceptance of the phenomena of human fallibility with an understanding of human mind traps that can lead to errors and tools to avoid those errors. These activities can result in a more effective and responsible workforce, a proactive leadership team, and an organization culture dedicated to avoiding errors. These activities send a strong message to current and future customers of an organizational commitment to proactive rather than reactive behavior.

In thinking of the end in mind, implementation of the techniques and methods discussed above will result in a reduction in human error, decrease of overall cost of quality and an increase in organizational bottom line. As shown in Figure 3, error reduction achieved from contributions from Error Avoidance Activities coupled with improvements in Appraisal Contributions, and Failure Contributions result in reductions in total Cost of Quality By incorporating these contributions the quality organization will be able to focus more on prevention/avoidance activities and few appraisal and corrective actions, e.g., internal rework before shipment and correction of failures that occur after the product is delivered to the customer.

Figure 3 Contributions to Reducing the Cost of Quality

In planning for changing to these new ways of doing things, everyone in the organization needs to be reminded of what Helen Keller said, “You cannot change the whole world, but you can change the world where you are”. Gandhi also put this in the proper perspective when he said, “You must be the change you want to see in the world.”


 

 

References

Committee on Quality of Health Care in American, Institute of Medicine (2000), to err is human: building a safer health system, Washington D.C. National Academy of Sciences

Crosby, P.B. (1979), Quality is Free, New York: McGraw Hill

Deming, Edward (1982), “Out of Crisis”, Cambridge Massachusetts, Massachusetts Institute of Technology

Feigenbaum, A.V. (1956), “Total quality control”, Harvard Business Review, Vol.34, No.6.p.93

Gladwell, Malcom, M, (2005), Blink, New York, Little, Brown and Company

Hart, C.A., Global Aviation Information Network (GAIN), Presentation, Federal Aviation Administration (undated)

Helmreich, R.L. Merritt, A.C., & Wilhelm, J.A. (1999). “The evolution of Crew Resource Management training in commercial aviation”, International Journal of Aviation Psychology. 9(1), 19-32

Machowski, F. and Dale, B. G. (1978), “Quality Costing: An examination of knowledge, attitudes and perceptions”, Quality Management Journal, Vol.5. No.3, p.84

Nova Public Broadcasting Series program , Why Planes Crash,

Okray R and Lubnau, T II, (2004-2011), Crew Resource Management for the Fire Service Tulsa Oklahoma, PennWell Corporation

Plunkett, J.J. and Dale, B.G. (1987), “A review of the literature on quality-related costs”, International Journal of Quality & Reliability Management, Vol. 4, No.1, p.40

Reason, J. (1990), Human error. New York: Cambridge University Press

Reason J. (1997), Managing the Risk of Organizational Accident, Aldershop, United Kingdom: Ashgate

Sandoval-Chavez, D.A. and Beruvides, M.G. (1998), “Using opportunity costs to determine the cost of quality: A case study in a continuous-process industry”, Engineering Economist, Vol.43, p.107

Schiffauerova, A. and Vince Thomson, V. Cost of Quality: “A Survey of Models and Best Practices”, International Journal of Quality and Reliability Management

Weick, K. and Sutcliffe, K , (2001) “Managing the Unexpected”, San Francisco, Jossey-Bass, John Wiley and Sons

- Larry Tew

© The Center for Error Management 2004-2011