HOME

TheInfoList



OR:

In the field of
human factors and ergonomics Humans (''Homo sapiens'') or modern humans are the most common and widespread species of primate, and the last surviving species of the genus ''Homo''. They are great apes characterized by their hairlessness, bipedalism, and high intellig ...
, human reliability (also known as human performance or HU) is the probability that a human performs a task to a sufficient standard. Reliability of
human Humans (''Homo sapiens'') or modern humans are the most common and widespread species of primate, and the last surviving species of the genus ''Homo''. They are Hominidae, great apes characterized by their Prehistory of nakedness and clothing ...
s can be affected by many factors such as age, physical
health Health has a variety of definitions, which have been used for different purposes over time. In general, it refers to physical and emotional well-being, especially that associated with normal functioning of the human body, absent of disease, p ...
, mental state, attitude,
emotion Emotions are physical and mental states brought on by neurophysiology, neurophysiological changes, variously associated with thoughts, feelings, behavior, behavioral responses, and a degree of pleasure or suffering, displeasure. There is ...
s, personal propensity for certain mistakes, and cognitive biases. Human reliability is important to the resilience of socio-technical systems, and has implications for fields like
manufacturing Manufacturing is the creation or production of goods with the help of equipment, labor, machines, tools, and chemical or biological processing or formulation. It is the essence of the secondary sector of the economy. The term may refer ...
,
medicine Medicine is the science and Praxis (process), practice of caring for patients, managing the Medical diagnosis, diagnosis, prognosis, Preventive medicine, prevention, therapy, treatment, Palliative care, palliation of their injury or disease, ...
and nuclear power. Attempts made to decrease human error and increase reliability in human interaction with technology include user-centered design and error-tolerant design.


Factors Affecting Human Performance

Human error, human performance, and human reliability are especially important to consider when work is performed in a complex and high-risk environment. Strategies for dealing with performance-shaping factors such as psychological stress, cognitive load,
fatigue Fatigue is a state of tiredness (which is not sleepiness), exhaustion or loss of energy. It is a signs and symptoms, symptom of any of various diseases; it is not a disease in itself. Fatigue (in the medical sense) is sometimes associated wit ...
include heuristics and biases such as confirmation bias, availability heuristic, and frequency bias.


Human reliability analysis

A variety of methods exist for human reliability analysis (HRA). Two general classes of methods are those based on probabilistic risk assessment (PRA) and those based on a cognitive theory of control.


PRA-based techniques

One method for analyzing human reliability is a straightforward extension of probabilistic risk assessment (PRA): in the same way that equipment can fail in a power plant, so can a human operator commit errors. In both cases, an analysis ( functional decomposition for equipment and task analysis for humans) would articulate a level of detail for which failure or error probabilities can be assigned. This basic idea is behind the Technique for Human Error Rate Prediction (THERP). THERP is intended to generate human error probabilities that would be incorporated into a PRA. The Accident Sequence Evaluation Program (ASEP) human reliability procedure is a simplified form of THERP; an associated computational tool is Simplified Human Error Analysis Code (SHEAN). More recently, the US Nuclear Regulatory Commission has published the Standardized Plant Analysis Risk – Human Reliability Analysis (SPAR-H) method to take account of the potential for human error.


Cognitive control based techniques

Erik Hollnagel has developed this line of thought in his work on the Contextual Control Model (COCOM) and the Cognitive Reliability and Error Analysis Method (CREAM). COCOM models human performance as a set of control modesstrategic (based on long-term planning), tactical (based on procedures), opportunistic (based on present context), and scrambled (random) – and proposes a model of how transitions between these control modes occur. This model of control mode transition consists of a number of factors, including the human operator's estimate of the outcome of the action (success or failure), the time remaining to accomplish the action (adequate or inadequate), and the number of simultaneous goals of the human operator at that time. CREAM is a human reliability analysis method that is based on COCOM.


Related techniques

Related techniques in safety engineering and
reliability engineering Reliability engineering is a sub-discipline of systems engineering that emphasizes the ability of equipment to function without failure. Reliability is defined as the probability that a product, system, or service will perform its intended functi ...
include
failure mode and effects analysis Failure is the social concept of not meeting a desirable or intended Goal, objective, and is usually viewed as the opposite of success. The criteria for failure depends on context, and may be relative to a particular observer or belief system ...
, hazop, fault tree, and SAPHIRE (Systems Analysis Programs for Hands-on Integrated Reliability Evaluations).


Human Factors Analysis and Classification System (HFACS)

The Human Factors Analysis and Classification System (HFACS) was developed initially as a framework to understand the role of human error in aviation accidents.Wiegmann and Shappell, 2003 It is based on James Reason's Swiss cheese model of human error in complex systems. HFACS distinguishes between the "active failures" of unsafe acts, and "latent failures" of preconditions for unsafe acts, unsafe supervision, and organizational influences. These categories were developed empirically on the basis of many aviation accident reports. "Unsafe acts" are performed by the human operator "on the front line" (e.g., the pilot, the air traffic controller, or the driver). Unsafe acts can be either errors (in perception, decision making or skill-based performance) or violations. Violations, or the deliberate disregard for rules and procedures, can be routine or exceptional. Routine violations occur habitually and are usually tolerated by the organization or authority. Exceptional violations are unusual and often extreme. For example, driving 60 mph in a 55-mph speed limit zone is a routine violation, while driving 130 mph in the same zone is exceptional. There are two types of preconditions for unsafe acts: those that relate to the human operator's internal state and those that relate to the human operator's practices or ways of working. Adverse internal states include those related to physiology (e.g., illness) and mental state (e.g., mentally fatigued, distracted). A third aspect of 'internal state' is really a mismatch between the operator's ability and the task demands. Four types of unsafe supervision are: inadequate supervision; planned inappropriate operations; failure to correct a known problem; and supervisory violations. Organizational influences include those related to resources management (e.g., inadequate human or financial resources), organizational climate (structures, policies, and
culture Culture ( ) is a concept that encompasses the social behavior, institutions, and Social norm, norms found in human societies, as well as the knowledge, beliefs, arts, laws, Social norm, customs, capabilities, Attitude (psychology), attitudes ...
), and organizational processes (such as procedures, schedules, oversight).


See also

* * (A Technique for Human Event Analysis) * , a technique used in the field of human reliability * * * * (Tecnica Empirica Stima Errori Operatori) * Incident pit, conceptual model from diving for explaining incident development and recovery * Success Likelihood Index Method


Footnotes


References

*

* * * * * * * * * * * * *

* * *

* *Federal Aviation Administration. 2009 electronic code of regulations. Retrieved September 25, 2009, from https://web.archive.org/web/20120206214308/http://www.airweb.faa.gov/Regulatory_and_Guidance_library/rgMakeModel.nsf/0/5a9adccea6c0c4e286256d3900494a77/$FILE/H3WE.pdf


Further reading

* * * * * *

* * *

* * * * * * * * * * * * * * * *CCPS, Guidelines for Preventing Human Error. This book explains about qualitative and quantitative methodology for predicting human error. Qualitative methodology called SPEAR: Systems for Predicting Human Error and Recovery, and quantitative methodology also includes THERP, etc.


External links


Standards and guidance documents


IEEE Standard 1082 (1997): IEEE Guide for Incorporating Human Action Reliability Analysis for Nuclear Power Generating StationsDOE Standard DOE-HDBK-1028-2009 : Human Performance Improvement Handbook


Tools


EPRI HRA CalculatorRiskSpectrum HRA software

Simplified Human Error Analysis Code


Research labs


Erik Hollnagel
at th
Crisis and Risk Research Centre
a
MINES ParisTech
at the US Sandia National Laboratories
Center for Human Reliability Studies
at the US
Oak Ridge National Laboratory Oak Ridge National Laboratory (ORNL) is a federally funded research and development centers, federally funded research and development center in Oak Ridge, Tennessee, United States. Founded in 1943, the laboratory is sponsored by the United Sta ...

Flight Cognition Laboratory
at NASA Ames Research Center
David Woods
at th
Cognitive Systems Engineering Laboratory
at The
Ohio State University The Ohio State University (Ohio State or OSU) is a public university, public Land-grant university, land-grant research university in Columbus, Ohio, United States. A member of the University System of Ohio, it was founded in 1870. It is one ...

Sidney Dekker's Leonardo da Vinci Laboratory for Complexity and Systems Thinking, Lund University, Sweden


Media coverage


“How to Avoid Human Error in IT“

“Human Reliability. We break down just like machines“
Industrial Engineer – November 2004, 36(11): 66


Networking


High Reliability Management group at LinkedIn.com
{{DEFAULTSORT:Human Reliability