http://aviationweek.com/bca/lessons-bedford-gulfstream-accident-part-2
N121JM wreckage, aerial photograph, from NTSB Accident Docket ERA14MA271

How many times have you mentally kicked yourself for doing something incredibly stupid? And upon reflection, you can’t even figure out *why* you performed this way?  This “all too human” process is fascinating and built into our human operating system. Examining and understanding the little daily goofs and lapses can improve your life, but more importantly, improve your aviation safety so the big, bad stuff can (hopefully) be avoided.

To unpack the psychological process at play here, we first need to understand that most of our activities in life are only semi-consciously decided and most often automatically enabled. Our human operating system copes with the daily overload of sensory data, decisions and actions by utilizing a series of scripts or schema largely out of sight and called “implicit.” This is how we can famously drive to our destination in a car and not remember anything about the trip (or that we intended to stop at the store). We can also type 180 words per minute but are usually unable to label the qwerty keys without tapping them out on a table. This “implicit knowledge” and associated “scripts and schema” are internal and and invisible. They also are not even filtered or examined by conscious oversight; operating in the shadows. (This “dual-process” brain theory was popularized in Daniel Kahneman’s book “Thinking Fast and Slow“). We continuously process and act in an automatic mode unless we have time and motivation to engage in effortful reflection. This natural optimizing has great survival value from efficiency, but in complex activities it fails to manage risk and can occasionally wreck airplanes.

Let’s examine a typically bad decision to drive home, despite being pretty buzzed, after a party. Imagine we arrive home just fine despite our incapacitation. Achieving this “success”, we experience a feeling of relief and accomplishment. Now consider the opposite example of having a totally sober designated driver taking us home but getting T-boned and injured. We have in the first case a bad decision resulted in a good outcome, but in the second case a good decision resulted in a bad outcome. And though decisions and outcomes actually stand alone, that is not how our human brain interprets these situations. The first decision is reinforced due to “success” and the second one discounted as “bad luck”. Letting this “implicit learning” into our subconscious as a standard of operation can have serious consequences for our future safety. Let’s see how our brains subconsciously “code” these experiences.

In the case where our decision was bad but the outcome positive, our “success” psychologically reinforces our decision with relief and a feeling of accomplishment; e.g. “that turned out OK.” Though defective, this decision and validating success delivers a warm buzz of dopamine that neurologically encodes in our brains as “acceptable”. Unless we consciously reflect later on this action, to critique and correct this mental coding, a habit can easily develop and become an implicitly learned bad procedure. Clearly luck was the primary operative factor in both cases, not skill. As humans, however, our “fast processor” tends to evaluate all decisions solely on the basis of the outcome rather than applying an objective standard or evaluating the quality of the decision. And once imported, an implicitly accepted standard of “only one drink” can easily slide to “only two drinks” in the same manner; and you see where this is ends up going. A more thoughtful approach, guided by objective reflection, (carefully decided) is the antidote to this implicit learning and the basis for “Standard Operating Procedures” in charter and airline flying. Implicit procedures are self-optimizing and haphazard and slide to the lowest level of acceptability based on luck. And we all, unfortunately, know pilots that fly in this manner.

An aviation example of this process at work might start with a successful outcome despite marginal weather; perhaps arriving into a Delta airspace safely under a 1200 ft ceiling (even though our personal minimums were “1500 for VFR in Delta”).  Unless we later reflect and critique this dubious “success” we now have an implicitly accepted “new normal”. Our optimizing human brain codes “that worked out OK” and this new standard becomes part of our pilot operating system, out of sight and never adjudicated by our “better pilot self”. Pretty soon we find ourselves with new and sketchy standards and we might not even know why or how this standard was established. These implicitly learned, automatic schema are imported “under the radar” and are accepted as operational–just like bad code embedded in computer program. And similarly, they may operate fine for a while until they fail suddenly and surprisingly. This neurological process is responsible for “normalizing deviance” that led to NASA’s safety issues with the Space Shuttle. On a “group think” level these same implicit procedures were working so we “go with it” (the only standard was “success”). Many of these errors are not big and obvious, but insidiously erode standards on every flight. And this is also how thousands of hours and increased experience can work against us by building complacency rather than excellence.

To successfully combat the implicit learning of questionable procedures, an “after flight critique and reflection” is essential for safety. Otherwise, our optimizing human brain is always at work creating implicit shortcuts–efficiency over safety. It is vital for future safety to always schedule a sacred time for personal analysis after every flight ; do it when you log the flight? This reflection process is essential to analyze what went right and wrong–and most importantly “why?” This is also the reason why written personal minimums (or professional SOPs) are necessary to keep a pilot honest and safe. Sliding standards, often implicitly learned, seem to always precede the crunch of aluminum.


Please “follow” our SAFE blog to receive notification of new articles. Write us a comment if you see a problem or want to contribute an article. We are always seeking more input on aviation improvements and flight safety. There are many highly qualified aviation educators out there! If you are not yet a SAFE member, please Join SAFE and support our mission of generating aviation excellence in teaching and flying. Our amazing member benefits alone make this commitment worthwhile and fun. Lastly, use our FREE SAFE Toolkit App to put pilot endorsements and experience requirements right on your smart phone and facilitate CFI+DPE teamwork. Working together we make safer pilots!

About the author 

David St. George (Lifetime Member)

David St. George learned to fly at Flanders Valley Airport in 1970. Proving that everyone is eventually trainable, he became an FAA Gold Seal Flight Instructor for airplanes (single and multi, instrument, and glider) and serves the Rochester FSDO as an FAA Designated Pilot Examiner. In this capacity, he gives flight tests at any level from sport pilot to ATP and CFI. For 25 years David was East Hill Flying Club's 141 Chief Instructor and manager. David holds multi and single engine ATP pilot certificates, with pilot ratings for glider and seaplane and several jet type ratings. He recently earned his 13th renewal as a Master Instructor and owns an Aeronca Champ so he can build hours for that airline job! http://learnturbine.com

You may also like

SAFE at #OSH22!

SAFE at #OSH22!

Active Learning; Rewire Your Brain!

Active Learning; Rewire Your Brain!

Leave a Reply

{"email":"Email address invalid","url":"Website address invalid","required":"Required field missing"}