Priming Observability for Situations

The OODA Loop

In a previous post, we called out several issues with the standard or typical example of the OODA loop. One particular point was the lack of detail surrounding the Decide phase of the model. While OODA explores some of the factors involved in the Orient phase that feeds into the Decide phase, it offers minimal elaboration on what is relayed and how it might be reasoned about. Here is where we can augment the OODA loop with another decision-making model, referred to as the Recognition-Primed Decision (RPD) model of rapid decision making.

Recognition-Primed Decision Model

The model asserts that individuals assess the situation and generate a plausible course of action (CoA), which is then evaluated using mental simulation. The authors claim that decision-making is primed by recognizing the situation and not entirely determined by recognition in the original paper. The model runs counter to the common thinking that individuals employ an analytical model in complex time-critical operational contexts. Multiple options are carefully evaluated, weighed up, and compared before choosing a response or action. The analytical model works best with inexperienced individuals, whereas experts employ more of a naturalistic decision-making method that is heuristic, holistic, or intuitive.

Situation Recognition

In the RPD model, an expert understanding of a situation depends mainly on the goals, cues, expectations, and the typical actions within such situations (prototypical patterns). The RPD model has three components: the matching component, diagnosis, and simulation. The matching component attempts to identify the current situation from the memory of prototypical situations. Further diagnostics are obtained if the situation is not recognized and (online) learning is engaged. Here pattern recognition is central to decision making. A pattern consists of cues, spatial and temporal relationships, cause-and-effect chains and reflects the operational goals and expectations.

Charting a Course

The intelligence or expertise we commonly hear in mission-critical situations is the extensive knowledge of patterns that make it extremely easy to identify the small but critical states that a system is within or is about to enter. Once a decision, or choice of action script, is made, a mental simulation of the anticipated consequence is done, and the expected outcome is compared with the goals. Here an effective and efficient mental model of the situation is paramount for a fast and accurate assessment. If the outcome is favorable, the action is taken; otherwise, alternative scripts are evaluated. If none of the mapped scripts are found to be acceptable, then further diagnosis is initiated. 

Reframing Observability

Most of us can readily recognize the essential aspects of this model in everyday life. Still, it is hard to pinpoint where our current Observability tooling approach supports complex distributed computing systems. The site reliability engineering (SRE) community’s current emphasis is on data collection, which is far too quickly and irresponsibly relabeled as information or worse knowledge. Acquiring information on a system equates to projecting the system’s future states with much less of a degree of uncertainty. How does a distributed trace, log, or event even come close to addressing such prediction capabilities and capacities? The lens (or model) from which we view the computing world has many engineers staring down at data and details, seeing trees, more so roots, while being utterly oblivious to the forest, ecosystem, and the nature at play (the dynamics of action). We’re failing at first base.