Biases Are Not Bugs

The word 'bias' carries negative connotations in everyday usage, implying a flaw or failing. In cognitive science, it means something more precise and, in context, more sympathetic: a systematic deviation from rational decision-making that arises from the use of mental shortcuts, known as heuristics, that are efficient under most circumstances. These shortcuts exist because the brain lacks the computational capacity to evaluate every decision from first principles. They evolved in environments where they generally produced good enough outcomes. The problem, for security practitioners, is that the environments in which we now operate have been deliberately designed by attackers to exploit them.

Understanding which biases are most relevant to cybersecurity, and how attackers operationalise them, is not merely an intellectual exercise. It is the necessary precondition for designing interventions that actually work, because an intervention that does not address the specific cognitive mechanism driving a behaviour is unlikely to change it.

Ten Biases, Ten Attack Vectors

Cognitive Bias

What It Is

How Attackers Exploit It

Security Example

Authority Bias

Tendency to comply with instructions from perceived authority figures

Spoofing executives, regulators, or IT departments

CEO fraud, fake IT helpdesk password resets

Urgency/Scarcity

Heightened response to time pressure or limited availability

Creating artificial deadlines and consequences for inaction

"Your account will be suspended in 24 hours"

Familiarity/Mere Exposure

Increased trust in things that feel recognisable

Using spoofed brand assets, known sender names, and familiar layouts

Pixel-perfect clone of a bank login page

Social Proof

Using others' behaviour as a guide when uncertain

Fabricated peer endorsement or implied consensus

"Your colleagues have already verified their accounts"

Reciprocity

Felt an obligation to return a favour or gesture of help

Offering assistance before making a request

Fake IT support resolving a minor issue, then asking for credentials

Optimism Bias

Belief that bad outcomes are less likely to happen to oneself

No direct exploit needed; this bias reduces vigilance generally

"I would notice a phishing email" (most people believe this)

Commitment and Consistency

Drive to behave consistently with prior commitments

Establishing small agreements before escalating the request

Multi-stage social engineering across several low-stakes interactions

Inattentional Blindness

Failure to notice unexpected stimuli when focused on a task

Embedding malicious content in legitimate-looking documents

Malicious macros in a routine supplier invoice

Hyperbolic Discounting

Preference for immediate rewards over larger future ones

Framing the attack as offering immediate benefit

Fake prize notifications, instant refund frauds

Affect Heuristic

Decisions driven by current emotional state

Inducing fear, excitement, or sympathy before the ask

Fake charity appeals following disasters; threat-of-arrest scams

Sources: Cialdini (2007); Kahneman (2011); Vishwanath et al. (2018).

The Bias of Believing You Are Not Biased

Before examining how these biases affect end users, it is worth dwelling on a finding that tends to be unwelcome in security circles: these same biases operate on security professionals with equal force, and in some respects with greater consequence.

The bias blind spot, first documented by Pronin and colleagues at Princeton, is the systematic tendency to recognise cognitive biases in others while underestimating their influence on one's own judgement (Pronin et al., 2002). Security professionals who have studied social engineering techniques tend to score particularly high on confidence measures, but not significantly higher on actual detection performance. The knowledge creates an illusion of immunity.

 

Professionals who study social engineering tend to score highly on confidence, but not significantly higher on actual detection. Knowledge of a bias does not neutralise it.

 

Bias

How It Affects Security Professionals

Practical Consequence

Confirmation Bias

Interpreting ambiguous evidence in incident response to fit the initial hypothesis

Tunnel vision during investigations; relevant indicators missed

Anchoring

Over-reliance on the first data point (e.g. CVSS score) when assessing risk

Low-CVSS vulnerabilities with high exploitability are de-prioritised

Availability Heuristic

Overestimating the likelihood of threats that are recent or memorable

Resource allocation skewed toward high-profile attack types over statistically more common ones

Planning Fallacy

Systematic underestimation of time and resources required for security projects

Patch cycles slip; remediation timelines are consistently over-optimistic

Automation Bias

Excessive trust in automated tool outputs; reduced critical evaluation

SIEM alerts are accepted or dismissed without adequate human judgement

Sunk Cost Fallacy

Continued investment in failing security tools or programmes because of prior spend

Legacy awareness platforms retained despite evidence of ineffectiveness

This table does not imply that security professionals are uniquely susceptible. It reflects that no professional group is immune to cognitive biases, and that high-stakes decisions made by security teams amplify the consequences when biases operate unchecked.

Urgency as the Master Exploit

If a single cognitive vulnerability had to be selected as the one most consistently exploited across attack types, urgency would be the clear candidate. Time pressure reliably degrades decision quality by suppressing the slow, deliberate processing that would otherwise detect inconsistencies. It is not simply that people make faster decisions under time pressure; it is that they make qualitatively different decisions, relying more heavily on System 1 pattern-matching and bypassing the verification steps that System 2 would normally impose.

Vishwanath and colleagues (2018) found that susceptibility to phishing increased significantly when emails included explicit time constraints, even when all other elements were held constant. This is why urgency cues are a near-universal feature of social engineering attacks across every channel: email, voice, SMS, and increasingly synthetic video. The attacker who creates a sense of urgency has effectively disabled the target's primary defence mechanism.

PRACTITIONER IMPLICATION: Any high-value transaction process that can be completed quickly, at the request of a single individual, under apparent time pressure, is architecturally vulnerable regardless of how well-trained the people involved are. The design question is how to build verification steps that feel like part of the normal process rather than an extraordinary interruption.

Social Proof at Scale

Social proof deserves particular attention as a bias that technology is amplifying rather than diminishing. The classic mechanism, the tendency to look to others' behaviour as a guide when uncertain, has always been exploitable through fabricated endorsements and implied consensus. What has changed is the ease with which fabricated social signals can be generated and the scale at which they can be deployed.

In an organisational context, social proof operates through the apparent behaviour of peers and authority figures. If employees observe that senior leaders do not use MFA consistently, or that their colleagues regularly share passwords for shared accounts, these observations are more powerful predictors of their own behaviour than any number of training modules advocating for the opposite. The implication is direct: visible, consistent, secure behaviour by respected figures in the organisation is one of the highest-leverage interventions available, and it costs nothing beyond intentional attention.

From Understanding to Action

The purpose of mapping biases to attack vectors is not to produce a more sophisticated threat taxonomy. It is to identify the specific mechanisms that any effective intervention must address. An awareness programme that does not engage with the fact that urgency suppresses System 2 processing is not addressing the actual cause of the behaviour it is trying to change.

Part 3 of this series translates this understanding into practical intervention design, using the EAST framework and choice architecture principles to show how security teams can change the environments in which these biases operate, making secure outcomes more likely without relying on employees to overcome their cognitive architecture through willpower alone.

 

References

Cialdini, R.B. (2007). Influence: The Psychology of Persuasion. Revised ed. Collins Business.

Kahneman, D. (2011). Thinking, Fast and Slow. Penguin Books.

Pronin, E., Lin, D.Y. and Ross, L. (2002). The bias blind spot: Perceptions of bias in self versus others. Personality and Social Psychology Bulletin, 28(3), pp.369-381.

Vishwanath, A., Harrison, B. and Ng, Y.J. (2018). Suspicion, cognition, and automaticity model of phishing susceptibility. Communication Research, 45(8), pp.1146-1166.