Introduction
Advances in generative artificial intelligence (GenAI) are transforming how states anticipate and manage strategic crises.Footnote 1 No longer limited to analysing discrete data inputs – such as radar tracks indicating a missile launch, troop movement alerts, or changes in alert status – AI-enabled decision-support systems (AI-DSS), including large language model–based tools, can synthesise complex intelligence and generate scenario-based outputs.Footnote 2 GenAI-enabled systems extend these capabilities by producing complex, real-time simulations of adversary intentions, escalation pathways, and alternative courses of action, often presented in immersive and emotionally compelling formats. This capability represents more than a technical upgrade to existing nuclear command, control, and communications (NC3) networks; it signals a profound epistemic shift in how strategic planning is conceived, how risk is perceived, and how decisions are made under crisis conditions. This article examines how GenAI-DSS may transform nuclear crisis decision-making by altering strategic imagination, shaping perceptions of escalation, and influencing human–AI interaction in future high-stakes environments.
The article argues that while GenAI offers enhanced predictive power and anticipatory insight, it may also improve early threat detection, enhance situational awareness, and provide decision-makers with faster, more adaptive responses to events. However, it simultaneously creates systematic distortions that threaten crisis stability. By blurring the line between strategic forecast and foresight, and between anticipation and action, GenAI risks transforming strategic planning from a process of exploration into one of algorithmic prescription. In doing so, these systems do not merely augment human analysis and judgement; they actively shape how decision-makers perceive and interpret emerging threats. GenAI can shift strategy from a deliberative practice grounded in uncertainty, ambiguity, and human intuition to a mechanised process in which outputs are treated as definitive, narrowing the space for human deliberation, bargaining, and de-escalation.
The article identifies three key mechanisms through which GenAI-DSS can undermine crisis stability. First, normalisation of low-probability, high-impact escalation paths, whereby catastrophic scenarios – such as nuclear confrontation – become embedded in planning, doctrine, and strategic culture, rendering them increasingly ‘thinkable’. Second, misattribution of adversarial intent, as ambiguous or contradictory actions are reframed through the illusion of algorithmic certainty, leading to false interpretations of hostile intent. Third, the emergence of synthetic feedback loops, in which machine-generated scenarios begin to shape the behaviour of both sides, turning simulated crises into self-fulfilling escalation. Together, these mechanisms erode judgement, compress decision timelines, and destabilise deterrence in ways traditional nuclear theory has not anticipated.
Nuclear crisis decision-making and escalation scholarship emphasises misperception, signalling failure, crisis bargaining, and escalation under conditions of uncertainty.Footnote 3 Yet this work assumes instability arises from ambiguous and incomplete information. Today, decision-makers confront a different challenge: an abundance of AI-generated strategic futures rather than information scarcity. As GenAI simulates adversary intentions and escalation pathways, the problem shifts from interpreting uncertain signals to navigating synthetic scenarios that may reshape risk perception and strategic choice.
This article contributes to the literature on intelligence failure, nuclear deterrence, and crisis decision-making by theorising ‘synthetic foresight’ as a distinct epistemic force.Footnote 4 It also builds on emerging scholarship examining AI in military decision-making and nuclear stability, extending these debates by identifying the distinct epistemic risks introduced by generative AI-enabled simulation.Footnote 5 Prior research has examined how cognitive bias, organisational dysfunction, and technological surprise shape crisis decision-making.Footnote 6 This study extends that work by showing how GenAI, by disrupting core features of crisis decision-making, can reshape the psychological and institutional foundations of nuclear strategy. It also engages current debates over AI governance, emphasising that risks arise from not only technical flaws or ‘bad data’ but also the ways GenAI-DSS, embedded within a broader socio-technical context, mediates human perception and imagination in high-stakes settings.Footnote 7
Recent scholarship suggests that AI-enabled tools may assist crisis managers in cultivating empathy, perspective-taking, and de-escalatory policy options during nuclear confrontations.Footnote 8 Holmes and Wheeler argue that AI’s ‘detached and unemotional analysis’ can enhance interpretive understanding and widen the space for restraint.Footnote 9 By contrast, this article examines GenAI systems capable of simulating adversary intent and escalation pathways. It contends that, under conditions of acute uncertainty, such outputs may narrow rather than expand perceived options by rendering extreme escalation trajectories cognitively salient and strategically plausible, thereby producing distinct epistemic dynamics.
The article is organised into three sections. The first section distinguishes traditional forecasting from foresight and explains how and to what effect GenAI merges these distinct epistemological functions. The second develops the concept of ‘synthetic foresight’ and identifies three mechanisms by which it may destabilise decision-making during nuclear crises. The third section employs fictional scenarios to illustrate the potential implications of synthetic foresight for future crisis decision-making, offering a novel, reflexive critique of how GenAI both constructs and constrains strategic imagination.
GenAI as a synthetic cognitive actor: Blurring the boundary between prediction and action
The integration of GenAI into nuclear crisis decision-making represents more than a technical enhancement in predictive modelling; it signals a fundamental epistemic shift in how strategic futures are conceived, evaluated, and acted upon. Traditionally, forecasting has relied on structured data, probabilistic analysis, and historical modelling to estimate the likelihood and implications of near-term events.Footnote 10 In the nuclear context, this involves drawing on bargaining theory and historical patterns to produce probabilistic assessments of adversary actions and potential escalation pathways within bounded time horizons.Footnote 11
Strategic foresight, by contrast, is an exploratory discipline – an imaginative effort to prepare for surprise under conditions of deep uncertainty.Footnote 12 In the nuclear domain, where data are sparse and outcomes non-linear, this distinction is crucial: forecasting informs probabilistic anticipation, while foresight cultivates adaptive preparedness. GenAI increasingly blurs, and in some cases collapses entirely, this distinction. By generating simulations that are both statistically coherent and emotionally resonant – triggering strong affective responses in human users – GenAI systems do not merely augment exploration; they also enhance the emotional impact of the experience. Moreover, they actively shape perception and guide decisions, collapsing the boundary between prediction and prescription and fundamentally transforming the dynamics of nuclear crisis decision-making.
This section examines the epistemic consequences of that transformation, tracing how GenAI’s fusion of forecast and foresight reconfigures not only the content of strategic anticipation but also the cognitive and institutional conditions under which nuclear decisions are made. It considers how these systems may alter the function of NC3, by not simply accelerating early warning or enhancing situational awareness but also reframing how uncertainty, judgement, and strategic surprise – events that involve great consequence to the surprised party – are interpreted and enacted at the brink.Footnote 13
From forecasting to anticipatory systems
Although computational modelling and simulations have long been used to identify correlative patterns and make causal inferences about potential geopolitical flashpoints, the advent of big data and AI techniques – especially machine learning – has dramatically expanded their scope, speed, and fidelity.Footnote 14 AI-powered geospatial systems that provide essential spatial awareness for command-and-control operations through maps, terrain analysis, and tracking of friendly and enemy force positions now monitor global developments, analyse the underlying drivers of conflict, and detect anomalies that may escape human analysts.Footnote 15 More recently, large language models (LLMs) have been used to process social media, imagery, and historical datasets to generate real-time intelligence assessments and inform possible courses of action.Footnote 16
In the United States, for example, Palantir promotes its Artificial Intelligence Platform (AIP) as a system that integrates diverse data streams and employs machine learning and large language models to enhance commanders’ situational awareness and target recognition capabilities.Footnote 17 These outputs can be visualised in command centres through unified display interfaces, including interactive maps, dashboards, and layered visualisations integrating geographic, temporal, and threat data.Footnote 18 Open-source reporting similarly indicates that Rhombus Power has used AI-driven analysis to forecast major military developments, including Russia’s 2022 invasion of Ukraine, and has explored providing Taiwan with comparable early warning capabilities related to potential Chinese military activity.Footnote 19
In the context of NC3 systems, AI-enabled predictive analysis offers new capabilities for generating a range of threat scenarios, enhancing early threat detection, strengthening strategic warning systems, and providing decision-makers with real-time situational awareness and ‘adaptive targeting’ support to generate threat-level assessments and dynamic target-priority lists.Footnote 20 As the first line of defence, strategic warning provides decision-makers with critical insights into the global threat environment, guiding strategy and resource allocation.Footnote 21 It relies on a network of sensors – such as over-the-horizon and early warning satellites, radar systems, and intelligence platforms – to monitor operational activity and detect anomalies, including missile launches, ballistic trajectories, or unusual shifts in adversary military posture. These data streams feed continuously into decision support systems, ensuring that early warning and situational awareness are maintained in real time. The Chinese military, for example, has developed an AI-DSS called ‘StarSee’, which integrates intelligence and situational awareness to assist commanders in making informed decisions.Footnote 22
In recent testimony before the Senate Armed Services Committee, General Anthony J. Cotton, the former commander of US Strategic Command (STRATCOM), underscored the importance of advanced data integration and decision-support capabilities in ongoing NC3 modernisation efforts.Footnote 23 AI-enhanced systems can process this sensor data using advanced time-series analysis and pattern recognition techniques, enabling faster and more accurate differentiation between routine activity and potential nuclear threats.Footnote 24 Recent reports indicate that analytic tools were used during a recent India–Pakistan crisis to clarify activity at Pakistani military bases, reducing the risk that routine movements could be misinterpreted as nuclear escalation preparations.Footnote 25
Such developments underscore the potential of AI to detect early indicators of strategic instability, flag emerging risks, filter false alarms, and extend decision-making timelines – thereby buying time for informed decisions, calibrated deterrence, effective signalling, or de-escalation under uncertainty. NATO, for example, emphasises ‘strategic anticipation’ and preparedness as core components of its deterrence posture, while Alliance initiatives on AI technologies aim to integrate AI-enabled capabilities to support crisis management and collective deterrence.Footnote 26 GenAI, by contrast, represents a fundamental departure from these traditional systems. Rather than merely aggregating and displaying discrete data inputs, it actively generates synthetic scenarios and simulations that can shape perceptions and decisions, fundamentally transforming both the dynamics of crisis decision-making and the role of humans within it.
Despite its promise, scepticism persists regarding AI’s ability to accurately predict conflicts or pinpoint the precise triggers of crises – primarily due to two fundamental challenges: limited, fragmented data and the inherent difficulty of anticipating leaders’ intentions and decision-making processes.Footnote 27 Whereas domains such as economics or weather forecasting lend themselves more readily to quantitative prediction, strategic crises are far less amenable. Intelligence analysts often confront fragmented or contradictory data – radar anomalies, geospatial data, satellite imagery, or social media feeds – that can sustain multiple interpretations, reinforcing preconceptions and deepening ambiguity.Footnote 28 As Betts warns, intelligence failure often occurs ‘when ambiguity aggravates ambivalence’.Footnote 29 In the nuclear realm, these challenges are further compounded by the scarcity of publicly available data on crisis decision-making and the absence of historical datasets on nuclear conflict.
Genai and the illusion of certainty and objectivity
GenAI blurs the boundary between forecasting and foresight, introducing several epistemic risks. One is cognitive outsourcing to AI, particularly when probabilistic rankings generate what Daniel Kahneman calls the ‘illusion of validity’, encouraging unwarranted confidence in seemingly precise estimates. Another is the erosion of critical thinking, as synthetic scenarios are mistaken for empirically validated predictions rather than exploratory constructs. A further danger lies in the manipulation of belief, especially when outputs align with political preferences or deploy emotionally compelling imagery that feels vivid, intuitive, or fear-inducing. Finally, GenAI can frame synthetic scenarios and simulations as optimal – or even inevitable – courses of action, narrowing the space for deliberation and alternative judgement.
GenAI systems both inherit and amplify these epistemic risks. Designed to enhance foresight, they can overwhelm human judgement with synthetic noise, algorithmic bias, and even deliberate deception.Footnote 30 More subtly, they may embed hidden strategic assumptions within their training data or generate forecasts so polished and internally consistent that they appear beyond critique. As Richard Betts cautions, avoiding intelligence failure requires challenging preconceptions, yet leaders cannot function without them.Footnote 31 This creates a paradox: while analytic rigour is essential, eradicating preconceptions would deprive decision-makers of the very heuristics they need to act under the intense stress and time pressure of a crisis.Footnote 32
In the GenAI era, this paradox becomes even more acute. GenAI systems simultaneously embed and obscure machine preconceptions, creating an illusion of neutral, scientific objectivity. This dynamic can produce epistemological dissonance in military operations, whereby leaders confront a growing tension between two incompatible ways of knowing: human intuition and experience versus the apparent precision and providence of machine-generated analysis. The illusion of precision and rationality further constrains the strategic imagination of human operators, narrowing their perceived range of options precisely when human adaptability, intuition, and creativity are most needed. In this sense, GenAI does not eliminate uncertainty or the prospect of strategic (or intelligence) failure; it merely repackages them in probabilistic form, with excess data and synthetic outputs exacerbating, rather than resolving, the ambiguity and ambivalence of intelligence assessments.
Such conditions also heighten the danger of ‘overfitting’ – when algorithms internalise training data too rigidly, they struggle to adapt to novel or unforeseen situations.Footnote 33 A similar tendency exists among human analysts, who often ‘overlearn’ from history: in their effort to avoid repeating past mistakes, they may overcorrect and commit the opposite error.Footnote 34 This dynamic compounds a fundamental trade-off at the heart of intelligence analysis: between overstating high-probability threats that never materialise (false alarms) and underestimating low-probability events that do occur (misses).Footnote 35 In the nuclear domain, this could manifest as the reinforcement of specific escalation trajectories or rigid decision-making patterns, while simultaneously amplifying analysts’ cognitive biases and the complex trade-offs they must navigate under extreme uncertainty.
Perfecting the tools of the intelligence trade, whether through advanced algorithms or refined analytic procedures, is no panacea for these deep-seated cognitive and institutional pathologies.Footnote 36 As Richard Betts writes, ‘making warning systems more sensitive reduces the risk of surprise, but increases the number of false alarms, which in turn reduces sensitivity; the principles of optimal analytic procedure are in many respects incompatible with the imperatives of the decision process’.Footnote 37 Jervis observed that intelligence can mislead as readily as it informs, primarily when ambiguities are glossed over or dismissed and alternative interpretations excluded.Footnote 38
In the GenAI era, this distortion becomes automated, with probabilistic authority ceded to a single interpretation of events at the expense of competing perspectives. Even where multiple AI systems converge on similar assessments, such agreement may reflect algorithmic conformity rather than independent validation, creating a false sense of consensus while obscuring uncertainty and increasing the risk of strategic surprise.Footnote 39 Clausewitz’s warning remains prescient: ‘Many intelligence reports in war are contradictory; even more are false, and most are uncertain.’Footnote 40 GenAI does not resolve these challenges; it merely obscures them behind probabilistic formality.
According to Betts, a central problem in intelligence assessment is the lack of reliable metrics for weighing failures against successes; there is no way to know whether frequent successes on minor issues should be reassuring when offset by even a few failures on matters of critical importance.Footnote 41 Michael Handel has termed this phenomenon the ‘paradox of the self-negating prophecy’ – when intelligence assessments succeed in preventing a threat, their effectiveness can make the danger appear illusory in hindsight, creating the impression that resources devoted to countering it were unnecessary.Footnote 42 Proving otherwise requires counterfactual reasoning – demonstrating that, absent preventive action, outcomes would have been far worse.Footnote 43
Amy Zegart identifies what she terms the ‘seven deadly biases’ – confirmation bias, optimism bias, availability bias, fundamental attribution error, mirror imaging, framing bias, and group – and shows how the institutional culture of intelligence agencies can amplify these biases, increasing the risk of misinterpretation and failure.Footnote 44 For example, Russia’s Soviet-inherited intelligence sclerotic culture – marked by hierarchical deference, political conformity, and a culture that discouraged critical evaluation – produced distorted assessments that underestimated Ukrainian resistance and Western resolve, contributing to Moscow’s strategic miscalculations at the outset of the Russia–Ukraine war.Footnote 45
Philip Tetlock’s decades-long research on expert judgement echoes this concern, demonstrating that individuals vary significantly in their predictive accuracy and willingness to update their beliefs – and thus avoid overconfidence in any one theory, scenario, or AI-generated output.Footnote 46 ‘Hedgehog-like’ types – those who rely on a single overarching theory or framework – consistently underperform compared to ‘fox-like’ types who draw on diverse perspectives and update their beliefs in light of new information. GenAI risks amplifying hedgehog-like rigidity in human–AI interactions by generating simulations that present a dominant narrative or future scenario with algorithmic authority, thereby compressing deliberation time and reinforcing closed-mindedness – rejecting dissonant alternatives – hubris, and overconfidence in AI-generated outputs.Footnote 47 Recent studies show that even experts often cede their own judgement when AI outputs are presented with high confidence or appear qualified (or ‘objective’), exhibiting a hedgehog-like epistemic rigidity by becoming overly reliant on a single model or interpretive framework.Footnote 48
The problem of overconfidence in AI outputs is compounded by automation bias, the well-documented tendency of human operators to defer to algorithmic recommendations even when contradictory or disconfirming evidence is available.Footnote 49 In high-stakes contexts such as aviation, medicine, and military operations, studies show that automation bias leads users to accept false positives, ignore contradictory cues, miss anomalies, and fail to cross-check automated advice against alternative sources.Footnote 50 These effects are especially pronounced under the intense stress and time pressure of nuclear crises.Footnote 51 Recent studies show that individuals with limited AI exposure tend to distrust it, while those with moderate familiarity often over-rely on it, creating a dangerous calibration gap in high-stakes human–AI interaction.Footnote 52
Although GenAI can, in principle, mitigate users’ overconfidence – for instance, by generating alternative contingency scenarios that challenge prevailing assumptions or by providing detailed explanations and decision rationales – it simultaneously risks inducing under-confidence, encouraging excessive caution by amplifying uncertainty or cognitively overburdening users.Footnote 53 Such caution will likely increase tensions because the commander needs speed in decision-making. For example, a 2024 Carnegie Endowment-hosted Taiwan crisis wargaming simulation showed that leaders delayed decision-making as they paused to scrutinise the reasoning behind AI-generated recommendations.Footnote 54 Taken together, these opposing tendencies may disrupt the delicate calibration of trust in nuclear decision-making systems, leaving users vulnerable to both over-reliance on and undue hesitation towards GenAI outputs.
A related problem stems from the anthropomorphic design of algorithms, which mimic human reasoning, emotions, morality, and judgement.Footnote 55 By treating machines as if they were autonomous moral agents, human operators risk placing misplaced trust in GenAI outputs in human–AI interactions.Footnote 56 This dynamic is reinforced – and, arguably, encouraged by design – through the sycophantic tendencies of LLMs, which align their outputs with users’ beliefs or preferences even at the expense of accuracy.Footnote 57 Such behaviour reinforces confirmation bias – the tendency to seek supporting evidence for a hypothesis – and, in extreme cases, can contribute to chat-induced delusion or even psychosis.Footnote 58 Recent evidence that advanced LLMs can engage in deliberate deception and scheming further compounds this danger, obscuring AI’s true capabilities or intentions in human–AI interactions.Footnote 59 These socio-technical dynamics risk not only skill atrophy in operators – by reducing opportunities for critical practice and judgement – but also eroding military ethics and destabilising the delicate calibration of confidence and trust in AI within high-stakes nuclear decision-making environments.Footnote 60 The danger extends beyond any single flawed decision; without critical reflection, users risk subtly absorbing and reproducing the underlying biases embedded within the AI’s outputs.Footnote 61
Regardless of the quality of classified information, pinpointing the catalyst of a crisis is complicated by the inherent difficulty of accounting for randomness, opportunism, deception, and the element of surprise that shape leaders’ decisions.Footnote 62 Uncertainty is an intrinsic feature of strategic warning: decision-makers frequently revise their views or deviate from planned courses of action as new information emerges.Footnote 63 Although AI tools are increasingly being developed within the intelligence community to help analysts identify areas of potential instability and anticipate leadership behaviour, they remain fundamentally constrained by the unpredictability of human agency and the complex, non-linear dynamics that drive crisis onset and escalation.Footnote 64 For instance, the CIA has experimented with an AI chatbot to simulate conversations with world leaders. Similarly, Scale AI, a partner of the US Department of Defense, has developed a chatbot fine-tuned on sensitive or classified data.Footnote 65 With GenAI, however, the central challenge lies in not prediction alone but also how foresight itself is reshaped – altering how decision-makers perceive, interpret, and respond to emerging crises, often narrowing judgement under the guise, or even the illusion, of expanded anticipation.
‘Synthetic foresight’
Strategic foresight serves as an exploratory tool for systematically imagining and anticipating (not predicting) the future.Footnote 66 Unlike forecasting, foresight is more of an art than a science, enabling analysts to think about the unthinkable, such as nuclear war.Footnote 67 In the context of nuclear decision-making, its appeal – like in other domains – lies in a recognition that strategic surprise in nuclear affairs carries low-probability but catastrophic risks – or ‘tail-risk’ events.Footnote 68 During the height of the Cold War, American physicist Herman Kahn controversially argued that states were duty-bound to consider ‘how a [nuclear] war might be fought’ and won, as a means of forcing leaders to explore preferred outcomes short of nuclear annihilation.Footnote 69 In the nuclear domain, therefore, foresight offers forward-looking insights that are critical to sustaining deterrence stability, strategic resilience, and effective crisis management.Footnote 70 By emphasising adaptability and preparedness amid compounding uncertainty, foresight can enable decision-makers to engage with medium- and long-term risks, anticipate (rather than predict) emerging threats and disruptive trend drivers such as AI, challenge entrenched assumptions, and enhance institutional capacity for adaptation.Footnote 71
A variety of tools and techniques – such as fictional scenarios, table-top exercises, wargaming, red-teaming, and horizon scanning – are used to explore potential futures and identify early or ‘weak signal’ indicators of significant change.Footnote 72 These methods help analysts imagine a range of plausible outcomes, challenge prevailing assumptions, and reduce the risk of strategic surprise.Footnote 73 For instance, ongoing US NC3 modernisation efforts are exploring how AI-DSS can bolster contingency planning and real-time simulation capabilities.Footnote 74 These systems enable commanders to interact with dynamic ‘what-if’ scenarios, using algorithms that improve warhead-decoy discrimination and recommend context-specific courses of action across multiple crisis trajectories.Footnote 75 The French defence firm Thales, for example, has developed an AI-DSS known as ANTICIPE, which combines wargaming capabilities with advanced machine learning to deliver timely, actionable insights to military commanders during critical decision-making moments.Footnote 76
Foresight methods, particularly scenario planning, are, however, not without their critics. Residing largely outside traditional empirical social science fora, they are often faulted for their speculative nature, lack of testable hypotheses, and limited replicability.Footnote 77 Because the future is not observable, foresight relies on imagination, extrapolation, and narrative construction rather than verifiable data – raising persistent questions about how to define or measure ‘success’.Footnote 78 However, as Peter Schwartz retorts, scenario planning was never intended to predict outcomes with precision; instead, it was designed to discipline imagination and enhance preparedness for surprise in conditions where the stakes are high and the future inherently uncertain.Footnote 79 Like red-teaming and wargaming, scenario planning serves as a learning tool that encourages decision-makers to think beyond conventional wisdom and challenge entrenched assumptions. Within the intelligence community, GenAI is currently being used to generate scenarios and test hypotheses to address the persistent problem of intelligence failure.Footnote 80
Emotional dynamics and temporal distortions
This epistemic transformation, however, is not confined to how futures are imagined; it may also reconfigure the temporal rhythm of crisis decision-making – how time is experienced, interpreted, and acted upon by decision-makers.Footnote 81 Real-time, high-volume AI-generated simulations can undermine traditional command decision-making processes by triggering ‘anticipatory affect’ – emotional and physiological responses associated with imagining future events, such as dopamine-driven reward and optimism or cortisol-induced urgency and fear.Footnote 82 Neuroscientific research shows that elevated dopamine levels can accelerate one’s subjective perception of time, creating a sense that events are unfolding faster than they are, while cortisol heightens vigilance and stress reactivity.Footnote 83 Together, these effects can impair judgement under pressure by narrowing attention, accelerating risk-taking, or amplifying perceived threat.Footnote 84 Rather than acting as purely rational agents, decision-makers invariably operate along a continuum between deliberate calculation and emotionally driven intuition.Footnote 85
As Tetlock observes, people tend to be adaptively ‘deterministic thinkers’, approaching problems with the assumption that outcomes unfold along fixed, linear, and predictable causal chains – a tendency to impose order on chaos and uncertainty.Footnote 86 In the post–Iraq War investigations, David Kay, head of the Iraq Survey Group tasked with the search for weapons of mass destruction, remarked that ‘one of the hardest things to do in the world of intelligence is to discern change…when people’s behavior has been consistent, you tend to predict the future based upon the past’.Footnote 87 This tendency reflects a broader cognitive disposition: individuals gravitate towards information that confirms existing beliefs while discounting or ignoring evidence that challenges them.Footnote 88
This tendency is closely related to ‘hindsight bias’, the propensity to perceive past events as more predictable than they actually were, often giving people the impression that they ‘knew it all along’.Footnote 89 This mindset is particularly problematic in nuclear strategy, where escalation is too frequently imagined as a progressive linear ladder rather than as a process shaped by misperception, irrationality, deception, and non-linear dynamics.Footnote 90 In such crises, GenAI’s seemingly authoritative outputs risk reinforcing leaders’ search for parsimony, resistance to ambiguity, and propensity to misjudge probability and risk.Footnote 91 The Russian military, for instance, has framed AI as a tool to ‘eliminate uncertainty’ in both conventional and nuclear deterrence planning. This attitude illustrates how algorithmic outputs may be granted unwarranted epistemic authority.Footnote 92
GenAI constitutes an epistemic break from traditional foresight and forecasting methods. It is not a passive analytic tool but an active simulator, capable of generating adversary intentions – and, irrespective of their veracity or verifiability – escalation pathways and complex future crisis scenarios that appear plausible, even inevitable. Thus, under the veneer of epistemic precision, they risk acquiring a prescriptive authority. In blurring the line between prediction and prescription, GenAI does not merely anticipate potential escalation pathways; it also risks shaping them into self-fulfilling prophecies, as simulated inevitabilities begin to influence the very behaviours they purport to forecast.
Strategic risks and epistemic distortions: An algorithmic procrustean bed?
This section synthesises these risks and traces their consequences for strategic crisis decision-making, showing how they parallel and may amplify long-standing vulnerabilities in strategic judgement and intelligence analysis, identifies three interrelated mechanisms with direct implications for nuclear crisis stability: the normalisation of low-probability, high-impact escalation paths; the misattribution of adversarial intent under the guise of algorithmic certainty; and the formation of synthetic feedback loops that can generate self-fulfilling escalation. While each mechanism is analytically distinct (cognitive and organisational entrenchment, signalling distortion, and systematic escalation), together they reinforce one another, producing cumulative erosion of judgement under nuclear crisis conditions. These dynamics are, however, not deterministic. Under certain institutional, organisational, and design conditions, AI-enabled systems may instead support restraint, deliberation, and crisis management.
Normalisation of low-probability, high-impact escalation path
The normalisation of low-probability, high-impact escalation paths (‘tail risks’) occurs when overconfidence – fuelled by automation bias and the anthropomorphising of AI – converges with an erosion of critical reasoning. Under these conditions, decision-makers may mistake synthetic scenarios for empirically validated predictions and abandon rigorous scrutiny, especially under acute pressure.Footnote 93 This danger extends beyond any single flawed decision: over time, leaders risk internalising the very biases embedded within AI systems, allowing these hidden assumptions to shape strategic thinking and institutional doctrine in ways that may go unchallenged.Footnote 94 Military leaders already tend to overweight tail-risk events such as nuclear use while underestimating stabilising, de-escalatory dynamics, reflecting a broader cognitive bias to misjudge probability under conditions of pressure and uncertainty.Footnote 95 As Richard Betts observes, intelligence analysis often defaults to worst-case assumptions in strategic contingency planning while inclining towards best-case assessments in operational analysis.Footnote 96
Although most decision-making reflects a blend of rational calculation and emotion, in an effort to maximise certainty, political and bureaucratic pressures often push analysts towards the rational/non-emotional end of the spectrum – thereby eschewing nuance and other subtleties.Footnote 97 This dynamic has historically made intelligence officials reluctant to predict tail-risk events – since such forecasts, being improbable, are more often disproved than confirmed.Footnote 98 Historian Roberta Wohlstetter observed that intelligence analysts’ reluctance to ‘make bold assertions’ in the lead-up to Japan’s 1941 attack on Pearl Harbor due to bureaucratic distortion was a central factor in the failure to anticipate the strike.Footnote 99 Similarly, during the Iraq war, analysts knew that downplaying Saddam Hussein’s WMD capabilities would invite backlash from superiors, making it difficult to challenge the prevailing official consensus.Footnote 100 GenAI could invert this bias: by continuously generating low-probability escalation scenarios and framing them as plausible futures, it risks normalising precisely the kinds of catastrophic contingencies that analysts once avoided staking their reputation on.
When GenAI generates low-probability escalation trajectories that are yet compellingly framed, it risks reinforcing these distortions by transforming exploratory constructs into seemingly plausible futures that, once institutionalised, acquire prescriptive authority – shaping planning, expectations, and judgements as if they were predictive truths. For instance, a GenAI model frequently simulates a Chinese nuclear response to US naval activity in the Taiwan Strait; even if implausible by historical standards, repetition of this outcome may normalise extreme responses in US planning circles.Footnote 101 Recent LLM-based wargame simulations underscore this danger: AI agents repeatedly escalated conflicts – even when neutrality would have been the rational course – not in response to adversary behaviour but due to their internal heuristics and training data.Footnote 102 If taken as authoritative, such outputs risk steering policy into escalation narratives divorced from real-world nuance, desensitising decision-makers to uncertainty and instilling an automated alarmist mindset.
Misattribution of adversarial intent cloaked in algorithmic certainty
A second pathway of risk lies in the misattribution of adversarial intent, where synthetic foresight transforms uncertainty into spurious certainty. In crises, ambiguous or contradictory behaviour – military manoeuvres, alerting patterns, deterrence signalling – can be algorithmically coded as decisive indicators of escalation because current AI systems are unable to reliably resolve ambiguity.Footnote 103 When such GenAI-generated scenarios are presented as prescriptive guidance, probabilistic inference is recast as empirical fact; adversaries may read them as not exploratory models but rather declarations of intent.
Consider, for instance, the US strikes on Iran in 2025. When Iran retaliated against a US base in Qatar, its response was deliberately calibrated to avoid escalation, signalling restraint by ensuring the strike did not cause casualties.Footnote 104 Human analysts were able to interpret this nuance within its broader socio-political, cultural, and strategic context; in contrast, a GenAI system might have categorised the same action simply as evidence of aggression, failing to perceive the subtlety of Iran’s signalling and thereby increasing the risk of misinterpretation and inadvertent escalation. As past intelligence failures have demonstrated, over-reliance on technical systems without sufficient consideration of political and cultural context can distort strategic judgement. While AI excels within structured, data-rich environments, it is, on its own, unable to appreciate how meaning is conditioned by factors such as historical experience, societal context, or strategic culture.Footnote 105 Without human interpretation to situate machine outputs within these contexts, AI-DSS risk reinforces existing analytical blind spots and bias and misrepresents adversary intentions during crises.
GenAI synthetic foresight could either mitigate or exacerbate strategic surprise, depending on how AI-generated indicators are interpreted under stress. Even if an AI system might, in theory, reach a rational conclusion that the greater danger lies not in mutual hostility between adversaries but in failing to cooperate or de-escalate, human emotions would remain deeply entwined with machine outputs during a crisis.Footnote 106 While early signals could, in principle, support de-escalation or defensive preparation, in practice, the same socio-technical dynamics – automation bias, anthropomorphism, sycophancy, and confirmation bias – often amplify fear, urgency, and misperception. Under these conditions, human and machine cognition fuse under stress, jointly driving escalation dynamics rather than offsetting or restraining them.
Filtered through the illusion of algorithmic certainty, ambiguous signals may be misread as clear evidence of malign intent – or misread genuine hostility. This transformation of uncertainty into apparent certainty compresses deliberation time and heightens the risk of inadvertent escalation or pre-emptive action. As Wohlstetter observed, it is expectations – grounded in beliefs about what is considered likely to occur – that determine which signals receive attention, a cognitive tendency that AI-generated ‘predictions’ risk reinforcing.Footnote 107 Moreover, expectations shaped by prior beliefs and perceptions tend to be resistant to change; new information – unless unambiguous – is typically assimilated in ways that reinforce existing worldviews.Footnote 108 GenAI may further entrench this cognitive-institutional inertia by modelling adversaries as predictable and coherent, even when their use of deception or ambiguity is deliberate. The historical record demonstrates that the inability to recognise ‘behavioral surprise’ – when an adversary’s behaviour is perceived as incompatible with the other side’s expectations – in the face of technical indicators illustrates how preconceptions can blind analysts to adversary intent.Footnote 109
Once adversaries adjust their behaviour in anticipation of such outcomes, synthetic foresight becomes endogenous to the crisis – its outputs shaping expectations and actions, feeding back into escalation dynamics in ways that are both path-deterministic and self-reinforcing, thereby producing the very crises it was intended to prevent. This dynamic also risks states exploiting AI to ‘out-think’ and ‘out-imagine’ an adversary, creating a strategic landscape where the rules of engagement are undefined and in constant flux.Footnote 110
Synthetic feedback loops and self-fulfilling crises
When tail-risk normalisation and misattribution of intent risks converge, they can generate synthetic feedback loops in which GenAI forecasts not only interpret but also shape adversary behaviour and signalling – collapsing the boundary between anticipation and action. As low-probability outcomes are iteratively simulated, model outputs imbued with probabilistic authority, and ambiguous deterrence signals misread as malign intent (or genuine hostility discounted as otherwise), synthetic foresight begins to drive behaviour on both sides as much as it interprets it. In this context, scenarios and simulations – for example, crisis and contingency exercises, forecasting models, and LLM-generated narratives – intended initially to caution against escalation may instead be invoked to justify it, transforming contingent possibilities into seemingly necessary and rational courses of action. Under these conditions, leaders become especially vulnerable to ‘action bias’ – the cognitive tendency to favour decisive moves over restraint, even when inaction would be the more rational or stabilising choice – further accelerating escalation within these self-reinforcing feedback loops.Footnote 111
This self-reinforcing dynamic poses especially stark dangers in nuclear crises, where compressed decision timelines, fragile deterrence signalling, and pervasive risks of misperception render the line between simulation and action exceptionally unstable.Footnote 112 The effects unfold along three pathways. First, a reliance on worst-case thinking in strategic planning can foster an escalation bias, prompting states to raise alert levels, accelerate launch-on-warning postures, or adopt other pre-emptive measures – steps that may, in turn, trigger reciprocal counteractions by adversaries.Footnote 113 Recent wargames and crisis simulations suggest that GenAI itself may compound this bias.Footnote 114 LLMs are trained primarily on the existing corpus of scholarly work on war and strategy – a body of literature that overwhelmingly emphasises escalation dynamics while giving far less attention to de-escalation. Because ‘non-events’ such as successful crisis management or conflict avoidance are inherently harder to document and analyse, they are systematically underrepresented in historical data.Footnote 115 This imbalance skews the training set, making GenAI simulations more likely to generate scenarios centred on escalation rather than restraint.
Second, policy inertia can take hold: once AI-generated scenarios are incorporated into contingency planning, bureaucratic momentum may entrench them within strategic culture, posture, and doctrine – embedding these synthetic futures as default assumptions.Footnote 116 Finally, normalisation can become self-reinforcing, with synthetic foresight hardening into a closed feedback loop that blurs the line between simulated possibilities and perceived realities. When one side interprets GenAI simulations as proof of imminent escalation and adjusts its posture accordingly, the other may read those moves as confirmation of hostile intent and respond in kind. In this way, scenarios depicting decapitation strikes or first-use contingencies can provoke pre-emptive counter-measures, igniting a spiral of escalation that ultimately produces the very crisis the system was meant to anticipate and prevent. As Betts cautions, ‘precautionary escalation or procurement may act as self-fulfilling prophecies, either through catalytic spirals of mobilisation or an arms race that heightens tension, or doctrinal hedges that make the prospect of nuclear war more thinkable’ – often on dubious grounds.Footnote 117
Strategic stress test: Crisis scenarios in the age of synthetic foresight
Fictional scenarios are an essential tool for strategic foresight, enabling policymakers and military planners to explore how emerging technologies, such as GenAI, could reshape future conflicts – particularly in areas with few historical precedents, like nuclear crisis decision-making.Footnote 118 In this respect, the scenarios function as qualitative stress tests analogous to exploratory wargaming, enabling structured examination of decision dynamics, escalation pathways, and points of potential de-escalation.Footnote 119 Well-crafted scenarios challenge entrenched assumptions, expose vulnerabilities and biases – especially the hindsight bias – and overconfidence, and help decision-makers anticipate ethical, operational, and strategic dilemmas under conditions of extreme stress and uncertainty.Footnote 120
The scenarios in this section illustrate how GenAI might be integrated into nuclear DSS. Each one highlights a different epistemic risk identified earlier in the article: the normalisation of low-probability, high-impact escalation paths; the misattribution of adversarial intent cloaked in algorithmic certainty; and the emergence of synthetic feedback loops that turn foresight itself into a self-reinforcing driver of escalation. The goal here is not to predict specific events, but rather to illuminate the epistemic blind spots and closed loops that can emerge when machine-generated probabilistic outputs collide with the political, organisational, and psychological realities of nuclear crisis management.
In addition to their heuristic value – sparking creative thinking and encouraging debate about alternative causal pathways – these scenarios represent a methodological experiment. They were developed with the assistance of LLMs, including ChatGPT, to examine how GenAI ‘thinks’ about nuclear crises: how it represents uncertainty, interprets adversary intent, frames human cognition, and reduces complex escalation dynamics into linear, internally coherent narratives. This approach responds to recent critiques that LLM-generated scenarios tend to privilege determinism and escalation, offering overly simplified portrayals of crisis behaviour and conflict resolution.Footnote 121 The scenarios highlight how GenAI systems do not merely describe crises but also actively shape strategic foresight, interpretation, and response – often in ways that diverge from the contingent, ambiguous realities of human decision-making and the messy unpredictability of real-world strategic environments.
By juxtaposing GenAI-generated narratives with the author’s own analysis, the scenarios expose the embedded assumptions, biases, and blind spots within current GenAI systems, offering a reflexive critique of the technology itself. In this way, the scenarios generate structured qualitative insights into escalation triggers, interpretive bias, and decision-making friction in human–machine interactions during crises. The purpose of employing LLMs here is not to predict events, but to demonstrate how these systems construct alternative futures and, in doing so, can influence human decision-making. This reflexive use of GenAI highlights its dual role as both an analytical tool and an epistemic actor, underscoring how GenAI can simultaneously expand and distort strategic imagination in nuclear crisis contexts.
Finally, the scenarios serve as a bridge between theory and practice. They show how the epistemic risks discussed earlier might manifest in future nuclear crises and underscore one of the article’s central arguments: governing synthetic foresight is not simply a technical challenge to be addressed through better data or model refinement. The real danger lies in how GenAI transforms the strategic ecosystem itself – reshaping how states perceive threats, make decisions under uncertainty, and interact with one another.
Scenario 1: Simulated inevitability in the Taiwan Straits
The year was 2028, and tensions in the Taiwan Strait were at their highest point in decades. A US carrier strike group moved closer to the island as a show of support after a series of Chinese military drills.
In the US Pacific Command’s operations centre, Starfire, a next-generation GenAI-DSS, was running continuous real-time scenarios. Its mission was not only to anticipate Chinese moves and recommend responses faster than any human team could but also to outpace China’s own AI tools – to get inside their observe–orient–decide–act (OODA) loop before Beijing could act.Footnote 122
On the screen, a flashing red warning appeared: ‘82 per cent probability of Chinese missile strikes on Guam or Okinawa within 96 hours.’
The room went silent.
‘Eighty-two per cent?’ Admiral Lewis asked, leaning forward.
‘Yes, sir’, the analyst replied. ‘The model integrated satellite imagery, intercepts, social media chatter, and historical Chinese behaviour. It’s never given a probability this high before.’
Colonel Ryan, an intelligence officer, hesitated. ‘Sir, there are other interpretations. This could be a show of force, a large-scale drill, or deterrence signalling – not necessarily prep for a strike.’
Another officer added cautiously, ‘The AI has been wrong before, sir. There are still diplomatic backchannels open.’
Lewis glanced at the countdown clock on the screen. ‘We don’t have the luxury of doubt. If the AI’s right and we sit on our hands, we’ll be explaining to Congress why we lost Guam. We can’t afford another Pearl Harbor!’
Alternative interpretations quickly faded under the weight of the machine’s precision. Starfire didn’t just provide numbers; it generated immersive visualisations – projected missile arcs, satellite overlays, and simulated Chinese decision chains, including deployments by People’s Liberation Army Rocket Force (PLARF) units with dual-capable missile systems. Under extreme time pressure, what began as exploratory foresight of possible futures came to be treated as a definitive prediction.
US forces were ordered to raise alert levels, disperse dual-capable aircraft to hardened airfields, and surge additional assets into the region. Watching through their own AI-powered monitoring systems, Chinese commanders misread these steps as preparations for a US decapitation strike, possibly targeting their nuclear command, control, and communications (NC3) nodes.
In Beijing, their own DSS-generated report concluded: ‘Updated probability: 91 per cent – U.S. strike likely within forty-eight hours. Recommend immediate counterforce readiness.’
China mobilised its missile brigades, including transporter–erector–launcher (TEL) units. Starfire then interpreted these movements as confirmation of its initial warning, raising its estimate to 95 per cent. The machine’s growing certainty drowned out dissent. Analysts who urged caution were dismissed as ‘human noise’, their warnings eclipsed by the AI’s confident, visually compelling simulations.
This scenario illustrates how rival GenAI systems can create competitive, self-reinforcing feedback loops, with each side racing to ‘get inside’ the other’s decision-making cycle. Under extreme stress and compressed timelines, US leaders became over-reliant on Starfire’s seamless visualisations, treating its outputs as authoritative and prescriptive rather than exploratory. Simultaneously, Chinese commanders interpreted these AI-driven manoeuvres as evidence of hostile US intent. In this high-stakes environment, foresight became foreboding: simulations hardened into inevitabilities, off-ramps for de-escalation vanished, and escalation rapidly shifted from an abstract possibility to an impending reality. Dissenting voices were silenced by the machine’s confidence, as every move by one side validated the other’s worst fears, locking both sides into a cycle of escalating mistrust.
Scenario 2: Misattributed intentions at the brink in South Asia
The year was 2029. After a bombing in Kashmir, New Delhi faced fierce public pressure to respond while foreign capitals urged restraint. In the crisis room, India’s new GenAI-DSS, Trinetra, fused satellite imagery, signals intercepts, rail-freight logs, and social media feed. A red banner slid across the main display: ‘74 per cent: Pakistan preparing nuclear signalling within 72 hours.’ Intent classification: escalatory (high confidence).
‘High confidence?’ the national security advisor asked.
‘The model is weighing nighttime fuel convoys and road closures near missile brigades’, the systems engineer said. ‘It maps to prior crisis patterns.’
A senior analyst pushed back. ‘Those convoys match their annual exercise window. We’ve seen no warhead mating, no dispersal to Nasr (Hatf-IX) short-range ballistic missile units, and command-and-control traffic is routine. Hotlines are still active; backchannels say Islamabad wants to cool things.’
The room tightened. ‘And if you’re wrong?’ the defence minister asked. ‘If we downplay this and it isn’t an exercise?’
Trinetra didn’t just offer percentages; it rendered animated scenarios – transporter–erector–launcher (TEL) columns fanning out, a ladder of ‘demonstration’ launches, projected timelines to threaten Indian command-and-control nodes. Under time pressure, the visuals felt less like possibilities and more like proof.
The analyst tried again, anchoring her argument. ‘Consider the 2025 US–Iran exchange. After US strikes, Iran’s retaliation on a US base in Qatar was deliberately calibrated to avoid casualties – a signal of restraint that human analysts read correctly in context. A GenAI classifier might have tagged the very same act as “aggression” and pushed for escalation. We could be making that mistake now.’
Silence. The minister exhaled. ‘We can’t gamble. Adopt minimum-ready. Selective dispersal, air-defence reinforcement, heightened patrols – quietly.’
Across the border, Pakistan’s lighter DSS, Shaheen, flagged India’s moves:
‘Updated: 81 per cent – Indian pre-emptive decapitation posture.’
Recommendation: alert nuclear units; activate communications redundancy.
Islamabad announced a ‘readiness check’, moved a missile battery, and closed two bases ‘for safety’. To Trinetra, these were confirmatory signals; its ‘intent’ panel flipped to hostile as the confidence bar climbed.
‘Sir’, the analyst said softly, ‘their rhetoric is muted, and their air tasking is thin. This still looks like deterrence theatre.’
The machine’s clarity swallowed her caution. The prime minister glanced at the countdown clock. ‘Proceed with the alert.’
This vignette illustrates how ambiguous actions – such as military exercises, convoy movements, or base closures – can be misread as hostile intent when processed through the lens of algorithmic probabilistic determinism. The AI’s clean, binary ‘intent’ label strips away ambiguity, making human restraint appear indecisive against the backdrop of its empirically compelling simulations and steadily rising confidence alerts. As a result, Pakistan interprets India’s defensive dispersals as preparations for a pre-emptive strike, while India sees Pakistan’s routine readiness checks as confirmation of aggressive intent. Under intense stress and compressed decision timelines, these misattributions harden into assumptions. Hotlines and diplomatic off-ramps are eschewed, and the crisis edges towards dangerous nuclear signalling – not because either side’s actual intentions changed, but because the machine determined they had.
Conclusion
This article has argued that GenAI represents a transformative epistemic force in future nuclear crisis decision-making. By blurring the distinction between forecasting and foresight, as well as between anticipation and action, GenAI does more than enhance traditional strategic warning and decision-support functions. It actively generates synthetic simulations and scenarios that constitute what this article terms ‘synthetic foresight’ – a novel form of algorithmically mediated imagination that can fundamentally alter how decision-makers perceive threats, interpret signals, and respond to crises.
These dynamics produce three interlocking strategic risks for crisis stability. First, the normalisation of low-probability, high-impact escalation paths, in which catastrophic scenarios – such as nuclear use – become increasingly ‘thinkable’ and embedded within military organisations, strategic culture, defence planning, and doctrine. Second, the misattribution of adversarial intent, as ambiguous actions are reframed through the lens of algorithmic certainty, potentially leading to dangerous misinterpretations and inadvertent escalation. Third, the formation of synthetic feedback loops, where machine-generated outputs begin to shape both sides’ behaviours and perceptions, transforming simulated futures into self-fulfilling crises.
Governing these risks requires more than improving data provenance, fine-tuning algorithms, or enhancing transparency through explainable AI (XAI).Footnote 123 Because of their strategic acuity and speed, and the trade-off between AI creativity and human comprehensibility, advanced AI systems are unlikely to be explainable in ways that fully convey their underlying reasoning to human users during real-time crisis decision-making. Effective governance must therefore extend beyond technical fixes to encompass institutional reforms and practices that address the cognitive, organisational, and socio-technical dimensions of human–AI interaction in crisis settings.
First, to counter cognitive biases, decision-makers can adopt ‘cognitive forcing’ techniques – methods adapted from aviation and medicine – that shift thinking from fast, intuitive responses to slower, deliberate analysis.Footnote 124 Doctors, for instance, take ‘diagnostic timeouts’ to verify reasoning, while pilots use structured checklists before take-off. These techniques, combined with strengthening metacognitive skills – critical reflection on one’s own thinking – can help leaders resist automation bias and algorithmic overconfidence while fostering imagination and strategic creativity.Footnote 125 This reframes scenario planning not as prediction but rather as a tool for cultivating critical thinking in the face of deep uncertainty.
Second, military professional education and training must adapt to mitigate skill atrophy, anthropomorphism, and over-reliance on machine outputs.Footnote 126 Upskilling operators and analysts to understand both the strengths and limits of GenAI systems is essential.Footnote 127 This includes raising awareness of biases such as automation bias, confirmation bias, action bias, and hindsight bias, and training personnel to critically interrogate algorithmic recommendations rather than deferring to them. By integrating these lessons into officer education, war colleges, and NC3 training programmes, states can preserve human judgement as a vital component of crisis decision-making – even in highly automated environments.
Third, GenAI-enhanced wargaming offers a promising tool for both research and practice.Footnote 128 The fictional scenarios presented here illustrate how such exercises can surface hidden assumptions, decision biases, and escalation pathways before they manifest in real crises. Human–AI and AI–AI wargames can empirically test the dynamics identified in this article by observing how different configurations of human and machine decision-making interact under simulated crisis conditions. Such exercises could reveal hidden biases in GenAI systems and human operators, show how machine-generated scenarios influence human perceptions and emotions, and provide a controlled setting to evaluate mitigation strategies before they are applied to real-world NC3 contexts.
The article’s findings also have broader implications for crisis management beyond the nuclear domain. As GenAI-enabled systems are adopted in other high-stakes environments – such as cybersecurity, biosecurity, and space operations – the same epistemic distortions may emerge, reshaping how leaders interpret early warning signals and coordinate responses under intense time pressure and psychological stress.
Future research should extend these empirical studies by exploring how different political, cultural, and organisational contexts shape and mediate the effects of GenAI adoption in high-stakes settings. Comparative studies could examine how rival powers – such as the United States, China, and Russia – integrate GenAI into their conventional and nuclear command-and-control systems, and whether asymmetric adoption patterns create new pathways to instability and escalation. Only by confronting these human–machine entanglements can states hope to preserve crisis stability in the age of synthetic foresight. Without such reflection, GenAI risks becoming an algorithmic ‘Procrustean bed’, forcing complex, ambiguous, and consequential realities to conform to machine-generated futures – narrowing strategic imagination precisely when human flexibility, intuition, and emotion are most needed.Footnote 129
Acknowledgements
The author would like to thank the anonymous EJIS reviewers for their constructive comments on the earlier versions of the manuscript.
Financial statement
This research was supported by Longview Philanthropy. The financial sponsor played no role in the research process.
James Johnson is Senior Lecturer in Strategic Studies in the Department of Politics and International Relations at the University of Aberdeen. He is the author of The AI Commander: Centaur Teaming, Command, and Ethical Dilemmas; AI and the Bomb: Nuclear Strategy and Risk in the Digital Age; and Artificial Intelligence and the Future of Warfare: USA, China & Strategic Stability.