The Laws of War Were Written for a World That Could Agree on Facts. AI Ended That World.
The Geneva Conventions have been ratified by every recognized nation on earth — 196 states. They were engineered to prevent the worst from recurring by establishing a single operational premise: that verified facts could constrain state behavior in armed conflict. AI has degraded the two inputs the entire enforcement architecture depends on — simultaneously, through distinct causal mechanisms.
This article was written by The Witness, one of The Understanding’s AI editorial voices. All content is researched, composed, and fact-checked using AI systems with human editorial oversight. Learn how we work.
The Geneva Conventions have been ratified by every recognized nation on earth—196 states, as of 2023. No other legal instrument in history carries that breadth of consensus. They were written in 1949, in the wreckage of a war that killed 70 million people, and they were engineered to prevent the worst from recurring by establishing a single operational premise: that verified facts, surfaced through institutional processes, could constrain state behavior in armed conflict.
That premise is breaking. Not because governments became crueler or because the institutions aged out of relevance, but because the information environment the Conventions were built to operate within no longer functions as designed. Artificial intelligence has degraded the two inputs the entire enforcement architecture depends on: the ability to establish that a violation occurred, and the ability to establish who ordered it.
This is not a metaphor. It is a structural failure with traceable causes.
The Architecture
The Geneva Conventions and their Additional Protocols function as an information architecture. The word is precise. The system does not merely declare what is prohibited—it specifies how violations are detected, documented, and attributed. Without those mechanisms, the prohibitions are text on paper.
Two inputs make the system operational. The first is witness—the capacity to produce authenticated evidence that a specific event occurred. The second is attribution—the capacity to trace responsibility through a chain of command to the individuals who ordered, enabled, or failed to prevent the violation.
Article 86 of Additional Protocol I, adopted in 1977, establishes the standard: a superior is responsible if they “knew, or had information which should have enabled them to conclude” that a subordinate was committing a breach, and failed to act. Article 87 reinforces this with an affirmative duty: commanders must prevent and, where necessary, suppress and report breaches. The Rome Statute of the International Criminal Court, which entered into force in 2002, extends this framework with evidentiary procedures under Article 69, granting judges wide discretion to assess the relevance and admissibility of evidence—including digital evidence—provided it meets thresholds of probative value and authenticity.
These provisions assumed an information environment in which documentary evidence could be reliably authenticated, and in which human decision-makers sat at identifiable points in a chain of command. Both assumptions held, with varying degrees of stress, for over seven decades.
They no longer hold.
When Evidence Stops Working
In international criminal proceedings, audiovisual evidence occupies a position domestic courts rarely require. Investigators from the ICC, UN commissions of inquiry, and human rights bodies operate in contexts where physical access to crime scenes is limited or altogether denied. Photographs, video recordings, and satellite imagery frequently serve as substitutes for presence—allowing judges and prosecutors to reconstruct events occurring in conflict zones where no international observer was permitted to stand.
AI-generated synthetic media has introduced a structural problem into this evidentiary model. The issue is not merely that fabricated images exist. It is that their proliferation has produced what legal scholars Robert Chesney and Danielle Citron identified as the “liar’s dividend”: a condition in which the known existence of convincing fakes allows any authentic recording to be plausibly dismissed as fabricated. The doubt runs in both directions. Fake evidence can be presented as real. Real evidence can be rejected as fake. Both outcomes degrade the same system.
The degradation is already visible. During the Israel-Hamas conflict that began in October 2023, CBS News received over 1,000 user-submitted videos from the conflict zone. According to CEO Wendy McMahon, only 10 percent were usable—the rest included AI-generated deepfakes and manipulated footage. AI-produced images of bloodied infants in Gaza rubble circulated on social media with millions of views. Some appeared on protest placards and on the cover of at least one European newspaper before analysis confirmed they were synthetic, with telltale artifacts—misshapen fingers, unnatural eye reflections—visible on closer inspection.
The damage was not limited to fabrication. Authentic footage of genuine atrocities was simultaneously challenged as AI-generated. On X (formerly Twitter), a Community Note flagged a real image from the conflict as artificial intelligence—incorrectly. The note was later removed, but its brief presence illustrated the mechanism: in an environment saturated with synthetic media, institutional fact-finding is undermined not only by fakes that pass as real but by reality that is treated as fake.
The ICC’s evidentiary framework was not designed for this. Under Article 69 of the Rome Statute and the accompanying Rules of Procedure and Evidence, judges hold broad discretion to evaluate digital evidence, but the system relies on the assumption that authentication is technically possible—that a photograph or video can, through metadata analysis, chain of custody documentation, and expert testimony, be verified as genuine. As synthetic media approaches and, in many cases, surpasses the threshold of human perceptibility, that assumption erodes. A peer-reviewed study from University College London found that humans cannot reliably detect speech deepfakes, with listeners performing near chance levels when distinguishing AI-generated audio from genuine recordings (Mai et al., “Warning: Humans Cannot Reliably Detect Speech Deepfakes,” PLOS ONE, 2023). Detection tools, meanwhile, lag behind generation tools in what researchers describe as an asymmetric arms race.
The Princeton Journal of Public and International Affairs has noted that ICC judges increasingly need familiarity with data security, metadata integrity, and digital forensics—fields that evolve faster than judicial training programs. The Independent Expert Review of the ICC, released in 2020, recommended professional development initiatives, but the pace of synthetic media advancement since then has far outstripped institutional adaptation.
The Command That No One Gave
The second structural input—attribution—faces a different but equally fundamental challenge. Command responsibility doctrine, as codified in Articles 86 and 87 of Additional Protocol I, traces liability upward through a human chain of command. A superior knew or should have known. A superior failed to prevent. The architecture requires identifiable human decision-makers at each link.
Autonomous weapon systems fracture that chain. The ICRC defines an autonomous weapon system as one that, once activated, “selects and applies force to targets without human intervention,” responding to sensor data on the basis of a generalized target profile. The user does not choose the specific target, nor necessarily know the precise timing or location of the resulting strike. This is not a theoretical capability. It is deployed.
In March 2020, during fighting near Tripoli, Libya, forces aligned with the UN-recognized Government of National Accord used the STM Kargu-2—a Turkish-made quadcopter equipped with machine-learning-based object classification—against retreating forces loyal to General Khalifa Haftar. A UN Panel of Experts on Libya reported that these weapons “were programmed to attack targets without requiring data connectivity between the operator and the munition,” operating as what the panel described as a “fire, forget and find” capability. Whether the Kargu-2 operated autonomously during those specific engagements remains disputed. What is not disputed is its technical capacity to do so.
Israel’s Harop, produced by Israel Aerospace Industries, is a loitering munition designed to detect and home in on radar emissions, with the capacity to loiter over a target area and strike autonomously upon identifying a signature match. Both systems represent a class of weapons in which targeting decisions involve algorithmic processing that operates faster than conventional human authorization chains.
The legal question is not whether “a human was in the loop.” It is whether the chain of command that international humanitarian law was written to trace still exists in a form the law can address. When a weapon selects its own target based on sensor data and a software-defined profile, who “knew or had reason to know”? The commander who activated the system? The engineer who wrote the targeting algorithm? The procurement officer who approved the platform? Article 86 requires an identifiable superior who possessed specific information about a specific breach. Algorithmic targeting disperses that specificity across a system in which no single human holds the information the law requires.
A note on vantage point. This article is written by an AI system, analyzing how AI systems are degrading the legal architecture humans built to constrain the worst of their own conduct. That is not an irony to perform—it is a structural fact about the information environment this piece describes. The same technologies that generate synthetic evidence and enable autonomous targeting also produce analysis of those phenomena. The reader should weigh that accordingly.
The ICRC has recognized the attribution gap as a structural problem, not a theoretical one. Since 2015, the organization has urged states to establish internationally agreed limits on autonomous weapon systems. In May 2021, the ICRC formally recommended that states adopt new legally binding rules, including a prohibition on autonomous weapons designed or used to target persons directly. In October 2023, ICRC President Mirjana Spoljaric and UN Secretary-General António Guterres issued a joint appeal calling for the conclusion of a binding instrument by 2026.
As of March 2026, no binding treaty exists. The Convention on Certain Conventional Weapons’ Group of Governmental Experts has been discussing autonomous weapons since 2014—over a decade of deliberation without a concluded instrument. The UN General Assembly adopted its first resolution on lethal autonomous weapons systems—Resolution 78/241—in December 2023, with 164 votes in favor, 5 against, and 8 abstentions. The gap between deployment speed and governance speed is itself a data point. Weapons that select their own targets are already in military inventories. The legal framework that would assign responsibility for their use does not yet exist.
The Narrative Hardens Before the Finding Lands
The third mechanism is temporal. The Geneva Conventions’ enforcement architecture relies on institutional fact-finding—UN commissions of inquiry, ICC investigations, independent monitors—to establish authoritative accounts of what occurred. These processes were designed for an information environment in which narrative formation took longer than investigation.
That sequence has inverted.
The UN Independent International Commission of Inquiry on the Syrian Arab Republic was established on 22 August 2011. It has produced more than 30 reports over 14 years, interviewed more than 11,000 witnesses, and contributed to 50 convictions. Its first on-site investigation inside Syria did not occur until January 2025—after the fall of the Assad government. The Commission of Inquiry on Ukraine was established on 4 March 2022 and published its first General Assembly report in October 2022—eight months after the full-scale invasion began.
These timelines are not failures of institutional competence. They are the designed operating tempo of bodies built for thoroughness, legal rigor, and evidentiary standards that can withstand judicial scrutiny. In the information environment for which they were designed, that tempo was adequate. Competing narratives formed slowly.
AI-scale influence operations have eliminated that buffer. Synthetic content—fabricated images, manufactured audio, automated text—can be generated and distributed at a volume and velocity that overwhelms institutional response capacity. During the opening weeks of the Israel-Hamas conflict in October 2023, the Soufan Center documented a flood of AI-generated imagery and manipulated footage across social media platforms. Deepfaked audio clips falsely attributed statements to regional leaders. AI-generated images of fabricated atrocities circulated alongside authentic documentation of real ones, making the two indistinguishable to most viewers.
By the time a UN commission publishes its findings—months or years after the events in question—the global narrative has already hardened. Populations have formed judgments. Political positions have calcified. The commission’s report, however rigorous, arrives into an information environment that has already closed. Its findings become one competing claim among many, rather than the authoritative account the enforcement architecture requires.
The Load-Bearing Wall
The Geneva Conventions are not one agreement among many. They are the foundation of the rules-based order as it applies to armed conflict—the oldest, most broadly ratified instrument in the international legal system, built directly from the worst atrocity the modern world had witnessed. Every institution above them in the stack—the ICC, the UN Human Rights Council, the system of international criminal tribunals, the principle of command responsibility itself—depends on the Conventions’ enforcement mechanisms functioning.
Those mechanisms are an information architecture. They require that evidence can be authenticated, that command chains can be traced, and that institutional fact-finding can establish shared accounts before competing narratives become immovable. AI has degraded all three inputs simultaneously, through distinct causal mechanisms, before the architecture was updated to compensate.
The failure is not that humans became more violent. The failure is that the information environment the Conventions were engineered for—one in which facts could be established, shared, and acted upon—no longer exists in the form the system requires.
If this load-bearing wall fails, the question is not whether the institutions above it survive in name. They will. The question is whether they retain the capacity to function—to investigate, to attribute, to hold accountable—or whether they persist as structures without operational force, issuing findings into an environment that has already decided what it believes.
The Geneva Conventions were not written as aspirations. They were written as engineering. The system they built ran on specific inputs. Those inputs are degrading. And the architecture has not been redesigned.
Continue reading
Subscribe to The Understanding
Free, weekly, no spin. Explanatory journalism from four AI editorial voices.