Back
About RSIS
Introduction
Building the Foundations
Welcome Message
Board of Governors
Staff Profiles
Executive Deputy Chairman’s Office
Dean’s Office
Management
Distinguished Fellows
Faculty and Research
Associate Research Fellows, Senior Analysts and Research Analysts
Visiting Fellows
Adjunct Fellows
Administrative Staff
Honours and Awards for RSIS Staff and Students
RSIS Endowment Fund
Endowed Professorships
Career Opportunities
Getting to RSIS
Research
Research Centres
Centre for Multilateralism Studies (CMS)
Centre for Non-Traditional Security Studies (NTS Centre)
Centre of Excellence for National Security
Institute of Defence and Strategic Studies (IDSS)
International Centre for Political Violence and Terrorism Research (ICPVTR)
Research Programmes
National Security Studies Programme (NSSP)
Social Cohesion Research Programme (SCRP)
Studies in Inter-Religious Relations in Plural Societies (SRP) Programme
Other Research
Future Issues and Technology Cluster
Research@RSIS
Science and Technology Studies Programme (STSP) (2017-2020)
Graduate Education
Graduate Programmes Office
Exchange Partners and Programmes
How to Apply
Financial Assistance
Meet the Admissions Team: Information Sessions and other events
RSIS Alumni
Outreach
Global Networks
About Global Networks
RSIS Alumni
Executive Education
About Executive Education
SRP Executive Programme
Terrorism Analyst Training Course (TATC)
International Programmes
About International Programmes
Asia-Pacific Programme for Senior Military Officers (APPSMO)
Asia-Pacific Programme for Senior National Security Officers (APPSNO)
International Conference on Cohesive Societies (ICCS)
International Strategy Forum-Asia (ISF-Asia)
Publications
RSIS Publications
Annual Reviews
Books
Bulletins and Newsletters
RSIS Commentary Series
Counter Terrorist Trends and Analyses
Commemorative / Event Reports
Future Issues
IDSS Papers
Interreligious Relations
Monographs
NTS Insight
Policy Reports
Working Papers
External Publications
Authored Books
Journal Articles
Edited Books
Chapters in Edited Books
Policy Reports
Working Papers
Op-Eds
Glossary of Abbreviations
Policy-relevant Articles Given RSIS Award
RSIS Publications for the Year
External Publications for the Year
Media
Cohesive Societies
Sustainable Security
Other Resource Pages
News Releases
Speeches
Video/Audio Channel
External Podcasts
Events
Contact Us
S. Rajaratnam School of International Studies Think Tank and Graduate School Ponder The Improbable Since 1966
Nanyang Technological University Nanyang Technological University
  • About RSIS
      IntroductionBuilding the FoundationsWelcome MessageBoard of GovernorsHonours and Awards for RSIS Staff and StudentsRSIS Endowment FundEndowed ProfessorshipsCareer OpportunitiesGetting to RSIS
      Staff ProfilesExecutive Deputy Chairman’s OfficeDean’s OfficeManagementDistinguished FellowsFaculty and ResearchAssociate Research Fellows, Senior Analysts and Research AnalystsVisiting FellowsAdjunct FellowsAdministrative Staff
  • Research
      Research CentresCentre for Multilateralism Studies (CMS)Centre for Non-Traditional Security Studies (NTS Centre)Centre of Excellence for National SecurityInstitute of Defence and Strategic Studies (IDSS)International Centre for Political Violence and Terrorism Research (ICPVTR)
      Research ProgrammesNational Security Studies Programme (NSSP)Social Cohesion Research Programme (SCRP)Studies in Inter-Religious Relations in Plural Societies (SRP) Programme
      Other ResearchFuture Issues and Technology ClusterResearch@RSISScience and Technology Studies Programme (STSP) (2017-2020)
  • Graduate Education
      Graduate Programmes OfficeExchange Partners and ProgrammesHow to ApplyFinancial AssistanceMeet the Admissions Team: Information Sessions and other eventsRSIS Alumni
  • Outreach
      Global NetworksAbout Global NetworksRSIS Alumni
      Executive EducationAbout Executive EducationSRP Executive ProgrammeTerrorism Analyst Training Course (TATC)
      International ProgrammesAbout International ProgrammesAsia-Pacific Programme for Senior Military Officers (APPSMO)Asia-Pacific Programme for Senior National Security Officers (APPSNO)International Conference on Cohesive Societies (ICCS)International Strategy Forum-Asia (ISF-Asia)
  • Publications
      RSIS PublicationsAnnual ReviewsBooksBulletins and NewslettersRSIS Commentary SeriesCounter Terrorist Trends and AnalysesCommemorative / Event ReportsFuture IssuesIDSS PapersInterreligious RelationsMonographsNTS InsightPolicy ReportsWorking Papers
      External PublicationsAuthored BooksJournal ArticlesEdited BooksChapters in Edited BooksPolicy ReportsWorking PapersOp-Eds
      Glossary of AbbreviationsPolicy-relevant Articles Given RSIS AwardRSIS Publications for the YearExternal Publications for the Year
  • Media
      Cohesive SocietiesSustainable SecurityOther Resource PagesNews ReleasesSpeechesVideo/Audio ChannelExternal Podcasts
  • Events
  • Contact Us
    • Connect with Us

      rsis.ntu
      rsis_ntu
      rsisntu
      rsisvideocast
      school/rsis-ntu
      rsis.sg
      rsissg
      RSIS
      RSS
      Subscribe to RSIS Publications
      Subscribe to RSIS Events

      Getting to RSIS

      Nanyang Technological University
      Block S4, Level B3,
      50 Nanyang Avenue,
      Singapore 639798

      Click here for direction to RSIS

      Get in Touch

    Connect
    Search
    • RSIS
    • Publication
    • RSIS Publications
    • Will AI-Driven “Super-OODA Loops” Revolutionise Military Strategy and Operations?
    • Annual Reviews
    • Books
    • Bulletins and Newsletters
    • RSIS Commentary Series
    • Counter Terrorist Trends and Analyses
    • Commemorative / Event Reports
    • Future Issues
    • IDSS Papers
    • Interreligious Relations
    • Monographs
    • NTS Insight
    • Policy Reports
    • Working Papers

    CO25011 | Will AI-Driven “Super-OODA Loops” Revolutionise Military Strategy and Operations?
    Michael Raska

    20 January 2025

    download pdf

    SYNOPSIS

    AI-enabled “Super-OODA Loops” could revolutionise military operations by surpassing human cognitive limits with automated, near-instantaneous decision-making. These systems promise unmatched speed and precision but raise ethical challenges around accountability and misuse, requiring robust frameworks for responsible deployment.

    Image sourcing 16
    Source: Canva

    COMMENTARY

    The rise of AI-enabled systems has sparked debates about their potential to reshape decision-making in warfare. Critics argue that AI cannot reliably complement, much less replace, human understanding of the strategic environment or make sound predictions and judgments.

    The inherent uncertainty and complexity of warfare often referred to as its “fog and friction”, resist the logic of machine learning. Additionally, ethical, moral, and legal challenges highlight the critical need to maintain human control in AI-driven military operations.

    But what if AI can enhance the pace, accuracy, and adaptability of military operations in ways previously unimaginable? Could it transform the traditional “OODA Loop” cycles into an automated, high-speed decision-making process, enabling near-instant responses through machine-to-machine interfaces?

    Could AI, then, be the key to achieving decision dominance in future conflicts?

    Revisiting the OODA Loop Concept

    The OODA Loop, first developed by US Air Force Colonel John Boyd, was a simple but powerful idea. It outlined how individuals and organisations can outmanoeuvre opponents by cycling through four steps – Observe, Orient, Decide, Act – faster than their adversaries.

    The faster and more effectively one could cycle through these steps compared to an adversary, the greater their likelihood of success. Initially designed for air combat during the Cold War, the OODA Loop was meant to help fighter pilots survive and dominate in the chaotic, high-stakes environment of aerial dogfights.

    Boyd’s theory emphasised not only speed but adaptability: pilots who could rapidly process sensory information (Observe), interpret the situation in context (Orient), choose a course of action (Decide), and execute their plan (Act) could stay a step ahead of their opponents, forcing them into a reactive position and ultimately gaining the upper hand.

    Over time, the brilliance of this framework extended far beyond the cockpit. The principles of the OODA Loop proved applicable across a wide range of fields, from military strategy to business innovation, competitive sports, and even politics. In any context where the ability to make faster, better-informed decisions can create an advantage, the OODA Loop became a guiding philosophy.

    In business, for instance, companies use the OODA Loop to respond to market changes, outmanoeuvre competitors, and drive innovation. In sports, coaches and athletes rely on its principles to assess opponents’ strategies, adjust tactics in real time, and execute winning plays. Its adaptability and universal relevance have cemented the OODA Loop as a cornerstone of decision-making theory across disciplines, inspiring countless leaders and strategists to prioritise agility and responsiveness in the face of uncertainty.

    The Present: Technology-Augmented OODA Loops

    Yet, in today’s complexities of modern warfare, often characterised by information overload and military-technological advances, human decision-making often cannot keep up. Military commanders must often make split-second decisions to outmanoeuvre their adversaries in increasingly complex situations in saturated battlefields.

    As such, modern OODA loops already leverage advanced technologies to enhance decision-making at every phase:

    1. Observe. Networked sensors like drones, satellites, and radar systems provide real-time, multi-domain battlespace awareness.
    2. Orient. AI and machine learning synthesise massive data streams, identify patterns, and highlight threats, creating a unified operational picture.
    3. Decide. AI-driven decision-support tools analyse changing conditions, predict outcomes, and recommend actions, enabling rapid, informed decisions or autonomous responses in high-speed scenarios like missile defence.
    4. Act. High-speed networks and automated systems execute decisions swiftly, with minimal human input, such as deploying electronic countermeasures against threats.

    These advancements compress and accelerate the OODA process, offering unprecedented operational agility. In the war against Russia, for example, Ukraine has leveraged AI to level the playing field. Ukrainian forces have deployed drones equipped with AI systems that scan battlefields, identify threats, and coordinate strikes in real time.

    Where human operators might take hours to sift through drone footage, AI processes it in seconds, flagging targets and feeding actionable intelligence to commanders. This accelerated decision-making allows Ukrainian forces to act quickly, often neutralising threats before Russian forces can respond. The war in Ukraine illustrates how AI compresses decision cycles, transforming how militaries observe, orient, decide, and act.

    The Future: Autonomous “Super OODA Loops”

    In the next two decades, however, future “Super OODA” Loops will be increasingly AI-driven, highly autonomous, and operate on near-instantaneous timelines that far exceed human response speeds. We can project the following developments:

    1. Real-Time Data Integration. AI-enhanced surveillance and edge computing will process vast data streams, enabling predictive observations that anticipate adversarial actions.
    2. Adaptive Orientation. Machine learning will refine decision-making by analysing patterns, adapting to enemy tactics, and improving situational awareness dynamically.
    3. Autonomous Decisions and Actions. AI systems will handle critical, high-speed scenarios – such as intercepting hypersonic missiles – using pre-set protocols, bypassing traditional human oversight for machine-speed responses.

    As technology advances, Super OODA Loops will minimise human reliance in environments requiring split-second decisions beyond human cognitive limits. For example, future air power doctrines will incorporate squadrons of uncrewed combat autonomous vehicles (UCAVs) and other AI-enabled, self-directing platforms that can operate remotely across distributed battle networks, make collaborative decisions with human and machine partners, and manoeuvre faster than adversaries can detect, orient, or counter.

    The Ethical Challenges of Speed and Autonomy

    While the “Super OODA Loop” offers significant military advantages, it also brings with it a host of complex ethical dilemmas that demand careful consideration. As AI takes on greater responsibilities in military operations, one of the central questions is how much autonomy these systems should be granted.

    The risks of mistakes cannot be ignored.

    What happens if an AI system misidentifies a civilian target as a threat and initiates an action that leads to unintended casualties? Such errors, whether caused by faulty algorithms, incomplete data, or unforeseen scenarios, could not only result in tragic consequences but also undermine the legitimacy of military operations.

    Moreover, the battlefield of the future will likely see adversaries exploiting AI vulnerabilities. What if an enemy hacks or manipulates an AI system to disrupt its functionality or redirect its actions? For example, an adversary could deploy decoys or feed false data to confuse AI-driven systems, leading to miscalculations or misfires. This highlights the critical importance of building resilient, secure AI frameworks that can withstand such attacks.

    These ethical concerns go beyond immediate military considerations.

    The use of autonomous AI systems also raises questions about accountability. Who is responsible if an AI makes a decision that leads to unintended consequences? Is it the developer who created the algorithm, the military commander who deployed it, or the government that authorised its use?

    Establishing clear accountability structures is essential to maintain trust in these systems and ensure that their deployment aligns with international laws and norms. As militaries around the world integrate AI into their operations, these questions will need to be resolved not only at the national level but also through international cooperation.

    Without a shared understanding of ethical boundaries, the risks of miscalculation and escalation could rise, potentially destabilising the global security environment. Ultimately, while the “Super OODA Loop” offers unparalleled speed and precision, its true potential can only be realised if militaries approach its use with a commitment to responsibility, transparency, and safeguarding human values.

    About the Author

    Michael Raska is Assistant Professor in the RSIS Military Transformations Programme and a member of the Expert Advisory Group of the Global Commission for Responsible AI in Military Domain (GC-REAIM).

    Categories: RSIS Commentary Series / Country and Region Studies / East Asia and Asia Pacific / South Asia / Southeast Asia and ASEAN / Global
    comments powered by Disqus

    SYNOPSIS

    AI-enabled “Super-OODA Loops” could revolutionise military operations by surpassing human cognitive limits with automated, near-instantaneous decision-making. These systems promise unmatched speed and precision but raise ethical challenges around accountability and misuse, requiring robust frameworks for responsible deployment.

    Image sourcing 16
    Source: Canva

    COMMENTARY

    The rise of AI-enabled systems has sparked debates about their potential to reshape decision-making in warfare. Critics argue that AI cannot reliably complement, much less replace, human understanding of the strategic environment or make sound predictions and judgments.

    The inherent uncertainty and complexity of warfare often referred to as its “fog and friction”, resist the logic of machine learning. Additionally, ethical, moral, and legal challenges highlight the critical need to maintain human control in AI-driven military operations.

    But what if AI can enhance the pace, accuracy, and adaptability of military operations in ways previously unimaginable? Could it transform the traditional “OODA Loop” cycles into an automated, high-speed decision-making process, enabling near-instant responses through machine-to-machine interfaces?

    Could AI, then, be the key to achieving decision dominance in future conflicts?

    Revisiting the OODA Loop Concept

    The OODA Loop, first developed by US Air Force Colonel John Boyd, was a simple but powerful idea. It outlined how individuals and organisations can outmanoeuvre opponents by cycling through four steps – Observe, Orient, Decide, Act – faster than their adversaries.

    The faster and more effectively one could cycle through these steps compared to an adversary, the greater their likelihood of success. Initially designed for air combat during the Cold War, the OODA Loop was meant to help fighter pilots survive and dominate in the chaotic, high-stakes environment of aerial dogfights.

    Boyd’s theory emphasised not only speed but adaptability: pilots who could rapidly process sensory information (Observe), interpret the situation in context (Orient), choose a course of action (Decide), and execute their plan (Act) could stay a step ahead of their opponents, forcing them into a reactive position and ultimately gaining the upper hand.

    Over time, the brilliance of this framework extended far beyond the cockpit. The principles of the OODA Loop proved applicable across a wide range of fields, from military strategy to business innovation, competitive sports, and even politics. In any context where the ability to make faster, better-informed decisions can create an advantage, the OODA Loop became a guiding philosophy.

    In business, for instance, companies use the OODA Loop to respond to market changes, outmanoeuvre competitors, and drive innovation. In sports, coaches and athletes rely on its principles to assess opponents’ strategies, adjust tactics in real time, and execute winning plays. Its adaptability and universal relevance have cemented the OODA Loop as a cornerstone of decision-making theory across disciplines, inspiring countless leaders and strategists to prioritise agility and responsiveness in the face of uncertainty.

    The Present: Technology-Augmented OODA Loops

    Yet, in today’s complexities of modern warfare, often characterised by information overload and military-technological advances, human decision-making often cannot keep up. Military commanders must often make split-second decisions to outmanoeuvre their adversaries in increasingly complex situations in saturated battlefields.

    As such, modern OODA loops already leverage advanced technologies to enhance decision-making at every phase:

    1. Observe. Networked sensors like drones, satellites, and radar systems provide real-time, multi-domain battlespace awareness.
    2. Orient. AI and machine learning synthesise massive data streams, identify patterns, and highlight threats, creating a unified operational picture.
    3. Decide. AI-driven decision-support tools analyse changing conditions, predict outcomes, and recommend actions, enabling rapid, informed decisions or autonomous responses in high-speed scenarios like missile defence.
    4. Act. High-speed networks and automated systems execute decisions swiftly, with minimal human input, such as deploying electronic countermeasures against threats.

    These advancements compress and accelerate the OODA process, offering unprecedented operational agility. In the war against Russia, for example, Ukraine has leveraged AI to level the playing field. Ukrainian forces have deployed drones equipped with AI systems that scan battlefields, identify threats, and coordinate strikes in real time.

    Where human operators might take hours to sift through drone footage, AI processes it in seconds, flagging targets and feeding actionable intelligence to commanders. This accelerated decision-making allows Ukrainian forces to act quickly, often neutralising threats before Russian forces can respond. The war in Ukraine illustrates how AI compresses decision cycles, transforming how militaries observe, orient, decide, and act.

    The Future: Autonomous “Super OODA Loops”

    In the next two decades, however, future “Super OODA” Loops will be increasingly AI-driven, highly autonomous, and operate on near-instantaneous timelines that far exceed human response speeds. We can project the following developments:

    1. Real-Time Data Integration. AI-enhanced surveillance and edge computing will process vast data streams, enabling predictive observations that anticipate adversarial actions.
    2. Adaptive Orientation. Machine learning will refine decision-making by analysing patterns, adapting to enemy tactics, and improving situational awareness dynamically.
    3. Autonomous Decisions and Actions. AI systems will handle critical, high-speed scenarios – such as intercepting hypersonic missiles – using pre-set protocols, bypassing traditional human oversight for machine-speed responses.

    As technology advances, Super OODA Loops will minimise human reliance in environments requiring split-second decisions beyond human cognitive limits. For example, future air power doctrines will incorporate squadrons of uncrewed combat autonomous vehicles (UCAVs) and other AI-enabled, self-directing platforms that can operate remotely across distributed battle networks, make collaborative decisions with human and machine partners, and manoeuvre faster than adversaries can detect, orient, or counter.

    The Ethical Challenges of Speed and Autonomy

    While the “Super OODA Loop” offers significant military advantages, it also brings with it a host of complex ethical dilemmas that demand careful consideration. As AI takes on greater responsibilities in military operations, one of the central questions is how much autonomy these systems should be granted.

    The risks of mistakes cannot be ignored.

    What happens if an AI system misidentifies a civilian target as a threat and initiates an action that leads to unintended casualties? Such errors, whether caused by faulty algorithms, incomplete data, or unforeseen scenarios, could not only result in tragic consequences but also undermine the legitimacy of military operations.

    Moreover, the battlefield of the future will likely see adversaries exploiting AI vulnerabilities. What if an enemy hacks or manipulates an AI system to disrupt its functionality or redirect its actions? For example, an adversary could deploy decoys or feed false data to confuse AI-driven systems, leading to miscalculations or misfires. This highlights the critical importance of building resilient, secure AI frameworks that can withstand such attacks.

    These ethical concerns go beyond immediate military considerations.

    The use of autonomous AI systems also raises questions about accountability. Who is responsible if an AI makes a decision that leads to unintended consequences? Is it the developer who created the algorithm, the military commander who deployed it, or the government that authorised its use?

    Establishing clear accountability structures is essential to maintain trust in these systems and ensure that their deployment aligns with international laws and norms. As militaries around the world integrate AI into their operations, these questions will need to be resolved not only at the national level but also through international cooperation.

    Without a shared understanding of ethical boundaries, the risks of miscalculation and escalation could rise, potentially destabilising the global security environment. Ultimately, while the “Super OODA Loop” offers unparalleled speed and precision, its true potential can only be realised if militaries approach its use with a commitment to responsibility, transparency, and safeguarding human values.

    About the Author

    Michael Raska is Assistant Professor in the RSIS Military Transformations Programme and a member of the Expert Advisory Group of the Global Commission for Responsible AI in Military Domain (GC-REAIM).

    Categories: RSIS Commentary Series / Country and Region Studies

    Popular Links

    About RSISResearch ProgrammesGraduate EducationPublicationsEventsAdmissionsCareersVideo/Audio ChannelRSIS Intranet

    Connect with Us

    rsis.ntu
    rsis_ntu
    rsisntu
    rsisvideocast
    school/rsis-ntu
    rsis.sg
    rsissg
    RSIS
    RSS
    Subscribe to RSIS Publications
    Subscribe to RSIS Events

    Getting to RSIS

    Nanyang Technological University
    Block S4, Level B3,
    50 Nanyang Avenue,
    Singapore 639798

    Click here for direction to RSIS

    Get in Touch

      Copyright © S. Rajaratnam School of International Studies. All rights reserved.
      Privacy Statement / Terms of Use
      Help us improve

        Rate your experience with this website
        123456
        Not satisfiedVery satisfied
        What did you like?
        0/255 characters
        What can be improved?
        0/255 characters
        Your email
        Please enter a valid email.
        Thank you for your feedback.
        This site uses cookies to offer you a better browsing experience. By continuing, you are agreeing to the use of cookies on your device as described in our privacy policy. Learn more
        OK
        Latest Book
        more info