Paragon Fellowship

US State Department

Evaluating AI Weapons
Technology for Arms Transfers



Project Description

Emerging technologies are critical to national security and are becoming more advanced every day. As a general purpose technology, Artificial intelligence (AI) can accelerate innovation and adoption in other domains of critical and emerging technologies. For militaries worldwide, AI offers up new opportunities as well as challenges. To better understand the risks involved in developing and using AI in the military domain, the 2025 Spring Cohort Paragon team conducted an extensive and academic-style research project, providing the Office of Regional Security and Arms Transfers (RSAT), within the Bureau of Political-Military Affairs at the U.S. Department of State (DOS), with a landscape capture of current AI usage in the security domain as well as recommending ideas and methods for guiding future policies regarding how AI embedded in defense articles/services could be transferred. This project explored existing case studies of ‘weaponized’ AI worldwide, conducted stakeholder interviews, and developed a framework for the USG to better address the intersection of emerging technologies and arms transfers.

Project Aims

Our team set out to provide analysis of the current uses of weaponized AI and to create a framework, i.e. ‘Weaponized AI Transfer Scorecard’, to help evaluate arms transfers in the future. We accomplished these goals through two deliverables:

1. Weaponized AI Landscape Overview
a. Definition of Key Terms: Defined terms necessary to understand different forms of AI and how AI may be used in the context of national security and defense.
b. Discussion of Contemporary AI Weapon Landscape: Mapped key players and the latest developments in the field of AI in the military and national security domain.
c. Overview of Weaponized AI on the Global Stage: Explored how countries around the world have been leveraging weaponized AI, including a forecast of possible future use.

2. Weaponized AI Framework
a. Key Principles: Defined and explained key principles to help assess a AI technology transfer case examples.
b. Case Studies: Evaluated relevant case studies
c. Stakeholder Interviews: Spoke to key subject matter experts

Methodology

Our research began with a broad literature review, which fellows then narrowed to develop a comprehensive framework for evaluating the transfer of AI-enabled military capabilities. Landscape Analysis: We conducted a thorough mapping of the ‘weaponized’ AI ecosystem, examining historical patterns of technological revolution in warfare, defining key terminology, and identifying convergent challenges like algorithmic opacity and data governance issues. We also conducted a market analysis and survey of the current AI and defense tech ecosystem.

1. Principles Development: We established seven core principles (Fairness, Transparency, Accountability, Data Governance, Sustainability, Cybersecurity, and Non-Proliferation) based on ethical frameworks, legal considerations, and technical limitations identified in the landscape analysis. We then refined each principle with requirements that RSAT could use nominally as a Transfer Scorecard.

2. Case Study Application: We applied our scorecard to diverse strategic partnerships, selecting representative cases across NATO allies, ANZUS treaty members, mutual defense partners, and space-based technologies. For each case, we systematically evaluated compliance with our established principles, documented findings, and identified patterns to validate the scorecard's effectiveness as a policy tool for decision-makers evaluating potential weapons transfers.

Project Deliverables

Weaponized AI Landscape Overview

The initial landscape analysis revealed a rapidly evolving ecosystem in which traditional defense contractors and civilian technology companies increasingly converge. The report distinguished between Lethal Autonomous Weapons Systems (LAWS) that can select and engage targets without human intervention after activation, and Intelligence, Surveillance, Target Acquisition, and Reconnaissance (ISTAR) platforms that enhance battlefield awareness while maintaining human oversight. The analysis identified several convergent challenges affecting AI weapons development:
• Definitional ambiguity
• Ethical considerations (i.e. Human Responsibility and Accountability Measures)
• The "black box" problem of algorithmic opacity
• Data governance challenges in military environments
• Reliability issues, including structural bias integration and operational unpredictability

Next, a market analysis showed that U.S. and Chinese investments dominate globally, with the U.S. leading in private sector funding and China producing more research publications. Fellows found that traditional defense contractors like Lockheed Martin and Northrop Grumman heavily invest in AI integration, while technology companies that previously avoided military applications are now entering the field. These developments reshape tactical decision-making and strategic considerations across military domains—from autonomous ground vehicles and drone swarms to AI-enhanced naval vessels and cyber operations. Finally, Deliverable #1 (Interim Report) examined weaponized AI developments across multiple geographic regions, highlighting case studies from Iran, Israel, North and South Korea, China, and the Arctic.


Weaponized AI Transfer Scorecard

Fellows created a comprehensive framework for evaluating potential transfers against seven fundamental principles in an attempt to help policymakers balance strategic partnerships with national security considerations. The seven core principles are as follows:

1. Fairness
◦ Evaluation of algorithmic fairness and bias mitigation;
◦ Assessment of pre-existing, technical, and emergent biases;
◦ Measurement of equalized outcomes across different contexts.

2. Transparency
◦ Addressing the "black box" problem of algorithmic opacity through explainable AI;
◦ Ensuring openness throughout the learning and training process;
◦ Applying software engineering principles to development.

3. Accountability
◦ Creating systems of human expertise and accountabilityin commanding and ordering AI deployment;
◦ Defining and scoping human-in-the-loop validation approaches;
◦ Establishing clear responsibility for AI-driven decisions.

4. Data Governance
◦ Ensuring data traceability and documentation;
◦ Standardizing anonymization, processing, and transformation practices;
◦ Clarifying terms regarding data ownership and updates.

5. Sustainability
◦ Assessing full lifecycle impacts from production to decommissioning;
◦ Evaluating energy consumption and infrastructure demands;
◦ Managing long-term economic liabilities.

6. Cybersecurity
◦ Examining the recipient country's cyber defense framework;
◦ Evaluating system design for zero-trust architecture;
◦ Implementing security audits and anomaly detection.

7. Non-Proliferation
◦ Preventing horizontal and vertical proliferation of weaponized AI;
◦ Addressing military technologies being integrated into civilian public safety infrastructure;
◦ Implementing controls against reverse engineering.

Fellows applied this framework to diverse real-world scenarios, including NATO relationships (U.S. to Estonia to Ukraine, and U.S. to France), strategic partnerships (U.S. to Israel), intelligence-sharing alliances (U.S. to Australia and New Zealand), mutual defense treaties (U.S. to South Korea), and dual-use technologies (Starlink, Large Language Models). These case studies demonstrated the scorecard's suitability and value across different strategic contexts.


Project Impact and Future Work

The Weaponized AI Transfer Scorecard provides RSAT with a tool to make forward-leaning decisions about AI-technology-defense transfers that balance strategic partnerships with security considerations. By systematically evaluating each transfer against established principles, the scorecard helps identify potential risks while facilitating beneficial exchanges with allies. As AI technology evolves (which it will inevitably and at increasing acceleration), this framework offers a foundation that can help steer policy knowledge on other matters relating to science and technology.

Contributors

Jacqueline Malaret (Project Lead)

Florida International University

Audrey Berlie

Stetson University

Alexa Celis

NYU

Hyerin Cha

George Mason University

Arshi Mahajan

Dartmouth College

Cate Schunk

University of Dallas

Phia Teller

NYU