DARPA Aims to Develop AI, Autonomy Applications Warfighters
Stay informed with our
free newsletters

This news is classified in: Defense Cyber Defense / IT Soldier

Mar 27, 2024

DARPA Aims to Develop AI, Autonomy Applications Warfighters Can Trust

An important goal of the Defense Advanced Research Projects Agency is developing artificial intelligence that is trustworthy for the Defense Department — particularly for making life-or-death recommendations to warfighters, said Matt Turek, deputy director of DARPA's Information Innovation Office.

AI, machine learning and autonomy are being used by about 70% of DARPA's programs in some form or another, Turek said today at a Center for Strategic and International Studies event.

Another reason AI development is such a priority is to prevent an unexpected breakthrough in technology, or "strategic surprise," by adversaries who might also be developing advanced capabilities, he said, adding that DARPA also aims to create its own strategic surprise.

Military Robots and Autonomous Systems - Market and Technology Forecast to 2030

Military Robots and Autonomous Systems - Market and Technology Forecast to 2030

Market forecasts by Region, by Technology, Application, Operation, Platforms, and by End-User. Technology and Market Overview, Events based Forecast, Opportunity Analysis, and Leading Companies Profiles

Download free sample pages More information

To accomplish those goals, DARPA is looking for transformative capabilities and ideas from industry and academia, Turek said.

One of the many ways the agency gets these capabilities and ideas is to hold various types of challenges where teams from the private sector can win prizes worth millions of dollars, he said.

An example of that, he said, is DARPA's Artificial Intelligence Cyber Challenge, which uses generative AI technologies — like large language models — to automatically find and fix vulnerabilities in open-source software, particularly software that underlies critical infrastructure.

Large language models involve processing and manipulating human language to perform such tasks as secure computer coding, decision-making, speech recognition and making predictions.

Turek said a unique feature of this challenge is the partnership between DARPA and state-of-the-art large language model providers that are participating in the challenges, including Google, Microsoft, OpenAI and Anthropic.

Most likely, large language model improvements will also benefit the commercial sector, as well as DOD, Turek said.

An example of the use of autonomy and of AI that DARPA has been testing with the Air Force involves its F-16 fighter jets, he said.

Turek said DARPA has four areas of AI research involving industry and academia partners:

  • Proficient artificial intelligence;
  • Confidence in the information domain, which includes tools that detect things like manipulated media;
  • Secure and resilient systems; and
  • Defensive and offensive cyber tools.

Turek noted that there's a lot of synergy across those four areas.

US Department of Defense
View original News release

Digital Forensics for National Security Symposium

Digital Forensics for National Security Symposium

National Harbor, MD
May 15 - 16, 2024

View agenda
Urban Operations Summit

Urban Operations Summit

National Harbor, MD
Jul 10 - 11, 2024

View agenda