What's new

The US Army just conducted a massive test of its battlefield artificial intelligence

F-22Raptor

SENIOR MEMBER
Jun 19, 2014
6,539
2
9,171
Country
United States
Location
United States
YUMA PROVING GROUND — After weeks of work in the oppressive Arizona desert heat, on Sept. 23 the U.S. Army carried out a series of live fire engagements at Yuma Proving Ground to show how artificial intelligence systems can work together to automatically detect threats, deliver targeting data and recommend weapons responses at blazing speeds.

Set in the year 2035, the engagements were the culmination of Project Convergence 2020, the first in a series of annual demonstrations utilizing next generation AI, network and software capabilities to show how the Army wants to fight in the future.


The Army was able to use a chain of artificial intelligence, software platforms and autonomous systems to take sensor data from all domains, transform it into targeting information, and select the best weapon system to respond to any given threat in just seconds. Army officials claimed that these AI and autonomous capabilities have shorted the sensor to shooter timeline—the time it takes from when sensor data is collected to when a weapon system is ordered to engaged—from 20 minutes to 20 seconds, depending on the quality of the network and the number of hops between where it’s collected and its destination.

“We use artificial intelligence and machine learning in several ways out here,” said Brigadier General Ross Coffman, director of the Army Futures Command’s Next Generation Combat Vehicle Cross-Functional Team. “We used artificial intelligence to autonomously conduct ground reconnaissance, employ sensors and then passed that information back. We used artificial intelligence and aided target recognition and machine learning to train algorithms on identification of various types of enemy forces. So, it was prevalent throughout the last six weeks.”

The first engagement of the day is informative of how the Army stacked together AI capabilities to automate the sensor to shooter pipeline. In that example, the Army used space-based sensors operating in low Earth orbit to take images of the battleground. Those images were downlinked to a TITAN ground station surrogate located at Joint Base Lewis McCord in Washington, where they were processed and fused by Prometheus.


Currently under development, Prometheus is an AI system that takes the sensor data ingested by TITAN, fuses it, and identifies targets. The Army received its first Prometheus capability in 2019, although it’s targeting accuracy is still improving according to one Army official at Project Convergence. In some engagements, operators were able to send in a drone to confirm potential threats identified by Prometheus.

From there, the targeting data was delivered to a Tactical Assault Kit—a software program that gives operators an overhead view of the battlefield populated with both blue and red forces. As new threats are identified by Prometheus or other systems, that data is automatically entered into the program to show users their location. Specific images and live feeds can be pulled up in the environment as needed.


All of that takes place in just seconds.

Now that the Army has its target, it needs to determine the best response. Enter the real star of the show: the FIRES Synchronization to Optimize Responses in Multi-Domain Operations, or FIRESTORM.


“What is FIRESTORM? Simply put it’s a computer brain that recommends the best shooter, updates the common operating picture with the current enemy situation, and friendly situation, admissions the effectors that we want to eradicate the enemy on the battlefield,” said Coffman.

Army leaders were effusive in praising FIRESTORM throughout Project Convergence. The AI system works within the Tactical Assault Kit. Once new threats are entered into the program, FIRESTORM processes the terrain, available weapons, proximity, number of other threats and more to determine what the best firing system to respond to that given threat. Operators can assess and follow through with the system’s recommendations with just a few clicks of the mouse, sending orders to soldiers or weapons systems in just seconds of identifying a threat.

Just as important, FIRESTORM provides critical target deconfliction, ensuring that multiple weapons systems aren’t redundantly firing on the same threat. Right now, that sort of deconfliction would have to take place over a phone call between operators. FIRESTORM speeds up that process and eliminates any potential misunderstandings.

In that first engagement, FIRESTORM recommended the use of an Extended-Range Cannon Artillery. Operators approved the algorithm’s choice, and promptly the cannon fired a projectile at the target located 40 kilometers away. The process from identifying the target to sending those orders happened faster than it took the projectile to reach the target.


Perhaps most surprising is how quickly FIRESTORM was integrated into Project Convergence.

“This computer program has been worked on in New Jersey for a couple years. It’s not a program of record. This is something that they brought to my attention in July of last year, but it needed a little bit of work. So we put effort, we put scientists and we put some money against it,” said Coffman. “The way we used it is as enemy targets were identified on the battlefield—FIRESTORM quickly paired those targets with the best shooter in position to put effects on it. This is happening faster than any human could execute. It is absolutely an amazing technology.”

Prometheus and FIRESTORM weren’t the only AI capabilities on display at Project Convergence. In other scenarios, a Grey Eagle drone was able to identify and target a threat using the on board Dead Center payload. With Dead Center, the Grey Eagle was able to process the sensor data it was collecting, identifying a threat on its own without having to send the raw data back to a command post for processing and target identification. The drone was also equipped with the Maven Smart System and Algorithmic Inference Platform, a product created by Project Maven, a major Department of Defense effort to use AI for processing full motion video. According to one Army officer, the capabilities of the Maven Smart System and Dead Center, but placing both on the modified Grey Eagle at Project Convergence would help them to see how they compared.

With all of the AI engagements, the Army ensured there was a human in the loop to provide oversight of the algorithms' recommendations. When asked how the Army was implementing the Department of Defense’s principles of ethical AI use adopted earlier this year, Coffman pointed to the human barrier between AI systems and lethal decisions.

“So obviously the technology exists, to remove the human right the technology exists, but the United States Army, an ethical based organization—that’s not going to remove a human from the loop to make decisions of life or death on the battlefield, right? We understand that,” explained Coffman. “The artificial intelligence identified geo-located enemy targets. A human then said, Yes, we want to shoot at that target.”

https://www.c4isrnet.com/artificial...efield-artificial-intelligence-in-the-desert/
 

F-22Raptor

SENIOR MEMBER
Jun 19, 2014
6,539
2
9,171
Country
United States
Location
United States
YUMA PROVING GROUND, AZ -- Army Secretary Ryan McCarthy said integrated tactical network sets being developed by the Network Cross-Functional Team will vastly improve the service's network capability and resilience, following a Project Convergence demonstration here.

The Army is working on developing a resilient network that will be essential to everything the service is doing, McCarthy and Army Futures Command chief Gen. Mike Murray told reporters this week. During the Project Convergence demonstrations, the service learned it needs to work on scaling the network for both ground-to-ground and air-to-ground operations.

"The network, in particular, there's capability drops in fiscal [years] [2021] and [2023] that are going to be the upgrades that vastly improve our network capability and resilience," McCarthy told reporters yesterday. The service "learned about a lot of the things that they need to focus on . . . I looked at yesterday as a good thing."

The Network CFT in June announced it had completed a final design review for CS21 and was ready to start fielding equipment to four infantry brigades. CS21 equipment includes single-channel radios, tactical assault kits, tethered drones and more capabilities.

During the Project Convergence demonstration in Yuma, the Network CFT leveraged currently fielded radio and networking capabilities, which included a mix of artificial intelligence, machine learning and autonomous technologies, and demonstrated technologies under consideration for inclusion in CS23. Those are low Earth orbit and medium Earth orbit satellite communications links and Tactical Data Fabric efforts.

For example, CS21 program of record and commercial-off-the-shelf radios were used to provide the service with a mesh network that enabled information exchanges between ground and aerial weapon platforms, sensors and decision-making agents on the ground. Data was passed to the command post through a mid-tier link during the demonstrations.

McCarthy told reporters the Army during the experiment learned what can be improved with the capability sets.

The service knows what it needs "to work on or to fix for Capability [Set] 21 and 23," he said, adding: "I was encouraged."

https://insidedefense.com/insider/mccarthy-capability-drops-will-vastly-improve-network-resilience
 

Users Who Are Viewing This Thread (Total: 1, Members: 0, Guests: 1)

Top