If you’ve been wondering when our military would start deploying artificial intelligence on the battlefield, the answer is that it appears to be already happening. The Pentagon has been using computer vision algorithms to help identify targets for airstrikes, according to a report from Bloomberg News. Just recently, those algorithms were used to help carry out more than 85 airstrikes as part of a mission in the Middle East.
The bombings in question, which took place in various parts of Iraq and Syria on Feb. 2, “destroyed or damaged rockets, missiles, drone storage and militia operations centers among other targets,” Bloomberg reports. They were part of an organized response by the Biden administration to the January drone attack in Jordan that killed three U.S. service members. The government has blamed Iranian-backed operatives for the attack.
Bloomberg quotes Schuyler Moore, chief technology officer for U.S. Central Command, who said, of the Pentagon’s AI use: “We’ve been using computer vision to identify where there might be threats.” Computer vision, as a field, revolves around training algorithms to identify particular objects visually. The algorithms that helped with the government’s recent bombings were developed as part of Project Maven, a program designed to spur new greater use of automation throughout the DoD. Maven was originally launched in 2017.
The use of AI for “target” acquisition would appear to be part of a growing trend—and an arguably disturbing one, at that. Late last year it was reported that Israel was using AI software to decide where to drop bombs in Gaza. Israel’s program, which it has dubbed “The Gospel,” is designed to compile vast amounts of data and then recommend a target to a human analyst. That “target” could be a weapon, a vehicle, or a live human. The program can suggest as many as 200 targets in a period of just 10-12 days, Israeli officials have said. Israelis also say that the targeting is just the first part of a broader review process, which can involve human analysts.
If using software to identify bombing targets would seem like a process prone to horrible mistakes, it’s worth pointing out that, historically speaking, the U.S. hasn’t been particularly good at figuring out where to drop bombs, anyway. If you’ll recall, previous administrations have been prone to confuse wedding parties with terrorist HQs. You might as well let a software program make those kind of calls, right? It seems like there could be less human culpability that way. If the wrong person gets killed or you bomb a medicine factory instead of a munitions stockpile, you can always just blame it on a software glitch.
+ There are no comments
Add yours