The Israeli military’s reported use of an untested and undisclosed artificial intelligence-powered database to identify targets for its bombing campaign in Gaza has alarmed human rights and technology experts who said it could amount to “war crimes”.
The Israeli-Palestinian publication +972 Magazine and Hebrew-language media outlet Local Call reported recently that the Israeli army was isolating and identifying thousands of Palestinians as potential bombing targets using an AI-assisted targeting system called Lavender.
“That database is responsible for drawing up kill lists of as many as 37,000 targets,” Al Jazeera’s Rory Challands, reporting from occupied East Jerusalem, said on Thursday.
The unnamed Israeli intelligence officials who spoke to the media outlets said Lavender had an error rate of about 10 percent. “But that didn’t stop the Israelis from using it to fast-track the identification of often low-level Hamas operatives in Gaza and bombing them,” Challands said.
It is becoming clear the Israeli army is “deploying untested AI systems … to help make decisions about the life and death of civilians”, Marc Owen Jones, an assistant professor in Middle East Studies and digital humanities at Hamid Bin Khalifa University, told Al Jazeera.
“Let’s be clear: This is an AI-assisted genocide, and going forward, there needs to be a call for a moratorium on the use of AI in the war,” he added.
The Israeli publications reported that this method led to many of the thousands of civilian deaths in Gaza.
On Thursday, Gaza’s Ministry of Health said at least 33,037 Palestinians have been killed and 75,668 wounded in Israeli attacks since October 7.
AI use ‘violates’ humanitarian law
“The humans that were interacting with the AI database were often just a rubber stamp. They would scrutinise this kill list for perhaps 20 seconds before deciding whether or not to give the go-ahead for an air strike,” Challands reported.
In response to widening criticism, the Israeli military said its analysts must conduct “independent examinations” to verify that the identified targets meet the relevant definitions in accordance with international law and additional restrictions stipulated by its forces.
It refuted the notion that the technology is a “system”, but instead “simply a database whose purpose is to cross-reference intelligence sources, in order to produce up-to-date layers of information on the military operatives of terrorist organisations”.
But the fact that there were “five to 10 acceptable civilian deaths” for every single Palestinian fighter who was an intended target shows why there are so many civilian deaths in Gaza, according to Challands.
Professor Toby Walsh, an AI expert at the University of New South Wales in Sydney, said legal scholars will likely argue that the use of AI targeting violates international humanitarian law.
“From a technical perspective, this latest news shows how hard it is to keep a human in the loop, providing meaningful oversight to AI systems that scale warfare terribly and tragically,” he told Al Jazeera.
‘War crimes’
The media outlets cited sources who said the Israeli army made decisions during the first weeks of the current conflict that “for every junior Hamas operative that Lavender marked, it was permissible to kill up to 15 or 20 civilians”.
The sources also said that if senior Hamas officials were the target then “the army on several occasions authorised the killing of more than 100 civilians in the assassination of a single commander”.
Ben Saul, the United Nations special rapporteur on human rights and counterterrorism, said if details in the report prove to be true, “many Israeli strikes in Gaza would constitute the war crimes of launching disproportionate attacks”.
If true, many Israeli strikes in Gaza would constitute the war crimes of launching disproportionate attacks: "the army decided during the first weeks of the war that, for every junior Hamas operative, it was permissible to kill up to 15 or 20 civilians" https://t.co/8JTgG4JVeQ
— Prof Ben Saul – UN SR Human Rights & Counterterror (@profbensaul) April 3, 2024
“Israel is currently trying to sell these tools to foreign entities, to governments that are looking to what Israel’s doing in Gaza, not with disgust, but actually with admiration,” said Antony Loewenstein, an Australian journalist and author of The Palestine Laboratory: How Israel Exports the Technology of Occupation Around the World.
“We’ll find out in the coming months and years who they may be … my sense is it’s gonna be countries that are currently saying they’re opposed to what Israel is doing.”