AI Kill-lists Drive Israeli Genocide in Gaza
Bappa Sinha
or reload the browser
AN investigation conducted by Israeli-Palestinian magazine “+972” has made explosive revelations about Israel’s use of artificial intelligence (AI)-generated “kill lists” to target supposed Palestinian militants in Gaza through bombing attacks that resulted in mass civilian casualties during its genocidal campaign in Gaza since 7 October, 2023. The report states that an AI computer programme called “Lavender” played a central role in the unprecedented bombing of Palestinians, especially during the early stages of the war. Lavender was developed by the Israeli Defence Force’s (IDF) elite Unit 8200. Lavender was used along with another AI system called “The Gospel.” While “The Gospel” identifies buildings and structures supposedly used by militants, Lavender identifies individuals and adds them to a kill list for targeted elimination.
THE MODUS OPERANDI
According to the report, during the first weeks of the war, the Israeli Army almost completely relied on Lavender to identify individuals and their homes for air strikes. Lavender marked as many as 37,000 Palestinians as “suspected militants” with links to Hamas. The IDF gave its officers a free hand to adopt Lavender’s kill list without any requirement for further human oversight or cross-check with ground intelligence regarding these individuals. Human oversight was limited to just 20 seconds per target only to verify that the target was male before authorising the bombing since female primary targets were picked by the AI programme were considered to be a mistake as Hamas does not recruit women in its military wing. This programme was used despite knowing that it often made mistakes and that it often picked targets who had no connection to militant groups.
The Israeli military systematically targeted individuals in their homes, typically at night when their entire families were present, rather than engaging them during military operations. This strategy was chosen because locating the individuals in their residence was easier. Additionally, automated systems, including one named “Where’s Daddy?” were employed specifically to track these targeted individuals and carry out bombings when they were inside their family homes. As a result, numerous Palestinians, predominantly women, children, the elderly and other non-combatants, were killed by Israeli airstrikes due to decisions made by the AI programme.
Information collected on most of the 2.3 million residents of Gaza through surveillance is fed into the Lavender system, which then analyzes and ranks the likelihood of each resident’s association with the military wing of Hamas. Lavender gives almost every single person in Gaza a rating from 1 to 100, as an indicator of how likely it is that they are Hamas militants. Lavender “learns” to identify characteristics of known Hamas militants, whose information had been fed into the machine as training data, and then tries to locate these same characteristics -- referred to as “features” in AI terminology -- among the general population. Typical “features” would include an individual’s visual information, mobile usage information, social media connections such as membership of WhatsApp groups, battlefield information, phone contacts, and photos. While humans select these features at first, the machine gradually comes to identify features on its own, thereby becoming completely opaque and unaccountable. An individual found to have several different incriminating features is given a high rating and thus automatically becomes a target for bombing. Rating thresholds above which individuals were chosen for bombing were decided arbitrarily, with thresholds lowered when Israeli officers ran out of targets to bomb. IDF officers knew well that the system would mistakenly flag individuals who had communication patterns similar to known Hamas militants, including police and civil defence workers, militants’ relatives, residents with names or nicknames identical to that of a militant, and people who used a phone that once belonged to a Hamas member which happens often enough, since devices of dead individuals get passed on to relatives in the warzone. Despite knowing that the system was in no way foolproof, the IDF relied on it extensively as a tool to generate targets for their genocide in an appalling reminder of Israeli callousness and cruelty.
MASSIVE CIVILIAN CASUALTIES
In an unprecedented move, the Israeli army decided that for every junior Hamas operative that Lavender marked, it was permissible to kill up to 15 or 20 civilians! In the past, the military had not authorized any “collateral damage” during assassinations of low-ranking militants. When it came to targeting such “junior militants” marked by Lavender, the army preferred to use only relatively inexpensive unguided missiles, commonly known as “dumb” bombs (in contrast to “smart” precision bombs), which can destroy entire buildings, wiping out many families as collateral damage. “You don’t want to waste expensive bombs on unimportant people -- it’s very expensive for the country and there’s a shortage [of those bombs]” -- went the Israeli logic.
In the event that the target was a senior Hamas official with the rank of battalion or brigade commander, the army authorized the killing of more than 100 civilians during the assassination of a single commander. Lavender and systems like “Where’s Daddy?” were thus combined with deadly effect, killing entire families and sometimes wiping out whole neighbourhoods. Most of the people killed were women and children. In order to assassinate Ayman Nofal, the commander of Hamas’ Central Gaza Brigade, the army authorized the killing of approximately 300 civilians, wiping out more than 15 houses in the bombing of the Al-Bureij refugee camp on Oct. 17, based on an imprecise pinpointing of Nofal’s location.
ROLE OF US BIG TECH
The tech infrastructure to run these deadly AI programmes may very well have come from US companies. In April 2021, the Israeli finance ministry announced the award of a contract for a $1.2 billion cloud computing system jointly built by Google and Amazon named “Project Nimbus.” The official statement said: “The project is intended to provide the government, the defence establishment and others with an all-encompassing cloud solution.”
A report published by the online publication “The Intercept”, mentioned that Google provided the Israeli government with the full suite of machine-learning and AI tools available through Google Cloud platform. Google documents indicated that the “Project Nimbus” cloud would give Israel capabilities for facial detection, automated image categorisation, object tracking, and even sentiment analysis that claims to assess the emotional content of pictures, speech, and writing. The tech community has discredited such dubious claims of discerning an individual’s emotions using facial expressions as pseudoscience. Many Google employees have become alarmed over the use of technologies such as AutoML, a Google AI tool offered to Israel through “Project Nimbus,” fearing both their inaccuracy and how they might be used for surveillance and military purposes. Google employees have been protesting their employer’s role and complicity in Israel’s genocide in Gaza for several months all across the US. “No Tech for Apartheid,” the organisers of the protest at Google offices, have said in a recent statement that Google has fired more than 50 employees for participating in these protests.
It is well established that US Big Tech monopolies such as Google and Amazon are entrenched in the US Military-Industrial-Surveillance complex and derive a significant share of their revenues from this association. The entire business model of companies such as Google and Facebook is based on pervasive surveillance of their users through their smartphones. The US intelligence agencies, such as the NSA, now have full access to this huge tranche of users' personal data. These same companies are now investing heavily in AI and see military applications, such as those mentioned in this article, as a huge source of revenue in the years ahead. Ethical concerns around the barbarity of using their dubious AI products for conducting genocide and furthering apartheid policies of the Israeli state are hardly going to be a roadblock for these companies doing business with the Israelis or any other US government-approved entities. These same companies have published and made loud pronouncements about “ethical AI charters” which are effectively a form of “ethics washing,” or essentially toothless self-regulatory pledges that provide only the appearance of scruples. The threat of AI is not from super-intelligent AI going rogue, as the “ethical AI charters” from these companies would have us believe, but from the business- as-usual activities of these tech monopolies and US-imperialism led countries whose behaviour can only be described as having “gone rogue” against any kind of world order.
or reload the browser