Foreign Affairs Forum

View Original

AI-powered systems used by Israel in its military operations in Gaza and Lebanon

Israel has been using several AI-powered systems in its military operations in Gaza and Lebanon, raising significant concerns about civilian casualties and potential violations of international law. Here are the key incidents and systems involved:

AI Targeting Systems

“Lavender” Systems

“Lavender” is an AI-based program used by the Israeli army to generate targets for assassination. According to intelligence officers interviewed by +972 Magazine and Local Call:

• The system has marked approximately 37,000 Palestinians as suspected “Hamas militants” for potential assassination.

• It played a central role in the bombing of Palestinians, especially in the early stages of the war.

• The system’s outputs were treated “as if it were a human decision”.

• Thousands of Palestinians, mostly women and children not involved in fighting, were killed due to the AI program’s decisions.

“The Gospel” System

“The Gospel” is another AI system used to mark buildings and structures that the army claims militants operate from.

“Where’s Daddy?” System

This system is designed to track targeted individuals and carry out bombings when they enter their family residences. It was used specifically to attack targets in their homes, often at night when families were present.

Consequences and Concerns

1. Indiscriminate Targeting: The AI systems have led to attacks on family homes, resulting in high civilian casualties.

2. Lack of Human Oversight: The reliance on AI has reduced human involvement in target selection, potentially increasing the risk of errors.

3. Data and Privacy Issues: These tools rely on ongoing surveillance of Palestinian residents of Gaza, raising concerns about privacy rights violations.

4. Potential War Crimes: Human rights organizations have warned that the use of these AI tools may lead to violations of international humanitarian law, particularly regarding distinction between military targets and civilians.

5. Scale of Destruction: The use of AI has enabled a massive increase in the number of daily targets, with over 1,000 targets struck daily in some instances.

Broader Implications

• Israel has reportedly been using Gaza as a testing ground for new military technologies, including AI systems.

• The use of these AI systems in warfare sets a concerning precedent for future conflicts globally.

• There are calls for increased regulation and accountability regarding the use of AI in military operations.

Conclusion

This situation highlights the urgent need for international discussions and regulations on the use of AI in Warfare Israel has implemented several AI-driven systems in its military operations in Gaza and Lebanon, prompting serious concerns regarding civilian casualties and potential infringements of international law.