AI's role in the Israel-Hamas war so far
Artificial intelligence is changing the world, and our new video series GZERO AI explores what it all means for you—from disinformation and regulation to the economic and political impact. Co-hosted by Taylor Owen, professor at the Max Bell School of Public Policy at McGill University and director of its Centre for Media, Technology & Democracy, and by Marietje Schaake, International Policy Fellow, Stanford Human-Centered Artificial Intelligence and former European Parliamentarian, this weekly video series will help you keep up and make sense of the latest news on the AI revolution.
In the first episode of the series, Taylor Owen takes a look at how artificial intelligence is shaping the war between Israel and Hamas.
As the situation in the Middle East just continues to escalate, today we're asking how is artificial intelligence shaping the war between Israel and Hamas? The short answer is that not as many expected it might. I think there's two cautions about the power of AI here, and one place where AI has been shown to really matter. The first caution is on the value of predictive AI. For years, many have been arguing that AI might not just help us understand the world as it is but might actually be able to help us predict future events. Nowhere has this been more the case than in the worlds of national security and policing.
Now, Gaza happens to be one of the most surveyed regions in the world. The use of drones, facial recognition, border checkpoints, and phone tapping have allowed the Israeli government to collect vast amounts of data about the Gazan population. Add to this, the fact that the director of the Israeli Defense Ministry has said that Israel is about to become an AI superpower, and one would think that the government might have had the ability to predict such events. But on October 7th, this was notably not the case. The government, the military, and Israeli citizens themselves were taken by surprise by this attack.
The reality, of course, is however powerful the AI might be, it is only as good as the data that's fed into it, and often if the data is biased or just plain wrong, so will the predictive capacity. So I think we need to be really cautious, particularly about the sales pitches being made by the companies selling these predictive tools to our policing and our national security services. The certainty which with they're doing so, I think, needs to be questioned.
The second caution I would add is on the role that AI plays in the creation of misinformation. Don't get me wrong, there's been a ton of it in this conflict, but it hasn't really been the synthetic media or the deep fakes that many feared would be a big problem in events like this. Instead, the misinformation has been low tech. It's been photos and videos from other events taken out of context and displayed as if they were from this one. It's been cheap fakes, not deep fakes. Now, there have been some cases even where AI deepfake detection tools, which have been rolled out in response to the problem of deep fakes, have actually falsely identified AI images as being created by AI. In this case, the threat of deep fakes is causing more havoc than the deep fakes themselves.
Finally, though, I think there is a place where AI is causing real harm in this conflict, and that is on social media. Our Twitter and our Facebook and our TikTok feeds are being shaped by artificially intelligent algorithms. And too often than not, these algorithms reinforce our biases and fuel our collective anger. The world seen through content that only makes us angry is just fundamentally a distorted one. And more broadly, I think calls for reigning in social media, whether by the companies themselves or through regulation, are being replaced with opaque and ill-defined notions of AI governance. And don't get me wrong, AI policy is important, but it is the social media ecosystem that is still causing real harm. We can't take our eye off of that policy ball.
I'm Taylor Owen, and thanks for watching.