GZERO AI Video

Should we regulate generative AI with open or closed models?

Title Placeholder | GZERO AI

Marietje Schaake, International Policy Fellow, Stanford Human-Centered Artificial Intelligence, and former European Parliamentarian, co-hosts GZERO AI, our new weekly video series intended to help you keep up and make sense of the latest news on the AI revolution. Fresh from a workshop hosted by Princeton's Institute for Advanced Studies where the discussion was centered around whether regulating generative AI should be opened to the public or a select few, in this episode, she shares insights into the potential workings, effectiveness and drawbacks of each approach.

We just finished a half week workshop that dealt with the billion-dollar question of how to best regulate generative AI. And often this discussion tends to get quite tribal between those who say, “Well, open models are the best route to safety because they foster transparency and learning for a larger community, which also means scrutiny for things that might go wrong,” or those that say, “No, actually closed and proprietary models that can be scrutinized by a handful of companies that are able to produce them are safer because then malign actors may not get their hands on the most advanced technology.”

And one of the key takeaways that I have from this workshop, which was kindly hosted by Princeton's Institute for Advanced Studies, is actually that the question of open versus closed models, but also the question of whether or not to regulate is much more gradient. So, there is a big spectrum of considerations between models that are all the way open and what that means for safety and security,

Two models that are all the way closed and what that means for opportunities for oversight, as well as the whole discussion about whether or not to regulate and what good regulation looks like. So, one discussion that we had was, for example, how can we assess the most advanced or frontier models in a research phase with independent oversight, so government mandated, and then decide more deliberately when these new models are safe enough to be put out into the market or the wild.

So that there is actually much less of these cutting, cutting throat market dynamics that lead companies to just push out their latest models out of concern that their competitor might be faster and that there is oversight built in that really considers, first and foremost, what is important for society, for the most vulnerable, for anything from national security to election integrity, to, for example, nondiscrimination principles which are already under enormous pressure thanks to AI.

So, a lot of great takeaways to continue working on. We will hopefully publish something that I can share soon, but these were my takeaways from an intense two and a half days of AI discussions.

More For You

Microsoft unveiled a new set of commitments guiding its community‑first approach to AI infrastructure development. The strategy focuses on energy affordability, water efficiency, job creation, local investment, and AI‑driven skilling. As demand for digital infrastructure accelerates, the company is pushing a new model for responsible datacenter growth — one built on sustainability, economic mobility, and long‑term partnership with the communities that host it. The move signals how AI infrastructure is reshaping local economies and what people expect from the tech shaping their future. Read the full blog here.

Armed Israeli soldiers walk through an alley in the Old City of Hebron, in the occupied West Bank, on February 7, 2026. The Israeli army routinely secures routes and gathering points when settlers visit the city.
Photo by Mosab Shawer/Middle East Images/StringersHub/Sipa USA

The Israeli government unilaterally passed measures that allow Jewish settlers to purchase land in the West Bank, overriding past laws that effectively banned the sale of property there to anyone other than Palestinian residents.