In March 2023, Italy became the first western country to block the advanced chatbot known as ChatGPT.
The Italian data protection authority, Garante, cited concerns over the protection of personal data when making this decision. It has given OpenAI, the California-based company that created ChatGPT, until the end of April to comply with its demands.
Garante said ChatGPT collects data in a way that’s incompatible with data protection law. Another reason given was the lack of age verification by the platform, which could expose children to harmful content. As a result, it used an emergency procedure to temporarily suspend the processing of personal data by OpenAI.
News about the temporary ban spread across the world, raising concerns about the consequences of decisions like this on the development of new artificial intelligence (AI) applications.
The move also coincided with a call by experts and business people – including OpenAI’s co-founder Elon Musk – to place limits on the development of AI-based applications until the risks could be better assessed.
The temporary ban could offer some important lessons about the proportionality and effectiveness of bans on developing technologies, about coordination between member states at the European level, and how to balance access to services with the need to protect children from accessing harmful content.
The order, issued on March 30, was signed by Pasquale Stanzione, the president of the Italian data protection authority. It followed a notification about a data breach concerning ChatGPT user data that had been reported ten days earlier.
Data processing
Garante briefly justified its measures by underlining the lack of information available to users, and data subjects, about the data processed by OpenAI. It also cited the large-scale processing of personal data to train generative systems such as ChatGPT.
OpenAI’s terms state that ChatGPT is provided only to users aged over 13. However, this did not satisfy Italy’s data protection authority, which was concerned about the lack of age verification.
The reaction by OpenAI was, first, to block access to ChatGPT in Italy and, second, to demonstrate its availability to collaborate with Garante on complying with the temporary order.

In its decision, Garante cited the large-scale processing of personal data to train generative systems such as ChatGPT. Shutterstock / Laymanzoom
Compliance would involve OpenAI implementing safeguards including the provision of a privacy policy, offering users the possibility of exercising individual rights over data protection, and providing information about the company’s legal basis for processing personal data.
Garante welcomed these commitments. It suspended the temporary order and requested that OpenAI implement these safeguards by the end of April 2023.
Harmonised framework
However, the case highlights at least three key lessons – namely, the lack of European coordination in regulating this technology, the effectiveness and proportionality of this measure, and the protection of children.
First, more European coordination is needed around the general issue of AI technology. The EU’s proposed AI Act is only one step towards a harmonised framework for ensuring the development of AI technologies that are aligned with European values. And as Italy’s ban has shown, the EU regulatory model can potentially become fragmented if national authorities go in their own directions.
In particular, the connection between AI and data protection empowers national authorities to react to the development of new AI technology. It also underlines the need for more coordination between European member states on regulation of all kinds.
Planning, not banning
Second, the measures adopted by the Italian data protection authority raise questions both about effectiveness and proportionality.
Regarding effectiveness, it’s worth noting that there were reports of a 400% surge in VPN downloads in Italy, potentially enabling users to get round the ban, following news of its introduction.
On the question of proportionality, a general ban does not seem to strike a balance between the conflicting constitutional interests at stake. The temporary measure does not mention how it takes into account the protection of other interests, such as the freedom of users to access ChatGPT.
Even though the ban is temporary, the situation might have benefited from the involvement at an earlier stage by other board members of the Italian data protection authority. A preliminary exchange with OpenAI could have avoided a ban altogether. This course of action could have anticipated the implementation of further safeguards to comply with data protection.
The best ways to protect children
Finally, the decision raises questions about the best ways to protect children from any harmful content created by these applications. Introducing an age verification system or alerts regarding harmful content could have been topics for discussion, had the parties been engaged in an ongoing dialogue.
This case offers an example of how general bans imposed on new technological applications are usually the result of quick reactions that do not involve a deep assessment of the effectiveness and proportionality of the measure.
Even if one argues that the decision tends towards protecting fundamental rights, primarily in data protection and safeguards for children, it leads to more uncertainty.
A preventative and collaborative approach with OpenAI would have minimised the risk of this service being blocked in Italy. Continued discussion between OpenAI and Italy’s authorities is critical.


OpenAI Pulls the Plug on Sora, Ending $1 Billion Disney Partnership
Federal Reserve Crisis: DOJ Standoff Threatens Powell's Succession and Rate Stability
Federal Judge Blocks Trump Administration's Move to End Temporary Protected Status for Somali Immigrants
Will a new border deal with the US open a backdoor into Kiwis’ personal data?
Palestinian Activist Leqaa Kordia Released from U.S. Immigration Detention After Judge's Order
CK Hutchison's Panama Ports Dispute Escalates as Arbitration Claims Surpass $2 Billion
Trump Administration Settles Lawsuit Barring Federal Agencies from Pressuring Social Media Censorship
Palantir's Maven AI Earns Pentagon "Program of Record" Status, Reshaping Military AI Strategy
Anthropic Sues Pentagon Over AI Blacklist, Citing Free Speech Violations
FEMA Reinstates $1 Billion Disaster Prevention Grant Program After Court Order
Federal Judge Blocks Pentagon's Blacklisting of AI Company Anthropic
Trump White House Unveils National AI Policy Framework for Congress
Jerome Powell May Stay on Fed Board Amid Criminal Investigation, Court Documents Reveal
Nintendo Switch 2 Production Cut as Holiday Sales Miss Targets
Makemation: a Nollywood movie that shows AI in action in Africa
Jeff Bezos Eyes $100 Billion Fund to Transform Manufacturing With AI 



