Info Pulse Now

HOMEcorporatetechentertainmentresearchmiscwellnessathletics

EU Offers Guidance on How AI Devs Can Obey Privacy Laws

By Fiona Jackson

EU Offers Guidance on How AI Devs Can Obey Privacy Laws

The opinion addresses how AI companies can be exempt from GDPR, when they can process personal data without the individuals' consent, and the consequences of developing a model with unlawfully processed data.

The European Data Protection Board has published an opinion addressing data protection in AI models. It covers assessing AI anonymity, the legal basis for processing data, and mitigation measures for impacts on data subjects for tech companies operating in the bloc.

It was published in response to a request from Ireland's Data Protection Commission, the lead supervisory authority under the GDPR for many multinationals.

The DPC sought more information about:

EDPB Chair Anu Talus said in a press release: "AI technologies may bring many opportunities and benefits to different industries and areas of life. We need to ensure these innovations are done ethically, safely, and in a way that benefits everyone.

"The EDPB wants to support responsible AI innovation by ensuring personal data are protected and in full respect of the General Data Protection Regulation."

An AI model can be considered anonymous if the chance that personal data used for training will be traced back to any individual -- either directly or indirectly, as through a prompt -- is deemed "insignificant." Anonymity is assessed by supervisory authorities on a "case-by-case" basis and "a thorough evaluation of the likelihood of identification" is required.

However, the opinion does provide a list of ways that model developers might demonstrate anonymity, including:

Kathryn Wynn, a data protection lawyer from Pinsent Masons, said that these requirements would make it difficult for AI companies to claim anonymity.

"The potential harm to the privacy of the person whose data is being used to train the AI model could, depending on the circumstances, be relatively minimal and may be further reduced through security and pseudonymisation measures," she said in a company article.

"However, the way in which the EDPB is interpreting the law would require organisations to meet burdensome, and in some cases impractical, compliance obligations around purpose limitation and transparency, in particular."

The EDPB opinion outlines that AI companies can process personal data without consent under the "legitimate interest" basis if they can demonstrate that their interest, such as improving models or services, outweigh the individual's rights and freedoms.

This is particularly important to tech firms, as seeking consent for the vast amounts of data used to train models is neither trivial nor economically viable. But to qualify, companies will need to pass these three tests:

Even if a company fails the balancing test, it may still not be required to gain the data subjects' consent if they apply mitigating measures to limit the processing's impact. Such measures include:

Technology lawyer Malcolm Dowden of Pinsent Masons said in the company article that the definition of "legitimate interest" has been contentious recently, particularly in the U.K.'s Data (Use and Access) Bill.

"Advocates of AI suggest that data processing in the AI context drives innovation and brings inherent social good and benefits that constitute a 'legitimate interest' for data protection law purposes," he said. "Opponents believe that view does not account for AI-related risks, such as to privacy, to discrimination or from the potential dissemination of 'deepfakes' or disinformation."

Advocates from the charity Privacy International have expressed concerns that AI models like OpenAI's GPT series might not be properly scrutinised under the three tests because they lack specific reasons for processing personal data.

If a model is developed by processing data in a way that violates GDPR, this will impact how the model will be allowed to operate. The relevant authority evaluates "the circumstances of each individual case" but provides examples of possible considerations:

The EDPB's guidance is crucial for tech firms. Although it holds no legal power, it influences how privacy laws are enforced in the EU.

Indeed, companies can be fined up to €20 million or 4% of their annual turnover -- whichever is larger -- for GDPR infringements. They might even be required to change how their AI models operate or delete them entirely.

SEE: EU's AI Act: Europe's New Rules for Artificial Intelligence

AI companies struggle to comply with GDPR due to the vast amounts of personal data needed to train models, often sourced from public databases. This creates challenges in ensuring lawful data processing and addressing data subject access requests, corrections, or erasures.

These challenges have manifested in numerous legal battles and fines. For instance:

Additionally, in September, the Dutch Data Protection Authority fined Clearview AI €30.5 million for unlawfully collecting facial images from the internet without user consent, violating GDPR. That same month, the Irish DPC requested the opinion be drawn up just after it successfully convinced Elon Musk's X to cease using European users' public posts to train its AI chatbot, Grok, without obtaining their consent.

Previous articleNext article

POPULAR CATEGORY

corporate

9808

tech

8831

entertainment

12396

research

5854

misc

13000

wellness

10208

athletics

13170