In its report, the ICO said highlighted what it described as “misconceptions” in the context of gen-AI use and data protection compliance. One of those misconceptions concerns the type of data that compliance efforts are focused on.
The ICO said: “Many organisations focus their generative AI compliance efforts around PII (personally identifiable information). However, to ensure compliance in the UK they should be considering processing of any ‘personal data’ (which is a broader and legally defined concept in the UK). Organisations must not undertake compliance based on a fundamental misunderstanding or miscommunicate their processing operations.”
Wynn said that, in different ways, both the EPDB and ICO have flagged the challenges businesses face in demonstrating that AI models are anonymous.
“In EU and UK data protection law, the concept of a ‘motivated intruder’ is relevant to determining whether data input to, or output from, an AI model constitutes personal data. It requires developers to consider whether a motivated intruder – whether a malicious hacker or a normal user acting with benign intentions – would be able to make connections from the data that would enable them to identify individuals from that data, even if the data in isolation is not attributable to an individual.”
The EDPB opinion and ICO response address a wide range of other data protection issues in the AI context, including how businesses using AI can lawfully process personal data when doing so.
According to the ICO, based on current practices, businesses can only scrape personal data from the internet to use to train gen-AI models if they have a valid ‘legitimate interest’ in that activity.
The GDPR provides that if the processing is necessary for the purposes of the legitimate interests pursued by the controller or by a third party, personal data processing can be lawful. However, the ‘legitimate interests’ ground can only be relied upon for processing personal data if the interests cited by the controller are not “overridden by the interests or fundamental rights and freedoms of the data subject which require protection of personal data […]”.
The EDPB’s opinion focused, in part, on how businesses can demonstrate the appropriateness of legitimate interest as a legal basis in the AI development and deployment phases. In doing so, it did not expressly rule out the possibility that other lawful bases for processing personal data in an AI context might be relied upon by businesses – including in the context of web scraping.
Malcolm Dowden of Pinsent Masons said the question of what constitutes a ‘legitimate interest’ has been the subject of debate and litigation for years.
“The extent to which AI-related innovation can, in and of itself, be considered to be a valid legitimate interest that businesses can cite in the context of AI-related personal data processing, has been the subject of discussion by UK law makers recently as part of their scrutiny of the Data (Use and Access) Bill before the UK parliament,” Dowden said.
“Advocates of AI suggest that data processing in the AI context drives innovation and brings inherent social good and benefits that constitute a ‘legitimate interest’, for data protection law purposes. Opponents believe that view does not account for AI-related risks, such as to privacy, discrimination or from the potential dissemination of ‘deep fakes’ or disinformation,” he said.
“Where bodies like the ICO and EDPB land on this issue is important, because it has the potential to remove barriers some see to AI development in Europe – notwithstanding that processing based on any legitimate interest acknowledged in this regard would still need to be shown to be necessary and not overridden by the rights and freedoms of data subjects,” he said.