Related Articles
On 15 May 2024, the UK Information Commissioner's Office ("ICO") announced its fourth call for evidence as part of its consultation series examining how data protection law applies to generative AI. Previous consultations in this series have focused on:
The focus of this fourth consultation is on how organisations developing, training and fine-tuning generative AI models can enable individuals to exercise their data protection rights.
As before, the ICO is seeking views from various people, ranging from developers and users of generative AI to legal advisors and consultants working in this area. The results of this and previous consultations will be used to shape the ICO's policy position on generative AI.
In this blog post, we explore the ICO's analysis and the input it is seeking.
Individual's rights Under data protection law, individuals have the following rights (restrictions and exemptions can sometimes apply) over their personal data when it is processed:
Across the AI lifecycle, organisations must have processes in place to enable and record the exercise of these rights.
The right to be informed is a prerequisite for exercising other rights under data protection law as individuals can only exercise them if they know if their information is being processed. The ICO stipulated that when generative AI developers collect personal data directly from individuals, for example to train or fine-tune a model, they must provide individuals with clear information about how this data is used and how they can exercise their rights according to Article 13 UK GDPR. Where the data is supplied by a third party, this obligation then lies with the third party. However, where developers collect personal data from other sources, for example through web scraping, they must still provide this information as set out in Article 14 UK GDPR. There are exceptions to this, as sometimes it is impossible or would require disproportionate effort to fulfil this obligation. Nonetheless, developers are still expected to take appropriate measures to protect individuals' rights and freedoms.
For example, developers must:
The ICO noted that it expects developers to have clear and easily accessible methods to help them respond to requests by individuals to access a copy of their data, regardless of the part of the AI lifecycle the data relates to. If developers claim they cannot fulfil these requests due to an inability to identify individuals, they must provide an explanation to the requestor, demonstrating why identification is not possible. The individual then has the option to provide additional information to aid in their identification, as outlined in Article 11(2) UK GDPR.
In its analysis, the ICO recognised that developers face challenges applying the right to erasure, restriction of and to object to processing due to memorisation issues inherent in AI models. AI models are prone to unintentionally outputting sections of the training data they have 'memorised' despite not being explicitly asked. To prevent this, developers often employ input and output filters. Input filters identify and modify user prompts, while output filters detect and adjust model outputs. Developers are also expected to consider the implications for the fairness and statistical accuracy of generative AI models where groups of individuals exercise these rights (such as a specific community).
Individuals' rights need to be respected throughout the AI lifecycle and supply chain, including during deployment. This also relates to the personal data that is submitted to the model once deployed and any outputs that can constitute personal data. Responsibility for fulfilling these rights lies with the organisations who are controllers or joint controllers at the various stages involved in the development and deployment of these generative AI models.
Based on its analysis, the ICO is seeking the following:
The ICO is interested in receiving evidence on tested, verifiable and effective methods that organisations are developing or using to meet their obligations in this area.
On 29 April 2024, NOYB (the organisation behind complaints against Meta's EU-US data transfers and more recently the "Consent or Pay" model) filed a complaint regarding OpenAI with the Austrian supervisory authority alleging that ChatGPT provides false information and OpenAI does not have any ability to correct it.
When asked about the complainant's (a public figure) birthday, ChatGPT repeatedly provided the wrong date of birth. However, OpenAI refused the complainant's request to correct or delete the data, claiming that it wasn't possible to do so. The only solution was for OpenAI to filter or block data on certain prompts (such as the name of the complainant), but it would be challenging to selectively filter out only specific details while retaining the rest of the context related to the complainant. OpenAI also allegedly failed to adequately respond to the complainant's access request.
Overall, it is clear that generative AI presents many challenges to ensuring individuals are informed about how their personal data is being processed and how to respond to data subject rights requests. It also highlights that data subject rights must not be an afterthought, but that rights mechanisms need to be considered early and built into the design of the AI system to respond to these requests.