Last week, Prime Minister asked “how can we write laws [to regulate AI] that make sense for something we don’t yet fully understand?”. The PM does not appreciate that his Government has already drafted a law that applies to the processing of personal data for AI purposes but which has the objective of diminishing the protection afforded to data subjects.
In this blog, I show, in the context of scientific research, how the proposed DPDI No 2 Bill” (the “Bill” ) permits personal data to be used, disclosed or transferred outside the UK for AI training and development purposes, in secret, without consideration of compatibility or the right to object to the processing.
The legislative steps in the Bill that achieves these objectives are as follows.
- Define “scientific research” so it includes private AI development and AI algorithm training (i.e from now on, read “scientific research” to include private sector AI research, training and development).
- Define “appropriate safeguards” for AI “scientific research” that are below the equivalent safeguards specified in the GDPR.
- Ensure that AI “scientific research” can be undertaken in the “legitimate interest of the controller or of a third party” if the previously mentioned “appropriate safeguards” apply. Thanks to this Bill, the “legitimate interests” lawful basis is likely emerge as the preferred alternative to “data subject consent” for most “scientific research” because of the reduction in the level of data subject protection.
- Ensure that any further processing for AI “scientific research” purpose is always deemed to be compatible with the purpose of obtaining, irrespective of that obtaining purpose or of the nature of the AI research.
- Introduce a transparency exemption which applies to existing data subjects described in any dataset. This exemption applies to the further use of personal databases for AI “scientific research” purposes by the controller or by any third party to whom the personal data is disclosed.
It is the lack of transparency which diminishes data subject rights. For instance, how do data subjects know they can exercise their rights if they do not know their personal data are being processed in the first place?
However, before we get to explaining all the above, the blog has to deal with the definition of “scientific research” and then “appropriate safeguards”.
Scientific Research
The analysis begins with the definition of “scientific research”. This definition reads:
“Any research that can reasonably be described as scientific, whether publicly or privately funded….”. and “include processing for the purposes of technological development or demonstration, fundamental research or applied research, so far as those activities can reasonably be described as scientific…”.
Does this includes AI learning or AI development? Well to get that answer I asked CHAT_GBT the question. It replied:
“Yes, scientific research often involves AI and machine learning, which are important tools in various scientific disciplines. … Machine learning is a subset of AI that focuses on developing algorithms and models that can learn from data and make predictions or decisions based on that learning”.
So AI training and development can fall within “scientific research”. The next outstanding issue is whether the appropriate safeguards apply.
Appropriate safeguards
The “appropriate safeguards” (defined in A.84B; some detail in A.84C) can be summarised as follows. To be helpful, I have added some commentary.
Safeguard 1: The processing of personal data does not permit identification of a living individual apart from the collection of personal data but the safeguard does not apply if the further AI “scientific research” purpose cannot be fulfilled without the identification of a living individual.
Commentary on Safeguard 1: The inference underpinning this safeguard is that the identification details in the personal data should initially be separated from the personal data after collection.
This separation could allow that in any subsequent processing, the relevant (non-identifiable) data are treated as anonymous at the time of that processing (even though the controller knows that, at a later time, how to reconstitute the data as personal data). This increases the possibility that, at the time of that processing, there would be no DP obligations arising from the further processing for AI (thanks to the new narrow definition of “personal data” in the Bill).
In any event, if the controller considers that the AI “scientific research” purpose cannot be fulfilled without using personal data, then the safeguard does not apply.
In practice, therefore, Safeguard 1 might prove to be not much of a safeguard.
Safeguard 2: The processing is “likely to cause substantial damage or substantial distress to a data subject to whom the personal data relates”.
Commentary on Safeguard 2: “Gosh” and “Crikey”! Processing likely to cause moderate damage/distress to data subjects (i.e. anything short of substantial) is below this safeguard’s threshold. This is touted as safeguard? Surely not!
Safeguard 3: No decisions with respect to any particular data subject are made unless is it medical research approved by SoS, medical ethics committee etc.
Commentary on Safeguard 3: this safeguard existed in the DPA1998 so it is nothing new. However, for medical research it allows the SoS indirect powers to approve future AI “medical research” (e.g. via an ethics body made up of SoS appointees which could approve medical research for the NHS undertaken by USA high tec industries such as the Palantir). Although this is touted as a safeguard, in practice it could easily become a licence that allows the Government to exploit NHS personal data.
Safeguard 4:show respect for data minimisation Principle or pseudonymisation.
Commentary on Safeguard 4: it’s noteworthy that an obligation with to respect “anonymising personal data” is missing”. In addition, the safeguard is not new and replicates the controller’s obligation in Article 25 (data protection by design and by default).
Comparison with the safeguards specified in Recital 156 of the EU_GDPR shows the latter is far more protective of data subjects than this obligation (see below).
Recital 156: “…Those safeguards should ensure … that technical and organisational measures are in place to ensure, in particular, the principle of data minimisation. The further processing of personal data for ..[“scientific research” purposes] …. is to be carried out when the controller has assessed the feasibility to fulfil those purposes by processing data which do not permit or no longer permit the identification of data subjects, provided that appropriate safeguards exist (such as, for instance, pseudonymisation of the data…)”
Note that Recital 156 equates pseudonymisation with the safeguard; the Bill merely equates “respecting for pseudonymisation” as the equivalent safeguard.
Finally, the S.o.S has powers to vary the “appropriate safeguards” (Article 84D). It would be optimistic to say that these powers would be used to strengthen the data subject protection when the Bill’s direction of travel is clearly in the opposite direction.
Lawful basis for AI
Much processing for “scientific research” can continue to be undertaken using the lawful basis of fully informed and freely given “data subject consent”. If data subjects give consent, they are obviously made aware of the AI “scientific research” purpose and are protected in that they can withdraw consent at any time.
Although the Bill makes changes to “consent” to make it more flexible for the “scientific research” purposes, the fundamentals of consent have not been changed by the Bill (e.g. records of consent have to be kept, withdrawal of consent is part of the Privacy Notice).
If, however, the lawful basis for such AI research changes to “legitimate interests”, then the right to object to the processing applies. To exercise this right, data subjects have to present grounds relating to their individual situation and, in addition, data subjects have to be made aware that the right to object applies (e.g. in a Privacy Notice).
As will be seen, these last two requirements can be negated by the Bill.
As an aside, the ICO has already determined that the “legitimate interests” lawful basis can be approved for medical research by the ICO (see references). In fact, I expect this lawful basis to emerge as the preferred option for all “scientific research” (because of the diminution of data subject protection)
Legitimate interests in detail
Consider the “legitimate interests” lawful basis in detail; this reads:
“Processing is necessary for the purposes of the legitimate interests pursued by the controller or by a third party, except where such interests are overridden by the interests or fundamental rights and freedoms of the data subject which require protection of personal data.…” (my emphasis)
If the processing causes no impact on the data subject, then there is unlikely to be overriding interests of the data subject to protect. Indeed it will be difficult for the data subject to provide grounds that their interests or rights should prevail, if there is no impact of the processing on them
Hence I expect the right to object to be refused and the processing for AI “scientific research” to continue unabated.
Note that this “legitimate interest” basis also allows a controller to lawfully disclose personal datasets to a third party for its legitimate interest in using such personal data for its AI “scientific research” (see reference at end).
Secrecy and A.13
The Bill states that if the “appropriate safeguards” (see above) associated with the processing for an AI “scientific research” purpose apply, there are circumstances when there is no need to tell existing data subjects about the further processing for this AI “scientific research” purpose. (I will explain “existing” later).
In further detail, exemption states that (my emphasis):
“If the controller intends to further process personal data…
(i) for (and only for) the purposes of scientific or historical research, the purposes of archiving in the public interest or statistical purposes, and
(ii) in accordance with appropriate safeguards [see above] and
providing the transparency information is impossible or would involve disproportionate effort” where
“Disproportionate effort” depends on “among other things, the number of data subjects, the age of the personal data and any appropriate safeguards applied to the processing”.
Clearly if a new AI project requires access to a large volume of personal data and the processing falls within the AI “scientific research” definition and the appropriate safeguards are applied, then this exemption is very likely to be engaged. It follows that there is no need to inform data subjects described by such personal data about the further AI “scientific research” purpose.
It means any database can be further used for AI “scientific research” purpose without any transparency to data subjects described in the data. As stated previously, such secrecy also applies when there is a disclosure by a controller of these personal data to a third party in its legitimate interests for an AI “scientific research” purpose.
Note that the exemption does not apply to new data subjects who are added to a dataset after use or disclosure for the AI “scientific research” purpose. This is because the disproportionate effort definition is unlikely to apply as new data subjects can easily be informed by an updated Privacy Notice describing this purpose.
I know this exemption is complicated! This because it only applies to not informing data subjects already described in a dataset to a new further AI “scientific research” purpose associated with the processing of their personal data when the appropriate safeguards etc apply.
Compatibility, secrecy and A.14
Suppose a controller discloses a personal dataset, not for its own AI “scientific research” purpose, but for the AI “scientific research” purpose of a third party. If the disclosing controller does not need to identify the further purpose to existing data subjects, what about the third party? Does that third party need to contact the data subject?
The answer is “no”, as there is a similar provision introduced as new Article A.14(6) that states the third party does not need to inform the data subject of its AI “scientific research” purpose if “disproportionate effort” applies and the appropriate safeguards apply.
If there were to be onward disclosure by a third party to another third party for its AI “scientific research” purpose, then the same exemption applies assuming disproportionate effort and the appropriate safeguards apply.
To avoid the risk that a further AI “scientific research” purpose is incompatible with the purpose of obtaining, the Bill provides that any further use or disclosure for that purpose is deemed compatible, if disproportionate effort and the appropriate safeguards apply (see new Article 8A(3)(b)).
In other words, “deemed compatibility” can also apply to any further third party disclosure for AI “scientific research” purposes.
Conclusion
The blog describes that, under the DPDI No 2 Bill, a controller can disclose, to any number of other controllers, a huge dataset of personal data in circumstances when there is no obligation to inform data subjects described in that dataset about the use or disclosure for an AI “scientific research” purpose so long as the disproportionate effort and appropriate safeguards criteria apply.
The blog also describes that the “appropriate safeguards” are not as safe as one would expect.
Because the UK has a transfer agreement with the USA, such a dataset can be transferred to the USA for another third party’s AI “scientific research” purpose, away from the UK/EU jurisdiction. Because there is an exemption from transparency, this transfer is unknown to each data subject.
Readers are probably aware that Clearview, Facebook, Google, Twitter and Yahoo are all high tech USA companies that have been fined by European Data Protection Regulators and are all global players in AI “scientific research”.
Many will find it surprising that a large dataset of personal data can be lawfully transferred from the UK to the USA for any third party AI “scientific research” purpose without data subjects described in the dataset knowing about the circumstances of such a use or disclosure.
Given the above, I ask the reader a simple question: “What possibly could go wrong?”. Answers by email to the Prime Minister who needs enlightenment (Email form on https://contact.no10.gov.uk/Contact.aspx).
Forthcoming Data Protection Courses
Our well received, all-day Zoom workshop (10.00-4.30) on the Data Protection and Digital Information No 2 Bill will be held on Thursday 7 December 2023. The fee will be £250+VAT. Email [email protected] for workshop agenda or to reserve a place on this session.
The following BCS Practitioner or Foundation courses can be attended in person, or via Zoom, or as a mixture (i.e. part Zoom, part attendance just in case “stuff happens on the day”).
- Data Protection PRACTITIONER Course is in London on Monday, 20 November 2023 to Friday, 24 November 2023 (5 days: 9.30am to 5.30pm).
- Data Protection FOUNDATION Course is in London on (December 12-14, 2023: Tuesday to Thursday, 3 days: 9.45am to 5.00pm) or
Full details on Amberhawk’s website (www.amberhawk.com) or obtained by emailing [email protected].
References
Use of legitimate interests for lawful basis in medical research : see para 3.16 of the Sandbox on Our Future Health: https://ico.org.uk/media/for-organisations/documents/4026890/our-future-health-regulatory-sandbox-final-report.pdf