SlideShare a Scribd company logo
How to regulate
foundation models:
can we do better
than the EU AI Act?
Lilian Edwards
Professor of Law,
Newcastle University
@lilianedwards
Lilian.edwards@ncl.ac.uk
April 2023
What are large or “foundation” models?
• GPT-2/3/3.5/4 (Open
AI/Microsoft)(prompt to text)(2019
on)
• “Large Language Model” or LLM
• ChatGPT
• DALL-E 2 (text to images – Google)
• Stable Diffusion (open source – text
to image)
• HarmonAI – makes AI generated
music (Stability)
• CoPilot (prompt generates computer
code – GitHub/OpenAI)
• Meta Make-me-A-Video (text to
video - Meta)
• ERNIE ( Baidu, China) (prompt to
text)
eg Stable Diffusion : img to img (open source third party code)
nurse
Stable Diffusion
doctor
DALL-E 2
ChatGPT
December
2022
How to regulate foundation models: can we do better than the EU AI Act?
How to regulate foundation models: can we do better than the EU AI Act?
Ecology of downstream deployers
Integration into search, Feb 2023
New York Times
How to regulate foundation models: can we do better than the EU AI Act?
Important (for law) features of large or
“foundation” models
• Generative – create text, images etc rather than merely
classifying or predicting (ML)
• Trained on unprecedentedly large datasets
• Often scraped from “public” Internet
• Impossible to manually review legality, privacy or harm
of every item in datasets
• Computationally expensive and retraining slow ->
• large tech co dominance
• GPT-4 training cost >$100mn
• environmentally worrying
• Training sets allow the model to assess probability of
next word, pixel etc – not direct copying
• Models are general , can have multiple uses, eg to write
a party invite, a racist attack or provide customer
support within an automated hiring system
• Generated content increasingly difficult to distinguish
from human-created content (disinfo, deepfakes)
• Outputs may be “hallucinations” Hoppner, 2023
Issues with large models
PHASE 1 STOCHASTIC PARROTS
• Don’t actually understand , just “parrot”
• Bias, discrimination, misrepresentation and stereotyping of
groups; hate speech
PHASE 2 WILL NO-ONE THINK OF THE ARTISTS?
• Image and video deepfakes
• Pastiche
• Copyright
PHASE 3 FAKE NEWS ON STEROIDS
• Fake news and “hallucination” (text + images]
• Education & plagiarism
• Digital Services Act
PHASE 4 – YOU HAVE ZERO PRIVACY, GET OVER IT
• GDPR
Solution 1 : the EU AI Act and “GPAI”
AIA “risk based” approach
• Unacceptable risk – ‘Complete’
prohibition, 4 examples – Article 5
• High-risk –Fixed categories of
risky domains, based on intended
use ; “essential requirements”
including dataset quality, human
oversight –
• Limited risk – Transparency
obligations for a few AI systems
(chatbots, deepfakes, emotion ID,
biometric categorisation) – Article
52
• Minimal risk – Codes of conduct –
Article 69
Photo Source: European Commission, Digital Strategy Website
https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai
Annex II - Products
• Machinery
• Toys
• Recreational craft and watercraft
• Lifts
• Equipment and protective systems
intended for use in potentially
explosive atmospheres
• Radio equipment
• Pressure equipment
• Cableway installations
• PPE
• Medical devices
• [...]
High-Risk AI systems (Designation) - Annexes II&III
• Annex III - Services
• Biometric identification and
categorisation of natural persons;
• Management and operation of critical
infrastructure;
• Education and vocational training;
• Employment, workers management and
access to self-employment;
• Access to and enjoyment of essential
private services and public services and
benefit;
• Law enforcement;
• Migration, asylum and border control
management;
• Administration of justice and democratic
processes
• Compliance with requirements (Art.8);
• Risk management system (Art.9);
• Data and data governance (Art.10); (“data quality”)
• “Training, validation and testing data sets shall be relevant, representative, free of errors and complete. They shall have the appropriate
statistical properties, including, where applicable, as regards the persons or groups of persons on which the high-risk AI system is
intended to be used.”
• Technical documentation (Art.11);
• Record-keeping (Art.12);
• Transparency and provision of information to users (Art.13);
• Human oversight (Art.14);
• “Human oversight shall aim at preventing or minimising the risks to health, safety or fundamental rights that may emerge when a high-
risk AI system is used in accordance with its intended purpose or under conditions of reasonably foreseeable misuse”
• Accuracy, robustness and cybersecurity (Art.15).
High-Risk AI systems (Requirements) – Arts. 8 - 15
How to regulate foundation models: can we do better than the EU AI Act?
Definition of GPAI in EU AIA
LGAIMs = Large General AI Models – Hacker, Engel and Mauer “Regulating ChatGPT and other Large generative AI Models”, February
2023
Issues : over-inclusive (no emphasis generality in “ability, task or output”); important for classification ->
Council position
‘general purpose AI system’ means an AI system that is trained on broad data at scale, is
designed for generality of output, and can be adapted to a wide range of tasks
EU Parliament, March 2023
Developers v deployers
• Developers?
• Akin to manufacturers – control
and have knowledge of the
training sets, weights, algorithms,
content moderation etc (esp if
closed source)
• Most high-risk obligations arise at
development stage (training &
human fine-tuning stages)
• Have practical power and
economic gains
• But can they handle
unforeseeable uses/risks? Or can
only tech giants  competition
issues? Open source providers?
• Deployers? [NB “users” in AIA!]
• Originally only duties on deployer
if they make a “substantial
modification” to the AI system, ie
become new provider
• Practicality ?: May be impossible
for them to fix or even audit issues
of data quality etc without access
to upstream source code, training
datasets etc (often
secret/proprietary – eg as with
GPT-3)
• “AI as a service” API/cloud model
will be prevalent till models
smaller
“Pick n Mix?”
Solutions in AIA process?
• 13 May 2022, French presidency added amendment excluding GPAIs from AIA
• Back and forth..
• Council position, Dec 2022, arts 4a-4c
• GPAI deemed high risk if they “may be used as high risk AI systems or as
components of high risk AI systems”
• Unless “explicitly excluded all high-risk uses”, but not if not in “good faith”
• European Parliament
• All generative AI models to go into high risk if
• they generate text that might be mistaken for human
• And all deepfakes and AV content showing something that “never
happened” unless an “obvious artistic work”
• & Providers owe cooperation & transparency obligations to downstream
users
• Commission to tweak the high-risk obligations by delegated Acts..
Foundation models v GPAI?? 19/4/23
“Foundation models”
• “an AI system model that is trained on broad data
at scale, is designed for generality of output, and
can be adapted to a wide range of distinctive
tasks.”
• Eg Chat GPT , Stable Diffusion
• “Trained on data scraped from entire Internet”?
Not just if labelled
• Stricter obligations - High risk ++?
• Adds sustainability obligations + independent
expert ex ante oversight
• Documented analysis and testing throughout
lifecycle
• Disclosures re copyright in training set; filters to
avoid delivery unlawful content
“GPAI”
• “AI system that can be used in and adapted to
a wide range of applications for which it was
not intentionally and specifically designed.”
• ?? E.g. “unlabelled data that need further
training by the provider, such as algorithms
developed to recognise skin cancer”
• Laxer regime, since obligations fall on high risk
providers who build on the data (?)
Deployers & value chain
• “non-binding standard contractual clauses that
regulate rights and obligations consistent with each
party’s level of control”
• Discrimination/ equality law
• Liability (product liability for AI)
• Copyright
• Content, hate speech, libel, fake
news (DSA)
• Privacy & data protection
• Personality & image rights
• Competition law
• See Hoppner
https://papers.ssrn.com/sol3/paper
s.cfm?abstract_id=4371681
• Apart from last though, only AIA
and DMA allows for structural ex
ante regulation
• Private ordering?
Solution 2 : everything but the AI Act!
1. Copyright & AI generated art: another
entire talk
• WaPost : “He used AI to win
a fine-arts competition. Was
it cheating?
• One judge said the striking
piece evoked Renaissance
art. But some critics
compared it to ‘entering a
marathon and driving a
Lamborghini to the finish
line.’”
• Effect on original artists? A
tool or a replacement??
• Boris Eldagsen The Electrician
• “winner” Sony World
Photography Awards,
• 17 April 2023
AI generated art and
copyright
• Often copyright in the art used as training dataset
inputs (eg Rutowski) – but they are not part of the
model
• No direct copying though sometimes a perfect copy
might emerge (memorisation)
• Is there actual copying of inputs?
• Is there (US) fair use or (UK/DSA) research or TDM
exception?
• Who owns the outputs?
• Again almost never direct copies
• “After the style of”..
• Derivative works?
• Some partial solutions : opt-out; haveIbeentrained ;
license, royalties or benefit sharing
• Litigation!
AI art litigation (US)
• Getty v Stability, based on copying of input works copyright Getty
• Transformative fair use?
• Anderson vs Stability & Deviant Art
• Aims to acquire rights over all OUTPUTS as derivative works from artistic works in
training set
• Defendants case to dismiss -
• NB Stability is open source, so you can analyse the underlying training sets
(cf Open (sic) AI & GPT-4)
If you thought artists were
p*ssed off by genAI..
… try the music
industry!
Can GPAI be compliant with GDPR?
• AIA high risk data quality and
transparency requirements actually
say nothing about privacy of users
• Machine learning process long
regarded as dubious but
• no federal DP law in US
• no quality of privacy,
confidentiality to data made
public in US
• GPAIs use permissionless public
data (eg Common Crawl)
• But - Replika decision, Italy, 2/2/23
• Primarily about exposure of children
to unsuitable sexualized material
• Unable to make valid contracts
Italy DPA Garantie vs ChatGPT, 31 March 2023
“
“
How to regulate foundation models: can we do better than the EU AI Act?
Open AI’s Italian Job(s)
What next?
• “reports of a 400% surge in VPN
downloads in Italy” 
• Spain, France, pan-European investigation
by EDPB
• Canada investigation
• Does GPAI fundamentally challenge the
GDPR?
• If so, which gives?? (and does it take ML
with it?)
• The end of the data Wild West?
• But are the privacy regulators really the
right ones to take generative AI on?
• (and will the UK become a light-touch AI
regulation” law haven” for Chat GPT!)
Private ordering – control?
• Contracts and licenses
• Eg Responsible AI Licensing
(RAIL)(FAccT ‘22)
• Model cards, model sheets
• Privacy policies
• EPSRC Generative AI terms of Service
project Jan-March 2023, report end
April !
• Technical controls
• Filters : eg Open AI NSFW filter
• API control eg Project December app
• Internal human moderation when
finetuning the model pre release
• Watermarks for generated content eg PAI
Guidelines for Synthetic Media
• Enforcement?
• Privity of contract
AGI-nising choices: Nuclear arm or buggy software?
"Should we let machines flood our
information channels with
propaganda and
untruth? Should we automate
away all the jobs, including the
fulfilling ones? Should we develop
nonhuman minds that might
eventually outnumber, outsmart,
obsolete and replace
us? Should we risk loss of control
of our civilization?" (source;
emphasis in original)
Future of Life Institute Open Letter, March
22, 2023
Narayanan and Kapoor https://aisnakeoil.substack.com/p/a-
misleading-open-letter-about-sci?
CREDITS
Parrot images by James Stewart, Edinburgh University ,
@datacontroversies
“after” various unknown artists, using MidJourney, 2023.
Stable Diffusion images created by and © Lilian Edwards, 2023, made
with own photo.
Image of Emily Bender with parrot © New York Times, 2023
Image on slide 12 taken with thanks from Thomas Hoppner
“ChatGPT, Bard & Co.: an introduction to AI for competition and
regulatory lawyers”, 2023 at https://www.hausfeld.com/en-
gb/what-we-think/competition-bulletin/chatgpt-bard-co-an-
introduction-to-ai-for-competition-and-regulatory-lawyers/

More Related Content

PPTX
AI Governance: Responsible and Trustworthy AI
PPTX
Responsible AI
PPTX
Global Governance of Generative AI: The Right Way Forward
PDF
“Responsible AI: Tools and Frameworks for Developing AI Solutions,” a Present...
PPTX
AI Governance and Ethics - Industry Standards
PDF
AI Act: Key Provisions of the First Major Law Regulating AI
PPTX
00_Main ppt_AI Act & EU AI Regulation.pptx
PDF
Navigating the New EU AI Act (Data & Analytics)
AI Governance: Responsible and Trustworthy AI
Responsible AI
Global Governance of Generative AI: The Right Way Forward
“Responsible AI: Tools and Frameworks for Developing AI Solutions,” a Present...
AI Governance and Ethics - Industry Standards
AI Act: Key Provisions of the First Major Law Regulating AI
00_Main ppt_AI Act & EU AI Regulation.pptx
Navigating the New EU AI Act (Data & Analytics)

What's hot (20)

PPTX
Generative AI and law.pptx
PDF
Generative-AI-in-enterprise-20230615.pdf
PDF
Introduction to AI Ethics
PPTX
The Future of AI is Generative not Discriminative 5/26/2021
PDF
What is Artificial Intelligence | Artificial Intelligence Tutorial For Beginn...
PDF
An Introduction to Generative AI
PPTX
Can ChatGPT be compatible with the GDPR? Discuss.
PDF
The Future is in Responsible Generative AI
PPTX
Generative AI Risks & Concerns
PPTX
Webinar on ChatGPT.pptx
PDF
Leveraging Generative AI & Best practices
PDF
𝐆𝐞𝐧𝐞𝐫𝐚𝐭𝐢𝐯𝐞 𝐀𝐈: 𝐂𝐡𝐚𝐧𝐠𝐢𝐧𝐠 𝐇𝐨𝐰 𝐁𝐮𝐬𝐢𝐧𝐞𝐬𝐬 𝐈𝐧𝐧𝐨𝐯𝐚𝐭𝐞𝐬 𝐚𝐧𝐝 𝐎𝐩𝐞𝐫𝐚𝐭𝐞𝐬
PPTX
PPTX
How to Teach and Learn with ChatGPT - BETT 2023
PDF
Explainability and bias in AI
PDF
AI 2023.pdf
PPTX
Responsible AI in Industry (ICML 2021 Tutorial)
PPTX
Explainable AI in Industry (KDD 2019 Tutorial)
PDF
The Rise of the LLMs - How I Learned to Stop Worrying & Love the GPT!
PDF
Design Ethics for Artificial Intelligence
Generative AI and law.pptx
Generative-AI-in-enterprise-20230615.pdf
Introduction to AI Ethics
The Future of AI is Generative not Discriminative 5/26/2021
What is Artificial Intelligence | Artificial Intelligence Tutorial For Beginn...
An Introduction to Generative AI
Can ChatGPT be compatible with the GDPR? Discuss.
The Future is in Responsible Generative AI
Generative AI Risks & Concerns
Webinar on ChatGPT.pptx
Leveraging Generative AI & Best practices
𝐆𝐞𝐧𝐞𝐫𝐚𝐭𝐢𝐯𝐞 𝐀𝐈: 𝐂𝐡𝐚𝐧𝐠𝐢𝐧𝐠 𝐇𝐨𝐰 𝐁𝐮𝐬𝐢𝐧𝐞𝐬𝐬 𝐈𝐧𝐧𝐨𝐯𝐚𝐭𝐞𝐬 𝐚𝐧𝐝 𝐎𝐩𝐞𝐫𝐚𝐭𝐞𝐬
How to Teach and Learn with ChatGPT - BETT 2023
Explainability and bias in AI
AI 2023.pdf
Responsible AI in Industry (ICML 2021 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)
The Rise of the LLMs - How I Learned to Stop Worrying & Love the GPT!
Design Ethics for Artificial Intelligence
Ad

Similar to How to regulate foundation models: can we do better than the EU AI Act? (20)

PPTX
Generative AI, responsible innovation and the law
DOCX
EU Artificial Intelligence Act (High-level summary of the AI Act)
PDF
Both Feet on the Ground - Generative Artificial Intelligence
PDF
FROM VICTIM TO VICTOR:HOW ASSET RESCUE SPECIALIST RECOVERED MY ASSET
PDF
“What Happens When Your Speed of AI Innovation Exceeds Your Ability to See th...
PPTX
Marsden AI coregulation in Europe, Macquarie AILS
PDF
Brandsmiths Intelligence On Ai Where Are We Now.pdf
PPTX
Research Summit 2025 - Path to AI Act.pptx
PPTX
[DSC DACH 23] AI Regulation - How to implement AI legally compliant? - Alexan...
PPTX
[DSC Europe 23] Shahab Anbarjafari - Generative AI: Impact of Responsible AI
PDF
Domande e risposte su AI (Commissione Europea, 12 dic 2023)
PPTX
Introduction to LLMs and their relevance for Official Statistics
PDF
Ai e-accountability
PPTX
[DSC DACH 24] Transparency as a catalyst for trustworthy and sustainable AI -...
PPTX
Compliance in AI: Policies and Best Practices
PPTX
Tendencias globales en la regulación de la IA y estándares tecnológicos asoci...
PDF
ChatGPT, Generative AI Data Security Considerations
PDF
TrustArc Webinar - Data Privacy Management in the Age of AI
PPTX
The EU ‘AI ACT’: a “risk-based” legislation for robotic surgery
PPTX
OpenChain Webinar - AI Legal Landscape - Slides
Generative AI, responsible innovation and the law
EU Artificial Intelligence Act (High-level summary of the AI Act)
Both Feet on the Ground - Generative Artificial Intelligence
FROM VICTIM TO VICTOR:HOW ASSET RESCUE SPECIALIST RECOVERED MY ASSET
“What Happens When Your Speed of AI Innovation Exceeds Your Ability to See th...
Marsden AI coregulation in Europe, Macquarie AILS
Brandsmiths Intelligence On Ai Where Are We Now.pdf
Research Summit 2025 - Path to AI Act.pptx
[DSC DACH 23] AI Regulation - How to implement AI legally compliant? - Alexan...
[DSC Europe 23] Shahab Anbarjafari - Generative AI: Impact of Responsible AI
Domande e risposte su AI (Commissione Europea, 12 dic 2023)
Introduction to LLMs and their relevance for Official Statistics
Ai e-accountability
[DSC DACH 24] Transparency as a catalyst for trustworthy and sustainable AI -...
Compliance in AI: Policies and Best Practices
Tendencias globales en la regulación de la IA y estándares tecnológicos asoci...
ChatGPT, Generative AI Data Security Considerations
TrustArc Webinar - Data Privacy Management in the Age of AI
The EU ‘AI ACT’: a “risk-based” legislation for robotic surgery
OpenChain Webinar - AI Legal Landscape - Slides
Ad

More from Lilian Edwards (20)

PPTX
What Do You Do with a Problem Like AI?
PPTX
The GDPR, Brexit, the UK and adequacy
PPTX
Slave to the Algorithm 2016
PPTX
Cloud computing : legal , privacy and contract issues
PPTX
Privacy, the Internet of Things and Smart Cities
PPTX
From Privacy Impact Assessment to Social Impact Assessment: Preserving TRrus...
PPTX
UK copyright, online intermediaries and enforcement
PPTX
The GDPR for Techies
PPTX
the Death of Privacy in Three Acts
PPTX
Revenge porn: punish, remove, forget, forgive?
PPTX
From piracy to “The Producers?
PPTX
The Death of Privacy in Three Acts
PPTX
Police surveillance of social media - do you have a reasonable expectation of...
PPTX
IT law : the middle kingdom between east and West
PPTX
What do we do with aproblem like revenge porn ?
PPTX
Slave to the Algo-Rhythms?
PPTX
9worlds robots
PPTX
The death of data protection
PPTX
The death of data protection sans obama
PPTX
Cdas 2012, lilian edwards and edina harbinja
What Do You Do with a Problem Like AI?
The GDPR, Brexit, the UK and adequacy
Slave to the Algorithm 2016
Cloud computing : legal , privacy and contract issues
Privacy, the Internet of Things and Smart Cities
From Privacy Impact Assessment to Social Impact Assessment: Preserving TRrus...
UK copyright, online intermediaries and enforcement
The GDPR for Techies
the Death of Privacy in Three Acts
Revenge porn: punish, remove, forget, forgive?
From piracy to “The Producers?
The Death of Privacy in Three Acts
Police surveillance of social media - do you have a reasonable expectation of...
IT law : the middle kingdom between east and West
What do we do with aproblem like revenge porn ?
Slave to the Algo-Rhythms?
9worlds robots
The death of data protection
The death of data protection sans obama
Cdas 2012, lilian edwards and edina harbinja

Recently uploaded (20)

PPTX
lecture 5.pptx on family law notes well detailed
PDF
8-14-25 Examiner Report from NJ Bankruptcy (Heller)
PDF
AI in Modern Warfare and Business Ethics Ortynska Law Ventures Cafe.pdf
PDF
devolution-handbook (1).pdf the growh of devolution from 2010
PDF
250811-FINAL-Bihar_Voter_Deletion_Analysis_Presentation.pdf
PPTX
Constitutional Law 2 Final Report.ppt bill of rights in under the constitution
PDF
APPELLANT'S AMENDED BRIEF – DPW ENTERPRISES LLC & MOUNTAIN PRIME 2018 LLC v. ...
PPTX
4-D...Preparation of Research Design.pptx
PPTX
Philippine Politics and Governance - Lesson 10 - The Executive Branch
PPT
SDEAC-2020-Leaves-of-Absence-Presentation-Daniel-De-La-Cruz.ppt
PPTX
Sexual Harassment Prevention training class
PPTX
PPT in Consti 2 Report (Week1).pptx under the contituiton
PPTX
Punjab Fertilizers Control Act 2025.pptx
PPTX
Behavioural_Approach_Public_Administration_Zambia_USA.pptx
PPTX
BL 2 - Courts and Alternative Dispute Resolution.pptx
PPTX
Classifying Different Branches of Law.pptx
PPTX
Indian Medical Device Rules or Institute of Management Development and Research
PPTX
Financial Rehabilitation and Insolvency Act
PDF
2022CH12581 - Civil Rights vs Morzak, Harrison, Chrisman et al. (Cook County,...
lecture 5.pptx on family law notes well detailed
8-14-25 Examiner Report from NJ Bankruptcy (Heller)
AI in Modern Warfare and Business Ethics Ortynska Law Ventures Cafe.pdf
devolution-handbook (1).pdf the growh of devolution from 2010
250811-FINAL-Bihar_Voter_Deletion_Analysis_Presentation.pdf
Constitutional Law 2 Final Report.ppt bill of rights in under the constitution
APPELLANT'S AMENDED BRIEF – DPW ENTERPRISES LLC & MOUNTAIN PRIME 2018 LLC v. ...
4-D...Preparation of Research Design.pptx
Philippine Politics and Governance - Lesson 10 - The Executive Branch
SDEAC-2020-Leaves-of-Absence-Presentation-Daniel-De-La-Cruz.ppt
Sexual Harassment Prevention training class
PPT in Consti 2 Report (Week1).pptx under the contituiton
Punjab Fertilizers Control Act 2025.pptx
Behavioural_Approach_Public_Administration_Zambia_USA.pptx
BL 2 - Courts and Alternative Dispute Resolution.pptx
Classifying Different Branches of Law.pptx
Indian Medical Device Rules or Institute of Management Development and Research
Financial Rehabilitation and Insolvency Act
2022CH12581 - Civil Rights vs Morzak, Harrison, Chrisman et al. (Cook County,...

How to regulate foundation models: can we do better than the EU AI Act?

  • 1. How to regulate foundation models: can we do better than the EU AI Act? Lilian Edwards Professor of Law, Newcastle University @lilianedwards Lilian.edwards@ncl.ac.uk April 2023
  • 2. What are large or “foundation” models? • GPT-2/3/3.5/4 (Open AI/Microsoft)(prompt to text)(2019 on) • “Large Language Model” or LLM • ChatGPT • DALL-E 2 (text to images – Google) • Stable Diffusion (open source – text to image) • HarmonAI – makes AI generated music (Stability) • CoPilot (prompt generates computer code – GitHub/OpenAI) • Meta Make-me-A-Video (text to video - Meta) • ERNIE ( Baidu, China) (prompt to text)
  • 3. eg Stable Diffusion : img to img (open source third party code)
  • 9. Integration into search, Feb 2023 New York Times
  • 11. Important (for law) features of large or “foundation” models • Generative – create text, images etc rather than merely classifying or predicting (ML) • Trained on unprecedentedly large datasets • Often scraped from “public” Internet • Impossible to manually review legality, privacy or harm of every item in datasets • Computationally expensive and retraining slow -> • large tech co dominance • GPT-4 training cost >$100mn • environmentally worrying • Training sets allow the model to assess probability of next word, pixel etc – not direct copying • Models are general , can have multiple uses, eg to write a party invite, a racist attack or provide customer support within an automated hiring system • Generated content increasingly difficult to distinguish from human-created content (disinfo, deepfakes) • Outputs may be “hallucinations” Hoppner, 2023
  • 12. Issues with large models PHASE 1 STOCHASTIC PARROTS • Don’t actually understand , just “parrot” • Bias, discrimination, misrepresentation and stereotyping of groups; hate speech PHASE 2 WILL NO-ONE THINK OF THE ARTISTS? • Image and video deepfakes • Pastiche • Copyright PHASE 3 FAKE NEWS ON STEROIDS • Fake news and “hallucination” (text + images] • Education & plagiarism • Digital Services Act PHASE 4 – YOU HAVE ZERO PRIVACY, GET OVER IT • GDPR
  • 13. Solution 1 : the EU AI Act and “GPAI” AIA “risk based” approach • Unacceptable risk – ‘Complete’ prohibition, 4 examples – Article 5 • High-risk –Fixed categories of risky domains, based on intended use ; “essential requirements” including dataset quality, human oversight – • Limited risk – Transparency obligations for a few AI systems (chatbots, deepfakes, emotion ID, biometric categorisation) – Article 52 • Minimal risk – Codes of conduct – Article 69 Photo Source: European Commission, Digital Strategy Website https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai
  • 14. Annex II - Products • Machinery • Toys • Recreational craft and watercraft • Lifts • Equipment and protective systems intended for use in potentially explosive atmospheres • Radio equipment • Pressure equipment • Cableway installations • PPE • Medical devices • [...] High-Risk AI systems (Designation) - Annexes II&III • Annex III - Services • Biometric identification and categorisation of natural persons; • Management and operation of critical infrastructure; • Education and vocational training; • Employment, workers management and access to self-employment; • Access to and enjoyment of essential private services and public services and benefit; • Law enforcement; • Migration, asylum and border control management; • Administration of justice and democratic processes
  • 15. • Compliance with requirements (Art.8); • Risk management system (Art.9); • Data and data governance (Art.10); (“data quality”) • “Training, validation and testing data sets shall be relevant, representative, free of errors and complete. They shall have the appropriate statistical properties, including, where applicable, as regards the persons or groups of persons on which the high-risk AI system is intended to be used.” • Technical documentation (Art.11); • Record-keeping (Art.12); • Transparency and provision of information to users (Art.13); • Human oversight (Art.14); • “Human oversight shall aim at preventing or minimising the risks to health, safety or fundamental rights that may emerge when a high- risk AI system is used in accordance with its intended purpose or under conditions of reasonably foreseeable misuse” • Accuracy, robustness and cybersecurity (Art.15). High-Risk AI systems (Requirements) – Arts. 8 - 15
  • 17. Definition of GPAI in EU AIA LGAIMs = Large General AI Models – Hacker, Engel and Mauer “Regulating ChatGPT and other Large generative AI Models”, February 2023 Issues : over-inclusive (no emphasis generality in “ability, task or output”); important for classification -> Council position ‘general purpose AI system’ means an AI system that is trained on broad data at scale, is designed for generality of output, and can be adapted to a wide range of tasks EU Parliament, March 2023
  • 18. Developers v deployers • Developers? • Akin to manufacturers – control and have knowledge of the training sets, weights, algorithms, content moderation etc (esp if closed source) • Most high-risk obligations arise at development stage (training & human fine-tuning stages) • Have practical power and economic gains • But can they handle unforeseeable uses/risks? Or can only tech giants  competition issues? Open source providers? • Deployers? [NB “users” in AIA!] • Originally only duties on deployer if they make a “substantial modification” to the AI system, ie become new provider • Practicality ?: May be impossible for them to fix or even audit issues of data quality etc without access to upstream source code, training datasets etc (often secret/proprietary – eg as with GPT-3) • “AI as a service” API/cloud model will be prevalent till models smaller “Pick n Mix?”
  • 19. Solutions in AIA process? • 13 May 2022, French presidency added amendment excluding GPAIs from AIA • Back and forth.. • Council position, Dec 2022, arts 4a-4c • GPAI deemed high risk if they “may be used as high risk AI systems or as components of high risk AI systems” • Unless “explicitly excluded all high-risk uses”, but not if not in “good faith” • European Parliament • All generative AI models to go into high risk if • they generate text that might be mistaken for human • And all deepfakes and AV content showing something that “never happened” unless an “obvious artistic work” • & Providers owe cooperation & transparency obligations to downstream users • Commission to tweak the high-risk obligations by delegated Acts..
  • 20. Foundation models v GPAI?? 19/4/23 “Foundation models” • “an AI system model that is trained on broad data at scale, is designed for generality of output, and can be adapted to a wide range of distinctive tasks.” • Eg Chat GPT , Stable Diffusion • “Trained on data scraped from entire Internet”? Not just if labelled • Stricter obligations - High risk ++? • Adds sustainability obligations + independent expert ex ante oversight • Documented analysis and testing throughout lifecycle • Disclosures re copyright in training set; filters to avoid delivery unlawful content “GPAI” • “AI system that can be used in and adapted to a wide range of applications for which it was not intentionally and specifically designed.” • ?? E.g. “unlabelled data that need further training by the provider, such as algorithms developed to recognise skin cancer” • Laxer regime, since obligations fall on high risk providers who build on the data (?) Deployers & value chain • “non-binding standard contractual clauses that regulate rights and obligations consistent with each party’s level of control”
  • 21. • Discrimination/ equality law • Liability (product liability for AI) • Copyright • Content, hate speech, libel, fake news (DSA) • Privacy & data protection • Personality & image rights • Competition law • See Hoppner https://papers.ssrn.com/sol3/paper s.cfm?abstract_id=4371681 • Apart from last though, only AIA and DMA allows for structural ex ante regulation • Private ordering? Solution 2 : everything but the AI Act!
  • 22. 1. Copyright & AI generated art: another entire talk • WaPost : “He used AI to win a fine-arts competition. Was it cheating? • One judge said the striking piece evoked Renaissance art. But some critics compared it to ‘entering a marathon and driving a Lamborghini to the finish line.’” • Effect on original artists? A tool or a replacement??
  • 23. • Boris Eldagsen The Electrician • “winner” Sony World Photography Awards, • 17 April 2023
  • 24. AI generated art and copyright • Often copyright in the art used as training dataset inputs (eg Rutowski) – but they are not part of the model • No direct copying though sometimes a perfect copy might emerge (memorisation) • Is there actual copying of inputs? • Is there (US) fair use or (UK/DSA) research or TDM exception? • Who owns the outputs? • Again almost never direct copies • “After the style of”.. • Derivative works? • Some partial solutions : opt-out; haveIbeentrained ; license, royalties or benefit sharing • Litigation!
  • 25. AI art litigation (US) • Getty v Stability, based on copying of input works copyright Getty • Transformative fair use? • Anderson vs Stability & Deviant Art • Aims to acquire rights over all OUTPUTS as derivative works from artistic works in training set • Defendants case to dismiss - • NB Stability is open source, so you can analyse the underlying training sets (cf Open (sic) AI & GPT-4)
  • 26. If you thought artists were p*ssed off by genAI.. … try the music industry!
  • 27. Can GPAI be compliant with GDPR? • AIA high risk data quality and transparency requirements actually say nothing about privacy of users • Machine learning process long regarded as dubious but • no federal DP law in US • no quality of privacy, confidentiality to data made public in US • GPAIs use permissionless public data (eg Common Crawl) • But - Replika decision, Italy, 2/2/23 • Primarily about exposure of children to unsuitable sexualized material • Unable to make valid contracts
  • 28. Italy DPA Garantie vs ChatGPT, 31 March 2023 “ “
  • 31. What next? • “reports of a 400% surge in VPN downloads in Italy”  • Spain, France, pan-European investigation by EDPB • Canada investigation • Does GPAI fundamentally challenge the GDPR? • If so, which gives?? (and does it take ML with it?) • The end of the data Wild West? • But are the privacy regulators really the right ones to take generative AI on? • (and will the UK become a light-touch AI regulation” law haven” for Chat GPT!)
  • 32. Private ordering – control? • Contracts and licenses • Eg Responsible AI Licensing (RAIL)(FAccT ‘22) • Model cards, model sheets • Privacy policies • EPSRC Generative AI terms of Service project Jan-March 2023, report end April ! • Technical controls • Filters : eg Open AI NSFW filter • API control eg Project December app • Internal human moderation when finetuning the model pre release • Watermarks for generated content eg PAI Guidelines for Synthetic Media • Enforcement? • Privity of contract
  • 33. AGI-nising choices: Nuclear arm or buggy software? "Should we let machines flood our information channels with propaganda and untruth? Should we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk loss of control of our civilization?" (source; emphasis in original) Future of Life Institute Open Letter, March 22, 2023 Narayanan and Kapoor https://aisnakeoil.substack.com/p/a- misleading-open-letter-about-sci?
  • 34. CREDITS Parrot images by James Stewart, Edinburgh University , @datacontroversies “after” various unknown artists, using MidJourney, 2023. Stable Diffusion images created by and © Lilian Edwards, 2023, made with own photo. Image of Emily Bender with parrot © New York Times, 2023 Image on slide 12 taken with thanks from Thomas Hoppner “ChatGPT, Bard & Co.: an introduction to AI for competition and regulatory lawyers”, 2023 at https://www.hausfeld.com/en- gb/what-we-think/competition-bulletin/chatgpt-bard-co-an- introduction-to-ai-for-competition-and-regulatory-lawyers/

Editor's Notes

  • #8: P Hacker “The propensity of ChatGPT particularly to hallucinate when it does not find readymade answers can be exploited to generate text devoid of any connection to reality, but written in the style of utter confidence”
  • #9: Clip art getty photos Book covers Early product concept, ads, design architecture mock ups Can add shadows, extend out paintings, produce programme credits, instant editorial cartooning
  • #13: Eg anthropomorphisation of CBT chatbots, postmortem avatars* “Socio-economic “ harms Energy – very costly to train giant models
  • #23: Bans from fan fora – lots of fantasy art but very derivative =- in its nature! But – effort? “He started with a simple mental image — “a woman in a Victorian frilly dress, wearing a space helmet” — and kept fine-tuning the prompts, “using tests to really make an epic scene, like out of a dream.” He said he spent 80 hours making more than 900 iterations of the art, adding words like “opulent” and “lavish” to fine tune its tone and feel. He declined to share the full series of words he used to create his art, saying it is his artistic product, and that he intends to publish it later. “If there’s one thing you can take ownership of, it’s your prompt,” he said.” A tool or a copy?
  • #28: Transparency is to downstream deployer not data subject; data quality is not about whether personal data that was permissioned
  • #33: Eg Project December postmortem avatar app – app withdrawn by GPT-3 API control as inappropriate, but not on application of end-user Later they stopped asking you to fill in a form re the probity of your applicatiuon and now its just money!