Velocity Media Blog New

AI and Liability: Key Insights from Recent EU Legislation

Written by Shawn Greyling | Jul 15, 2024 8:27:01 AM

The rapid evolution of artificial intelligence (AI) technology necessitates a comprehensive legal framework to address potential liabilities and ensure safety, transparency, and accountability. The European Union (EU) is pioneering this effort with its Artificial Intelligence Regulation, commonly referred to as the AI Act, which is set to come into force in the summer of 2024. This article explores the key takeaways from the AI Act and other significant EU legislative initiatives, focusing on the Revised Product Liability Directive (Revised PLD) and the proposed AI Liability Directive (AILD). We will examine the implications for businesses, consumers, and the broader AI ecosystem.

Covered in this article

Understanding Artificial Intelligence and Liability
Overview of Recent EU Legislative Initiatives
Key Takeaways from the EU Legislative Initiatives
The AI Liability Directive (AILD)
Impact on Businesses and Consumers
Challenges and Opportunities
Future Outlook
Conclusion
FAQs

Understanding Artificial Intelligence and Liability

Artificial intelligence and liability intersect at the point where the actions or inactions of AI systems can cause harm to users or property. As AI systems become more autonomous and complex, determining liability for damages becomes increasingly challenging. The EU's legislative initiatives aim to address these challenges by establishing clear rules and guidelines for AI developers and users.

Overview of Recent EU Legislative Initiatives

The EU has introduced several legislative measures to regulate AI, with the AI Act being the most comprehensive. The AI Act aims to ensure that AI systems used in the EU are safe, transparent, traceable, non-discriminatory, and environmentally friendly. In addition to the AI Act, the EU is revising its product liability regime to address the unique challenges posed by AI. Key legislative initiatives include:

  • Revisions to the Product Liability Directive (85/374/EEC) (Revised PLD)
  • Introduction of the AI Liability Directive (AILD)

Key Takeaways from the EU Legislative Initiatives

The Revised Product Liability Directive (Revised PLD)

Adopted by the European Parliament in March 2024, the Revised PLD is awaiting approval by the European Council. It introduces several amendments to address AI-specific issues:

Definition of “Product” and “Defect”

  • Product: Includes software, encompassing AI systems. Free and open-source software developed or supplied outside commercial activity is excluded, but manufacturers integrating such software may be liable for defects.
  • Defect: Now includes the product's ability to self-learn and acquire new features. AI systems must be designed to prevent hazardous behaviour.

Expanded Defendant Categories

  • Companies that modify products outside the manufacturer's control can be held liable.
  • Manufacturers of defective components integrated into a product are also liable.

Presumption of Defectiveness and Causation

  • Imposes a presumption of defectiveness if the product fails mandatory safety requirements or malfunctions during normal use.
  • Establishes a causal link between defect and damage, aiding claimants in proving their cases.

Definition of “Damage”

  • Includes loss or corruption of data and medically recognised psychological harm.
Legislation Key Features Impact on Businesses Impact on Consumers
AI Act - Comprehensive legal framework for AI - Must ensure AI systems are safe, transparent, traceable, non-discriminatory, and environmentally friendly - Enhanced protection and transparency
Revised Product Liability Directive (Revised PLD) - Expanded definitions of “product” and “defect”- New categories of defendants - Presumptions of defectiveness and causation - Liability for software upgrades, self-learning AI systems, and cybersecurity- Need to disclose relevant information - Easier to claim compensation for AI-related damages- Protection against defective AI products
AI Liability Directive (AILD) - Assists in making non-contractual fault-based claims - Disclosure of evidence and presumption of causation - Must disclose evidence for high-risk AI systems- Need to address technical complexities in proving fault - Simplifies proving fault in AI-related claims- Increased ability to seek compensation for AI-caused damages
Common Elements - Focus on safety, transparency, and consumer protection- Enhanced claimant support mechanisms - Must update compliance and risk frameworks- Ensure AI products meet new EU standards - Better recourse and compensation mechanisms for AI-related harms

The AI Liability Directive (AILD)

The AILD assists claimants in making non-contractual fault-based claims for damage caused by AI systems. Key aspects include:

Disclosure of Evidence and Presumption of Non-Compliance

  • Providers or users of high-risk AI systems can be ordered to disclose evidence if claimants prove the plausibility of their claims.

Presumption of Causation

  • Presumes a causal link between a fault and the AI system's output or failure to produce an output if claimants meet specific conditions.

Impact on Businesses and Consumers

The new regulations will significantly impact businesses developing or using AI systems. Companies must adapt their risk and compliance frameworks to align with the Revised PLD and AILD requirements. For consumers, these initiatives enhance protection and provide clearer avenues for recourse and compensation in cases of harm caused by AI systems.

Challenges and Opportunities

Challenges

  • Compliance costs and technical barriers.
  • Navigating the complexities of new legal requirements.

Opportunities

  • Enhanced consumer trust in AI technologies.
  • Potential for innovation in developing safer and more reliable AI systems.

Future Outlook

The AI Act and associated directives are just the beginning of a broader regulatory framework that will evolve with advancements in AI technology. Businesses must stay informed and proactive in adapting to these changes to mitigate risks and leverage growth opportunities.

Conclusion

The EU's legislative initiatives on AI liability mark a significant step towards ensuring AI technologies' safe and responsible use. Businesses involved in AI must thoroughly understand and comply with these new regulations to minimise liability risks and capitalise on emerging opportunities. Staying informed and prepared is crucial as the AI landscape continues to evolve.

FAQs About AI EU Legislation

1. What is the AI Act?

The AI Act is a comprehensive legal framework introduced by the EU to regulate the use and development of artificial intelligence systems, ensuring they are safe, transparent, traceable, non-discriminatory, and environmentally friendly.

2. When will the AI Act come into force?

The AI Act is expected to come into force in the summer of 2024.

3. What are the key components of the Revised Product Liability Directive (Revised PLD)?

The Revised PLD includes expanded definitions of “product” and “defect,” new categories of defendants, obligations for manufacturers to disclose information, and presumptions of defectiveness and causation to assist claimants.

4. What is the AI Liability Directive (AILD)?

The AILD is a proposed EU directive to assist claimants in making non-contractual fault-based claims for damage caused by AI systems, focusing on disclosure of evidence and presumption of causation.

5. How will the new EU regulations impact businesses?

Businesses will need to update their risk and compliance frameworks, ensure AI systems meet new safety and transparency standards, and be prepared to defend against liability claims.

6. What new liabilities do AI system manufacturers face under the Revised PLD?

Manufacturers are liable for defects arising from software upgrades, AI systems' ability to self-learn, and failure to supply necessary security updates. Liability extends to integrated defective components and modified products.

7. How does the Revised PLD define “defect” in AI systems?

A defect includes the AI system’s ability to self-learn, acquire new features, and comply with cybersecurity requirements. The product must prevent hazardous behaviour and provide necessary updates.

8. What kind of damages are covered under the Revised PLD?

The Revised PLD covers material damage, loss or corruption of data (excluding professional data), and medically recognised psychological harm.

9. What is the significance of the presumption of defectiveness in the Revised PLD?

The presumption of defectiveness helps claimants prove their case by assuming a product is defective if it fails safety requirements, malfunctions during normal use, or if the manufacturer fails to disclose relevant information.

10. How can businesses prepare for the new EU AI regulations?

Businesses should conduct risk assessments, ensure compliance with new standards, maintain detailed documentation, and update contractual protections and insurance policies to address new liabilities.