Embracing a New Era: The World’s First Major AI Regulation Act

March 29, 2024 | What's Hot | 4 min read

It happened.

On March 13, 2024, the European Parliament adopted the Artificial Intelligence Act (AI Act), the first major law to regulate artificial intelligence.

Lawmakers established a comprehensive framework that addresses the complexities and potential risks associated with AI technology. As a strong advocate for responsible AI development, I believe this legislation marks a significant step forward in ensuring that AI serves humanity’s best interests while minimizing potential harm.

The essence of this pioneering law lies in its adoption of a risk-based approach to AI regulation. It categorizes AI applications into distinct risk levels. Each category is accompanied by tailored regulations and oversight mechanisms, reflecting a nuanced understanding of the diverse capabilities and implications of AI systems.

  • Unacceptable risk: These AI systems and practices pose a clear threat to fundamental rights and are prohibited. Examples include AI systems that manipulate human behavior or exploit vulnerabilities such as age or disability. Biometric systems like emotion recognition in workplaces or real-time categorization of individuals are also prohibited.
  • High risk: AI systems identified as high-risk must comply with strict requirements, including risk-mitigation systems, high-quality data sets, activity logging, detailed documentation, clear user information, human oversight, and robust cybersecurity. Examples of high-risk AI systems include those systems that impact critical infrastructures like energy and transport, medical devices, and systems determining access to education or employment.
  • Limited risk: Limited risk systems must increase transparency about how and when AI is used to bring about their solution. For example, systems intended to directly interact with natural persons, such as chatbots, must inform individuals that they are interacting with an AI system. Deployers of AI systems generating or manipulating deepfakes must disclose that the content has been generated through AI or has been manipulated.
  • Minimal risk: No restrictions are imposed on minimal-risk AI systems, such as AI-enabled video games or spam filters. However, companies may opt to adhere to voluntary codes of conduct.

Why a Risk-Based Approach to AI Is the Right Approach

One of the most commendable aspects of this legislation is its stringent controls on high-risk AI use cases. Recognizing the potential for significant societal impact and ethical implications, the law mandates intense approvals and controls for AI systems deemed to pose a high risk to individuals, communities, or fundamental rights. By subjecting such applications to rigorous scrutiny and oversight, the legislation seeks to mitigate the potential harms associated with AI while promoting innovation and progress and deter such development of systems that may produce those harms.

Equally crucial are the transparency and accountability requirements imposed on limited-risk AI use cases. While these applications may not pose as immediate or severe risks as their high-risk counterparts, they still warrant careful monitoring, adaptation and regulation.

The law emphasizes the importance of transparency in AI decision-making processes, ensuring that individuals understand how AI systems operate and can hold developers accountable for their actions. By fostering transparency and accountability, the legislation promotes trust and confidence in AI technologies, essential for their widespread adoption and acceptance.

However, the true challenge lies in navigating the ambiguity inherent in distinguishing between limited-risk use cases and those that could potentially escalate into high-risk scenarios.

As AI continues to snake its way through various aspects of society, identifying and addressing these gray areas will be paramount. It requires ongoing vigilance, collaboration, and a willingness to adapt regulatory frameworks in response to emerging threats and challenges.

Despite these complexities, I am optimistic about the future of AI regulation. The enactment of this law represents a significant milestone in our collective journey towards harnessing the potential of AI for the greater good.

By adopting a risk-based approach and prioritizing transparency and accountability, we are actively working to lay the foundation for a more responsible and sustainable AI ecosystem. And as we embark on this new era of AI regulation, let’s remain vigilant, proactive, and committed to ensuring that AI serves as a force for good in our rapidly evolving world.

Read Next: The Jobs A.I. Can and Cannot Replace (and Why You Shouldn't Worry)

Subscribe to the Skillsoft Blog

We will email when we make a new post in your interest area.

Select which topics to subscribe to:

Start Thinking about AI Guardrails Now

My final thought is this: The AI Act is a decisive step in challenging how society thinks about and interacts with AI. In fact, it puts human beings in the driver’s seat – allowing us to contribute to the narrative of AI as it unfolds.

Even if the law doesn’t impact your organization due to its geographic location or the risk level in which you deploy AI, there’s a lot you can take away from it.

For example: one of the most impactful ways your organization can contribute to the conversation around AI in the near term is to establish an AI policy for its employees. A formal AI policy defines how employees can use AI within your organization and typically covers ethical usage, bias and fairness standards, compliance requirements, and other critical guidelines and guardrails.

As AI best practices evolve in your country, so will your AI policy.