October 28, 2025

The AI Risk Pyramid: Europe’s Strategy for Control?

Imagine if artificial intelligence was no longer just a tool, but a decision-maker? In Albania, an AI tool named Diella was appointed to a ministerial role few weeks ago, with a task to make government services more efficient and transparent. 

This avant-garde move begs a crucial question: as AI becomes a part of our daily lives, who sets the rules? In Europe, it’s the EU AI Act, the first comprehensive legal framework for artificial intelligence in the world. It’s a landmark regulation designed to stimulate tech innovation while still protecting fundamental rights, but how do you regulate something that evolves almost daily? 

The EU’s answer is a unique risk-based approach, a tiered system that classifies AI systems based on their potential to cause harm.

At the top are unacceptable risk AI systems, which are outright banned. These include AI used for “social scoring” by governments, like systems that grant or deny public services based on behavior. Real-time biometric identification in public spaces is also mostly prohibited, with only a few exceptions for serious crimes. 

Next are high-risk AI systems. These are allowed but come with many obligations. This category includes AI used in critical infrastructure, like managing electricity grids, or in essential public services, such as evaluating creditworthiness. AI in employment also falls here. These systems must meet requirements for data quality, accuracy, human oversight, transparency, and cyber security. Companies deploying high-risk AI will also need to conduct fundamental rights impact assessments to make sure they don’t harm vulnerable groups disproportionately. 

The numbers show why this is critical. Over 70% of companies are projected to integrate AI into their HR processes by 2027, for resume screening, performance evaluations etc. Furthermore, the adoption of AI in critical infrastructure is expected to surge by over 27% in the next five years.  

Below that are limited risk systems, which mostly involve transparency obligations. For example, chat bots and deep fakes must clearly state that they are AI generated. Finally, minimal risk AI, like spam filters or video games, faces very few restrictions, encouraging innovation in areas with little potential for harm. 

However, some argue that its complexity and compliance costs could inhibit European innovation, especially for small and medium sized enterprises. Others worry that the definitions of “high-risk” might not keep up with evolving technology, which would leave loopholes. The economic stakes are huge, with the global AI market projected to grow from $230 billion last year to over $1.7 trillion by 2032

The EU AI Act is a global experiment. Will this model inspire the rest of the world to provide a legal framework for artificial intelligence – remains to be seen. 

Watch the Video:

Share

Sign up for our newsletter

Explore More

Brussels, 1 December 2025 – B&K Agency is pleased to announce that as of 1 December, Heath Sloane joins the B&K Agency team as Director of Geopolitical Intelligence to kickstart geopolitical intelligence practice. In this role, he will be responsible for developing research and analysis on global security dynamics,

Read more

18 November 2025 / Brussels, Belgium – B&K Agency facilitated the policy roundtable ‘Protecting Transatlantic Innovation’ hosted by the Tholos Foundation and MEP Tomáš Zdechovský (EPP, Czechia) in the European Parliament. The event featured a discussion exploring how a stronger transatlantic partnership can protect digital innovation, enhance the competitiveness

Read more

2026 will not reward those who wait and see. PR and advocacy are shifting fast — from AI-driven storytelling and micro-targeted messaging to new rules in public trust, crisis response, and political communication. The campaigns that succeed next year will be the ones that understand the new playbook. Join

Read more