In 2017, the European Council emphasized the need to address artificial intelligence (AI) trends while maintaining high standards for data protection and ethics. By 2023, concerns about ChatGPT's misuse led to Italy's temporary ban. The European Union (EU) is now working on the AI Act to create ethical AI guidelines.

What is the EU AI Act?


Artificial Intelligence=
The EU AI Act aims to create a unified legal landscape for AI development and deployment across member states. By establishing consistent rules, the regulation seeks to prevent a fragmented market while fostering innovation and ensuring a high level of protection for citizens.

It promotes the uptake of human centric and trustworthy AI while ensuring high level of health, safety, fundamental rights, including democracy and rule of law, the environmental protection against the harmful effects of AI, and to support innovation. It also ensures the free movement of AI and prevents the Member States from imposing restrictions on the development, marketing and use of AI unless explicitly authorized by this regulation. Its aim is to foster trustworthy AI in Europe and beyond, and to ensure that AI systems respect fundamental human rights, safety and ethical principles.
Cost of WCC

What are the risks of AI?

While AI’s ultimate goal is to increase human well-being, the very components driving the socio-economic advantages of AI can also introduce new risks or negative outcomes for individuals or society. Those risks can cause harm to public interest and fundamental rights that are protected by Union Law, which are: material or immaterial, including physical, psychological, societal or economic harm.
Fire Wall AI

The act offers a fair and measured regulatory framework for AI that takes a horizontal approach, focusing solely on essential requirements to tackle associated risks and issues without overly restricting technological progress or unreasonably inflating the costs of bringing AI solutions to market.

The legislation uses a risk-based approach (RBA) by categorizing risks in four categories, with each level requiring a different degree of regulation:
  • 1

    Minimal risk: Such systems will not be subject to EU regulations.

  • 2

    Limited risk: Such systems will need to focus on transparency with their customers in effort to comply with the legislation.

  • 3

    High risk: Such systems must undergo rigorous risk assessment.

  • 4

    Unacceptable risk: Such systems will be banned in the EU as they pose a threat to human rights.


  • Conformity assessments will be conducted on high-risk AI systems to evaluate their compliance with the regulation, along with an emergency procedure, "allowing law enforcement agencies to deploy a high-risk AI tool that has not passed the conformity assessment procedure in case of urgency". High-risk AI systems will need to adhere to stringent mandates, encompassing risk mitigation mechanisms, high-quality datasets, activity logging, comprehensive documentation, transparent user communication, human supervision, and elevated levels of robustness, accuracy, and cybersecurity. Regulatory sandboxes will foster responsible innovation and the creation of AI systems that meet regulatory standards.
    Risks of AI


    Will innovation face difficulties?

    While these requirements may initially seem to hinder innovation, they could ultimately foster a more sustainable and trustworthy AI ecosystem in finance. The clear regulatory framework provides certainty for firms investing in AI development, potentially encouraging more widespread adoption of AI technologies.
    Data
    The regulation's emphasis on data quality and bias mitigation could lead to more reliable and fair AI systems, enhancing customer trust and potentially opening new market opportunities. Moreover, the focus on explainability and transparency may drive innovations in interpretable AI, an area of growing importance in finance.

    What will be the main challenges for financial services?

    • Compliance burden: The regulation categorizes AI applications based on their on a risk-based approach. High-risk applications, such as AI-powered credit scoring or fraud detection systems, will face stricter compliance requirements. Financial institutions, fintechs and regtech companies will need to invest in resources and expertise to ensure their AI systems comply with these regulations by developing more robust documentation processes and by training its staff on AI oversight.

    • Data governance: The Act emphasizes responsible data use and prohibits discriminatory AI systems. Companies will need to scrutinize their data collection practices and ensure that AI models are trained on unbiased data sets to avoid perpetuating biases in financial decision-making.

    • Explainability and transparency: The "black box" nature of some AI models can be problematic. The act emphasizes the need for explainable AI, requiring the different companies to understand how their AI systems arrive at decisions and be able to explain them to regulators and consumers. For multinational firms, the challenge of complying with the act alongside other regional regulations (such as those in the US or Asia) may create additional complexity.

    • While it undoubtedly introduces regulatory challenges for financial services, it also presents significant opportunities. By fostering a robust framework for responsible AI development, the Act can serve as a catalyst for innovation. Enhanced risk management, sophisticated fraud detection systems, and personalized customer experiences are just some of the potential benefits. As firms navigate this new landscape, a strategic approach focused on compliance, data governance, and explainability will be crucial. Ultimately, the EU AI Act positions the financial sector to harness the power of AI while safeguarding consumer interests and maintaining trust.

      What happens in case of noncompliance?

      Firms failing to adhere to the regulations will face penalties. These penalties will vary. For breaches involving prohibited AI applications, fines could amount to €35 million or 7% of the company's global annual turnover (whichever is greater), while breaches of other obligations could result in fines of €15 million or 3%. Supplying inaccurate information may lead to fines of €7.5 million or 1%. The AI Act includes more proportional limits on administrative fines for small and medium-sized enterprises (SMEs) and startup in case of violations.
      Penalties

      How can financial institutions turn AI regulation into a competitive advantage?

      Despite the compliance burden, the EU AI Act presents opportunities for forward-thinking financial institutions. Firms that embrace the principles of ethical and trustworthy AI may gain a competitive edge, both in customer trust and regulatory compliance. The regulation's support for AI's innovation hubs and regulatory sandboxes could provide valuable testing grounds for new AI-driven financial products and services. Additionally, as the EU positions itself as a leader in trustworthy AI, EU-based financial firms adhering to these high standards may find new opportunities in the global market.

      What's next for financial services in the AI era?

      The EU AI Act marks a pivotal moment for the financial services industry. As firms grapple with the complexities of this new regulatory landscape, a strategic approach is essential. Conducting a comprehensive inventory of AI systems and assessing their risk profiles is the first step towards compliance. Building robust AI governance structures, coupled with investments in explainable AI and data quality, will be crucial for mitigating risks and fostering trust.
      Secure
      Active engagement with regulators and industry bodies is paramount. By participating in shaping the implementation of the Act, financial institutions can influence the regulatory environment and avoid potential pitfalls. Moreover, embracing ethical AI principles can position firms as industry leaders, enhancing their reputation and attracting customers who value responsible business practices.

      The EU AI Act represents a significant shift in the regulatory landscape for AI in financial services. While it presents compliance challenges, it also offers a framework for building more trustworthy, robust AI systems. Financial institutions that successfully navigate this new environment may find themselves well-positioned for sustainable growth in an AI-driven future.
0 comments
Add your comment

Related articles

An AML EWRA overall risk assessment allows financial institutions to identify and appropriately manage the ML / FT risks...

Risk Based Approach Tue 18 June 2019

Explore the list of GDPR enforcement actions, fines, and penalties against companies, institutions, and organization gro...

EU Tue 19 February 2019

Conducting KYC on corporate entities requires knowledge on the structure of companies. Learn what essential documents ar...

Risk Based Approach Thu 08 August 2024

Hiring a consultant for your financial institution? Avoid mistakes and learn what you need to know before you choose to ...

FinTech Thu 08 November 2018
Experts in risk management and regulatory compliance

Pideeco is a consultancy firm providing legal services, business solutions, operational assistance and educational material for professionals in the financial industry.

We are based in Brussels and we specialize in regulatory risk compliance services covering the Eurozone.

Pideeco combines professional Regulatory knowledge and technical expertise to safeguard your business’ reputational and operational risk. Our unique customer-centric approach helps us build strategical and legitimate cost-efficient remedies.

Working with us means reaching out to complementary people, allowing for original thinking and innovative vision.

Our Network Learn more about us