Veriff
LibraryFraud centerFraud NewsNavigating AI compliance in 2025: Key insights from Veriff's webinar

Navigating AI compliance in 2025: Key insights from Veriff's webinar

AI is reshaping industries at an unprecedented pace, but with innovation comes greater regulatory challenges. How can businesses stay ahead of evolving compliance requirements like the EU AI Act? In Veriff’s latest webinar, top industry experts break down what you need to know about navigating the complex world of AI regulations and their global impact.

Header image
Author
Chris Hooper
Director of Brand at Veriff.com
February 19, 2025
Fraud
AI Compliance
Share:
On this page
Understanding the EU AI Act and its global impact
AI risk classification: What businesses need to know
Key dates to watch for AI compliance
Key compliance steps for businesses (EU AI Act & GDPR)
Final thoughts: Preparing for the future of AI compliance
Recommended AI guidelines for businesses

As artificial intelligence (AI) continues to transform industries worldwide, businesses face increasing regulatory scrutiny. In Veriff’s latest webinar, industry experts Chris Hooper (Director of Brand, Veriff), Tea Park (Attorney at Law, NJORD Law Firm), and Jeremy Joly (Co-founder and CPO of Spektr) discussed the evolving compliance landscape, focusing on the EU AI Act and its global implications.

This article summarizes the key takeaways to help businesses navigate AI compliance in 2025, with practical guidance on meeting EU AI Act requirements, GDPR alignment, and high-risk AI system obligations.

Understanding the EU AI Act and its global impact

The EU AI Act is one of the most significant regulatory frameworks governing AI technologies. With a broad scope, it applies to:

  • Companies based in the European Union
  • Any organization providing AI-driven services or products to EU users

Key definitions under the Act

  • AI System – A machine-based system designed to operate with varying levels of autonomy and adaptiveness, using inputs to generate outputs such as predictions, content, recommendations, or decisions that influence physical or virtual environments.
  • Provider – The entity developing an AI system.
  • Deployer – An organization implementing an AI system in its operations.
  • Importer – A company bringing AI solutions developed outside the EU into the region.
  • Distributor – A business that resells or facilitates AI solutions in the market.

Understanding where your business fits within this framework is essential for compliance.

AI risk classification: What businesses need to know

One of the most critical aspects of the EU AI Act is its risk-based approach, which classifies AI systems into four categories:

Unacceptable risk (strictly prohibited)

These AI applications are banned due to their potential harm to individuals and society. Examples include:

  • Social scoring systems
  • AI used for predictive policing
  • Real-time remote biometric identification (except in law enforcement with strict safeguards)

High-risk AI systems (extensive compliance obligations)

Businesses must implement stringent documentation, risk management, and human oversight for high-risk AI systems, including applications in:

  • Healthcare (e.g., AI-assisted diagnosis)
  • Finance (e.g., credit scoring, fraud detection)
  • Education and employment (e.g., AI-driven hiring assessments)
  • Limited-risk AI systems (transparency obligations)
  • AI applications in this category require transparency measures, such as user notifications when interacting with AI-generated content. Examples:
  • Chatbots
  • AI-generated media and deepfakes

Limited-risk AI systems (transparency obligations)

AI applications in this category require transparency measures, such as user notifications when interacting with AI-generated content. Examples:

  • Chatbots
  • AI-generated media and deepfakes

Minimal risk AI systems (best practice recommendations)

These AI applications have no major compliance obligations but are encouraged to follow best practices for responsible AI use.

Key dates to watch for AI compliance

  • February 2, 2025 – General EU AI Act compliance deadline
  • August 2, 2025 – Implementation of high-risk AI system obligations
  • August 2, 2026 – Stricter enforcement of compliance requirements
  • August 2, 2027 – Full-scale regulatory oversight comes into effect

Key compliance steps for businesses (EU AI Act & GDPR)

Conduct an AI audit

Start by identifying all AI-driven processes within your organization and classifying them according to the EU AI Act’s risk categories: Unacceptable, High-Risk, Limited, or Minimal. Once categorized, assess any compliance gaps to ensure you meet the necessary regulations without risking over-compliance or falling short of requirements.

Align with GDPR and data protection regulations

AI systems frequently handle personal data, making GDPR compliance essential. To ensure this, organizations should adopt privacy-enhancing techniques such as data anonymization and minimization to reduce exposure to personal information, synthetic data generation to train AI models without using real user data, and encryption alongside secure data handling to protect sensitive information. Additionally, adherence to lawful processing principles, including transparency and respecting data subject rights, is vital for maintaining compliance and building trust.

Establish human oversight mechanisms

The EU AI Act requires “meaningful human intervention” in AI-driven decisions, emphasizing the importance of maintaining human oversight in high-risk applications. To comply, organizations are encouraged to appoint AI compliance officers or establish AI governance teams to oversee implementation and adherence to regulations. Additionally, it is crucial to ensure that humans retain the final decision-making authority in scenarios where AI poses significant risks. Regular testing of AI models for bias, fairness, and accountability is also essential to align with the Act's standards and promote ethical AI practices.

Strengthen third-party AI vendor compliance

If your AI solutions depend on external vendors or APIs, make sure contracts clearly outline compliance responsibilities, verify that AI providers adhere to the EU AI Act and GDPR standards, and conduct regular audits and impact assessments of third-party systems.

Stay updated on AI regulatory changes

AI regulations are changing quickly, and businesses need to stay ahead. It's important to track updates to the EU AI Act and other regional policies, engage in industry discussions and regulatory forums, and participate in AI ethics groups. As new rules are introduced, companies should adapt their AI governance frameworks to remain compliant.

AI Compliance Guide – Webinar

Stay ahead of AI regulations with our webinar covering compliance frameworks, updates, and best practices—watch on-demand now!

Final thoughts: Preparing for the future of AI compliance

As AI regulations take shape, businesses must proactively prepare for compliance, rather than react to penalties. With strategic planning, clear documentation, and a focus on ethical AI deployment, organizations can stay ahead of regulatory challenges while leveraging AI for innovation.

Veriff’s webinar emphasized the importance of transparency, accountability, and adaptability in AI governance. By following these insights, businesses can ensure compliance and build trust with customers and regulators alike.

Recommended AI guidelines for businesses: