Skip to main content

Command Palette

Search for a command to run...

AI Tools for HR at European SMEs: What Is Safe to Deploy in 2026

How HR leads at EU SMEs can deploy AI for hiring, onboarding, and reviews without triggering EU AI Act Annex III obligations.

Updated
8 min read
AI Tools for HR at European SMEs: What Is Safe to Deploy in 2026
D
PhD in Computational Linguistics. I build the operating systems for responsible AI. Founder of First AI Movers, helping companies move from "experimentation" to "governance and scale." Writing about the intersection of code, policy (EU AI Act), and automation.

TL;DR: How HR leads at EU SMEs can deploy AI for hiring, onboarding, and reviews without triggering EU AI Act Annex III obligations.

Most AI vendors selling HR tools will not tell you that their product may be classified as high-risk under the EU AI Act. Why this matters: the EU AI Act classifies automated or semi-automated employment decision systems as high-risk under Annex III, point 4. That classification does not apply to every HR tool, but the line between a compliant workflow and a notifiable high-risk system can be a single configuration decision. A 35-person Dutch professional services firm, for example, can use AI to draft job descriptions and generate structured interview question banks with minimal compliance overhead. The same firm using a CV ranking score as the primary screening filter is squarely in Annex III territory, with conformity assessment, mandatory human oversight, and candidate explanation rights all required.

This article gives HR leads and operations leaders at growing mid-sized companies a practical map of which HR AI use cases are safe, which trigger high-risk obligations, and how to structure your tooling to stay on the right side of both the EU AI Act and GDPR Article 22. You will leave with a clear classification of three HR AI tool categories, the "assist not decide" principle that determines your compliance exposure, and a five-question checklist to run against any HR AI vendor before signing.


The Three HR AI Tool Categories

HR AI tools divide into three functional categories. Each carries a different compliance profile.

CV screening and ATS enrichment. Tools in this category parse CVs, flag keyword matches, score candidates against job criteria, or rank applicant pools. This is the highest-risk category under the EU AI Act. When a tool produces a ranking or score that filters candidates into or out of a shortlist without a human making an independent assessment, it is functioning as an automated decision-making system. Annex III, point 4 covers "AI systems intended to be used for recruitment or selection of natural persons, in particular for advertising vacancies, screening or filtering applications, evaluating candidates in the course of interviews or tests."

Onboarding automation. This category includes Q&A chatbots that answer new-hire questions, document generation tools that produce employment contracts or policy summaries from templates, and workflow automation that routes tasks to the right team. These tools do not make employment decisions. They assist HR staff with administrative work. Compliance exposure here sits primarily with GDPR data minimisation and processor agreements, not with the EU AI Act's high-risk framework.

Performance review assistance. Tools that help managers write structured review summaries, flag sentiment patterns in self-assessments, or generate suggested rating language fall into this category. The risk profile depends on how the output is used. A tool that drafts a summary for a manager to edit and approve is low-risk. A tool whose output feeds directly into a promotion or termination recommendation without independent manager review starts to resemble an automated employment decision system.


The EU AI Act Annex III Classification in Plain Terms

The EU AI Act's Annex III designates certain AI system categories as high-risk regardless of their technical design. Employment-related systems are in that list. Point 4 covers AI used in hiring, promotion, and task allocation decisions.

High-risk classification triggers four obligations for whoever deploys the system:

  • A conformity assessment must be completed before deployment.
  • Human oversight must be structurally implemented: a human must be able to intervene, override, or halt the system.
  • Candidates have a right to explanation when an AI system has influenced a decision about them.
  • The system must be registered in the EU database for high-risk AI systems.

These are not administrative formalities. They require documented processes, staff training, and vendor cooperation to obtain technical documentation. For most HR leads at a 30-person team, this compliance stack is disproportionate relative to the benefit of fully automated screening.


The "Assist Not Decide" Principle

The practical way to stay outside Annex III is to structure your HR AI tools so that they assist a human decision-maker rather than replace one. This reflects how the regulation draws the line.

An AI tool that generates a list of suggested interview questions based on a job description is assisting an HR lead. The HR lead selects which questions to use. No automated employment decision occurs.

An AI tool that scores CVs and presents a ranked shortlist where the hiring manager simply approves the top five is a different situation. The ranking score is doing the filtering work. The human is ratifying, not independently assessing. Regulators are likely to treat that as an automated decision in substance, regardless of the approval step.

The configuration rule for your people operations team: the AI output should require meaningful human judgment to act on, not just human confirmation of an AI output. Document how your HR staff use the tool's output. That documentation is your first line of defence in any supervisory inquiry.


GDPR Article 22 and Automated Hiring Decisions

Even before the EU AI Act came into full effect, GDPR Article 22 created obligations around automated decision-making in hiring. Article 22 gives individuals the right not to be subject to a decision based solely on automated processing if that decision produces a legal or similarly significant effect on them.

A hiring rejection based solely on a CV screening algorithm's output falls within scope. The legal basis options are narrow: explicit consent (difficult to obtain in a hiring context), necessity for entering a contract (limited application for pre-hire decisions), or a specific EU or Member State law authorising the processing.

In practice, this means any growing professional services firm using automated CV filtering without a documented human review stage is exposed under both GDPR Article 22 and, post-August 2026, the EU AI Act's Annex III regime. The two frameworks reinforce each other. A system that satisfies the EU AI Act's human oversight requirement for high-risk systems will typically also satisfy Article 22 requirements for human involvement in significant decisions.


Five Questions to Ask Any HR AI Vendor Before Signing

Use these questions before committing to any HR AI tool. The vendor's answers, and their willingness to answer in writing, tell you most of what you need to know about their compliance readiness.

  1. Does your system produce a ranking, score, or classification that filters candidates into or out of a shortlist? If yes, how is human oversight implemented in the workflow?

  2. Have you conducted a conformity assessment under the EU AI Act for use in employment-related decisions? Can you share the technical documentation and Declaration of Conformity?

  3. How does your system handle a candidate's right to explanation under the EU AI Act and GDPR Article 22? Is an explanation log generated automatically?

  4. What data is processed, where is it stored, and who acts as data processor under GDPR? Is a Data Processing Agreement included in the contract?

  5. If we configure the tool so that your AI output is advisory only and our HR team makes all final decisions, does the system architecture support and document that configuration?

A vendor who cannot answer questions two and three in writing is not ready for deployment in a compliant EU hiring workflow.


FAQ

Does the EU AI Act apply to our HR software if we are not the ones who built it? Yes. The EU AI Act distinguishes between providers (who build and place AI systems on the market) and deployers (who use those systems in their operations). As a deployer, you have obligations under Article 25, including implementing human oversight, following the provider's instructions for use, and monitoring the system for issues. You cannot transfer your deployer obligations to the vendor.

We only use AI to draft job descriptions and write interview questions. Do we need to worry about Annex III? No. Drafting job descriptions and generating interview question banks does not constitute an automated employment decision. These are content-generation tasks that assist HR staff. As long as the AI is not filtering, scoring, or ranking candidates, you are outside the Annex III high-risk classification.

What counts as "meaningful human oversight" for a CV screening tool? The oversight must be genuine, not performative. The EU AI Act requires that a human can understand the AI system's output, identify errors or bias, and override the system without the tool preventing or penalising that intervention. Document cases where the human reviewer diverges from the AI ranking. That record demonstrates active oversight.

Our HR team is using a tool we did not formally approve. What are the risks? Shadow AI in HR is a serious exposure point. If an employee is using an unapproved CV screening or performance analysis tool, your organisation may be operating a high-risk AI system with no conformity assessment, no documented oversight process, and no GDPR data processing agreement in place. A governance policy requiring all HR AI tools to be reviewed before use is the minimum control.


Further Reading

If you are not yet certain where your current HR tooling sits relative to EU AI Act obligations, the AI Readiness Assessment is a structured starting point.