If your employees use ChatGPT for writing emails, Gemini for data analysis, or Copilot in Excel - you have been subject to the AI Act since February 2025. It does not matter whether you have 3 people in your company or 300.
Most small business owners across Europe are unaware of this obligation. Yet enforcement of fines starts in August 2026. This article explains what you need to do, how much time you have, and how to prepare - without legal jargon.
Note: This article is for informational purposes only and does not constitute legal advice. For specific legal questions, consult a lawyer specialising in technology law.
What is the AI Act and why it applies to you
The AI Act (EU Regulation 2024/1689) is the world's first comprehensive law regulating artificial intelligence. It was passed by the European Parliament in March 2024 and applies across all EU member states.
The regulation does not only concern companies that build AI systems. It applies to any entity that uses AI. If your employee uses an AI tool at work - even the free version of ChatGPT - your company is a "deployer" under the AI Act.
That means concrete obligations. The most important one for small businesses is Art. 4 - the requirement to ensure AI competency among people who use these tools.
Art. 4 - the AI Literacy obligation
Article 4 of the AI Act is clear: providers and deployers of AI systems must ensure a "sufficient level of AI literacy" among people who operate or use those systems.
In practice, this means:
- You need to know which AI tools are used in your business - including those employees installed on their own (Shadow AI).
- Everyone using AI should understand how these tools work, their limitations, and potential risks.
- You need to document this - saying "they know" is not enough. You need evidence: training records, materials, attendance lists.
Key point: AI Literacy does not mean every employee must become an AI expert. It means that people using AI at work understand what they are doing, what the tool's limitations are, and when a result needs human verification.
Art. 4 does not require trainer accreditation - it requires documented team competency and training quality. What counts is practical knowledge, complete documentation, and genuine preparation of staff to work with AI.
What you specifically need to do
Preparing for Art. 4 of the AI Act comes down to three steps.
Step 1: Audit of AI tools in your business
Create a list of every AI tool your business uses. Not just the official ones - check what employees have installed on their own too. Common places where AI "hides":
- ChatGPT, Claude, Gemini - for writing emails, proposals, content
- Microsoft Copilot - built into Office 365
- Canva AI, Adobe Firefly - for graphics
- Grammarly, DeepL - for translations and proofreading
- CRM tools with AI features (e.g. HubSpot, Salesforce Einstein)
- Browser extensions with AI
For each tool, record: who uses it, what for, and what data it processes. This will be the basis for scoping your training. The tool audit is also a good moment to plan AI process automation -- which of these tools can be combined into a coherent workflow.
Step 2: Employee training
Training should cover:
- How AI works - what a language model is, why it "hallucinates", what a prompt is
- Tool limitations - when AI gets things wrong, what data can be entered and what cannot
- Data security - what happens to data entered into ChatGPT, the difference between the free and paid versions
- Practical use - how to prompt effectively, how to verify outputs, when not to trust AI
- Legal context - key facts about the AI Act and company obligations
Training does not need to last a week. For most small businesses, a one-day workshop tailored to your sector and the tools you actually use is enough. Also check whether you qualify for AI training funding from KFS or BUR -- it covers up to 80% of costs.
Step 3: Documentation
After training you should have:
- Training plan - a programme tailored to roles in the business (different for sales, different for accounting)
- Training materials - presentation, exercises, checklists
- Attendance records - who participated and when
- Completion certificates - for each participant
- Post-training report - findings, recommendations, follow-up plan
This documentation is your proof of compliance with Art. 4. In the event of an inspection, you present specific files - not promises.
Implementation timeline
The AI Act does not take effect all at once. Different provisions have different deadlines:
-
February 2, 2025
Art. 4 - AI Literacy. Obligation to ensure AI competency. Formally in force from this date.
-
August 2, 2025
Prohibited AI practices. Ban on social scoring systems, subliminal manipulation, and emotion recognition in the workplace.
-
August 2, 2026
Enforcement of fines. Supervisory authorities may impose financial penalties for non-compliance. This is the deadline you need to prepare for.
-
August 2, 2027
High-risk AI systems. Full requirements for AI systems in recruitment, lending, and healthcare.
Practical note: You have until August 2026 to have your documentation in order. That is less than 5 months from the publication date of this article. Do not leave it to the last minute - preparing the audit, training, and documentation takes several weeks.
Fines for non-compliance
The AI Act provides for three tiers of fines:
- Prohibited AI practices: up to 35 million euros or 7% of annual turnover
- High-risk AI systems and Art. 4 (AI Literacy): up to 15 million euros or 3% of annual turnover
- Providing false information to supervisory authorities: up to 7.5 million euros or 1% of annual turnover
For a small business these figures may seem abstract. But the regulation refers to "proportionate penalties" - meaning the supervisory authority will consider the size of the business, the severity of the breach, and whether you took any corrective action.
The worst position is doing nothing. A business that can show a training plan, documentation, and completion certificates - even if imperfect - is in a far better position than one that was unaware of the obligation.
ChatGPT and GDPR - is your business breaking the law
Pasting personal data (customer names, tax IDs, HR records, email contents) into the consumer version of ChatGPT violates GDPR. ChatGPT, Gemini, and Claude in their free tiers use your prompts to train models, which qualifies as transferring data outside the EEA without legal basis. Solution: "Team/Enterprise" tiers disable training, or choose tools with explicit "data privacy" mode.
The most common mistake by a business owner: an employee pastes a customer email into ChatGPT and asks "rewrite this in a friendlier tone". The email contains a name, email address, sometimes a phone number. That is personal data under GDPR. ChatGPT (Free, consumer Plus) stores it and may use it to train the model. This violates Art. 6 GDPR (no legal basis for processing) and Art. 44-49 (transfer of data outside the European Economic Area without safeguards).
3 most common ChatGPT data leaks in small businesses
- Customer lists for analysis: "Help me group these customers" + a spreadsheet with names, phone numbers, addresses. The entire spreadsheet goes into model training.
- Business email content: "Write a reply to this email" + the pasted message with sender data. Name, email, signature with company - all goes to OpenAI.
- HR documents: "Summarize this employment contract". Name, ID number, employee address - the whole document in the prompt, the whole document in training.
Scale of the problem: a 2024 Cyberhaven study found that 11% of data pasted into ChatGPT by employees is sensitive (customers, finance, HR). In a typical 30-person business that means several leaks per week - completely unintentional.
How to use ChatGPT in a GDPR-compliant way
- Choose a business tier: ChatGPT Team/Enterprise, Claude Team, Gemini Business. These tiers disable training on your data by default and include a Data Processing Agreement (DPA). Cost: from approximately 25 EUR/user/month.
- Create an AI policy for employees: one A4 page. What is allowed (anonymous content, code, concepts). What is not (customer personal data, HR records, financial data, passwords). Each employee signs that they have read the policy.
- Run training on safe AI usage: 90% of leaks are not hacker attacks - they are employee mistakes. Training + checklist + concrete "yes/no" examples = 80% risk reduction.
GDPR and AI Act compliance go hand in hand. Art. 4 of the AI Act requires training employees on AI use - the same training covers GDPR concerns. See our AI training for small businesses - it covers both regulations in a single workshop.
How 30Elevate can help
We combine practical AI knowledge with the realities of small business. Our AI training covers both practical skills and the documentation required under Art. 4 of the AI Act.
What you get after the workshop:
- Training plan tailored to your industry and team roles
- Training materials (presentation + checklists)
- Hands-on exercises with the AI tools you actually use
- Attendance records and completion certificates for each participant
- Post-training report with recommendations
One workshop - and your team's competency is raised and your documentation is in order. We do not promise an "AI Act compliance certificate" (because no such formal document exists), but we give you everything Art. 4 requires.
The trainer holds Google AI certifications and has hands-on experience deploying AI systems in businesses. The training is delivered in language that works for a business owner - not a developer.
Frequently asked questions
Does the AI Act apply to my small business?
Yes. If your employees use any AI tools - even ChatGPT for writing emails - Art. 4 of the AI Act requires you to ensure adequate competency. Business size and sector do not matter.
What fines apply for lack of AI Literacy compliance?
Violations of Art. 4 of the AI Act carry fines of up to 15 million euros or 3% of annual turnover (whichever is higher). For small businesses the amounts are proportionally lower, but still significant. Fines can be imposed from August 2026.
Do I need an AI Act certificate?
A formal "AI Act compliance certificate" does not exist. Art. 4 requires documented competency - a training plan, materials, attendance records, and a completion certificate. What matters is the trainer's expertise and the quality of documentation.
When does AI Act enforcement begin in the EU?
Art. 4 (AI Literacy) has formally applied since February 2, 2025. Enforcement of fines begins in August 2026. This gives businesses time to prepare, but it is best to start now.
What is the AI Literacy obligation under Art. 4?
Art. 4 requires that every person operating an AI system in a company has a sufficient level of competency. This includes understanding how AI tools work, recognizing limitations and risks, and being able to critically evaluate AI-generated outputs. The obligation applies to both employees and the business owner.
How do I conduct an AI tools audit in my company?
Create a list of all AI tools used in your company - from ChatGPT to Canva AI to CRM automations. For each tool, record: who uses it, what for, what data it processes, and what risk category it falls under according to the AI Act. Most small business tools fall into the minimal or limited category.
What AI documentation must a business maintain?
Art. 4 requires documentation of: a list of AI tools used, a training plan with scope and timeline, training materials, attendance records, and completion certificates. A simple folder with these elements is sufficient in case of an inspection.
Get your business ready for the AI Act
AI Literacy workshop for your team - practical skills and complete documentation for Art. 4 of the AI Act. See training details or get in touch.
Book a workshop