TWINLADDER
TwinLadder logoTwinLadder
CC BY-SA 4.0 Open Standard

The Twin Ladder Standard

From Compliance to Competence

Article 4 of the EU AI Act requires every organisation using AI to ensure sufficient AI literacy among its staff. The Twin Ladder Standard is the open methodology that measures whether you have achieved it — and shows you how to get there.

v1.0.0Published March 2026

What is Article 4?

The EU AI Act is Europe’s landmark regulation for artificial intelligence. Most of the Act focuses on high-risk AI systems — medical devices, recruitment tools, credit scoring. But Article 4 is different. It applies to everyone.

What Article 4 actually says

  • Every organisation that deploys or provides AI systems must ensure that its staff has a sufficient level of AI literacy.
  • AI literacy means understanding what AI can and cannot do, being aware of the risks, and knowing how to use AI tools responsibly.
  • This requirement is proportionate — it takes into account the technical knowledge, experience, and role of each person. A software engineer and a marketing manager need different levels of understanding.
  • It has been enforceable since 2 February 2025. Penalties: up to €15 million or 3% of global annual turnover, whichever is higher.

Who does it cover?

  • Not just regulated industries — every European organisation that uses AI in any capacity.
  • Not just IT departments — every person who interacts with AI tools in their work.
  • Not just high-risk applications — all AI use, from drafting emails with ChatGPT to screening CVs with automated software.
  • If your team uses ChatGPT, Copilot, AI-powered recruitment tools, contract review software, or any product with AI embedded — Article 4 applies to your organisation.

Why it matters now

Article 4 is already enforceable, but most organisations have no way to measure whether they comply. There is no official EU checklist, no ISO standard for AI literacy, and no agreed definition of “sufficient.” That is the gap the Twin Ladder Standard fills.

What solutions exist — and what is missing

Dozens of AI frameworks exist. Most of them solve a different problem.

The governance layer is well served

ISO 42001 covers AI management systems. The NIST AI Risk Management Framework addresses risk identification and mitigation. EU AI Act compliance checklists help organisations map their obligations. These frameworks are valuable — they tell you whether you have the right policies, processes, and documentation.

The competence layer is not

  • Governance tells you whether you have an AI policy. It does not tell you whether your people understand it.
  • You can pass an ISO 42001 audit and still have a workforce that cannot explain what a hallucination is.
  • You can have a perfect AI acceptable use policy and still fail Article 4 if no one has been trained to follow it.
  • Article 4 specifically requires competence — literacy, understanding, capability. Not just documentation.

How Twin Ladder is different

We measure competence, not just governance

Our six pillars map directly to what Article 4 requires: deployment competence, training, evidence of capability — not just policies on paper.

A clear compliance floor

Score 52 or above and you have a defensible position. Below that, you have measurable gaps to close. No ambiguity.

Open methodology, proprietary platform

The standard is published under CC BY-SA 4.0 — free to use, adapt, and redistribute. Think TCP/IP: the protocol is open, the services built on it are commercial. Anyone can adopt the methodology; Twin Ladder provides the best implementation.

Risk-calibrated and evidence-gated

A pharmaceutical company and a design studio face proportionate standards. And you prove compliance with evidence, not self-declarations.

Six pillars of AI competence

Each pillar answers a specific question about your organisation’s readiness. Together, they cover everything Article 4 requires.

Swipe horizontally to browse all pillars

Maturity Levels

Level 0

Exploring

0–25

Staff have heard of AI but cannot articulate capabilities or risks. No formal awareness activities.

Level 1

Developing

26–50

Leadership aware of AI obligations. Most staff have vague understanding of AI but cannot name specific risks.

Level 2

Implementing

51–75

Organisation-wide AI briefings completed. Staff can identify AI systems they use and describe key risks.

Level 3

Optimising

76–100

Continuous awareness programme. Staff proactively identify emerging AI risks. Context-specific understanding for all roles.

Maturity Levels

Level 0

Exploring

0–25

No AI use policy exists. Data protection in AI contexts not addressed.

Level 1

Developing

26–50

AI use policy drafted but not enforced. Informal guidance on acceptable use. GDPR acknowledged but not integrated.

Level 2

Implementing

51–75

Active AI use policy with defined acceptable and prohibited uses. GDPR compliance integrated into AI governance. DPIAs conducted for high-risk tools.

Level 3

Optimising

76–100

Comprehensive, regularly reviewed AI policy. Privacy-by-design principles embedded. Cross-regulatory compliance framework (AI Act + GDPR) fully operational.

Maturity Levels

Level 0

Exploring

0–25

No structured AI training. Staff learn informally or not at all.

Level 1

Developing

26–50

Generic AI training available. Self-directed learning. No role differentiation or completion tracking.

Level 2

Implementing

51–75

Role-specific training programme delivered to all AI-interacting staff including contractors. Completion tracked. Refresh cycles in place.

Level 3

Optimising

76–100

Personalised learning paths. Competence verified through scenario-based assessments. Continuous development culture. Third-party literacy verified contractually.

Maturity Levels

Level 0

Exploring

0–25

AI tools used ad hoc. No inventory of AI systems. Shadow AI prevalent.

Level 1

Developing

26–50

Partial AI inventory. Some tools assessed. Human review optional. Access controls informal.

Level 2

Implementing

51–75

Complete AI systems inventory with risk classification. Human oversight for consequential decisions. Verification protocols in place.

Level 3

Optimising

76–100

Automated AI systems monitoring. Continuous verification. Tool governance integrated into procurement. Shadow AI effectively eliminated.

Maturity Levels

Level 0

Exploring

0–25

No documentation of AI competence efforts. No audit trail.

Level 1

Developing

26–50

Some records exist. Informal documentation. Could not survive regulatory audit.

Level 2

Implementing

51–75

Centralised training records and evidence portfolio. Needs assessment documented. Can demonstrate effort under audit.

Level 3

Optimising

76–100

Comprehensive evidence framework. Automated compliance dashboards. Proportionality reasoning documented. Benchmark-ready data.

Maturity Levels

Level 0

Exploring

0–25

No AI governance structure. No designated responsible person.

Level 1

Developing

26–50

Informal responsibility. No dedicated oversight. Ad-hoc reviews when issues arise.

Level 2

Implementing

51–75

Named AI governance owner. Periodic review cycle. Ethics considerations documented. Incident response procedures exist.

Level 3

Optimising

76–100

Board-level AI oversight. Cross-functional governance committee. Continuous regulatory monitoring. Proactive risk anticipation.

Four maturity levels

Every organisation starts somewhere. The four levels describe a progression from no formal AI awareness to embedded, continuously improving AI competence. The compliance floor — the minimum Article 4 expects — sits at the boundary between Level 1 and Level 2.

Level 0

Exploring

Score: 0–25

No formal AI awareness programme. AI tools adopted ad hoc by individuals. No usage policy exists. Staff cannot articulate what AI tools they use or the risks involved. High likelihood of non-compliance.

Level 1

Developing

Score: 26–50

Some awareness training delivered. An AI acceptable use policy drafted but not yet consistently enforced. A tool inventory started but incomplete. Governance gaps remain. Working toward compliance but not yet there.

Level 2

Implementing

Score: 51–75

Structured training programme in place, tailored to roles. AI policy enforced organisation-wide. Evidence of competence documented and auditable. This is the compliance floor — the minimum standard Article 4 expects.

Level 3

Optimising

Score: 76–100

Continuous improvement embedded. External benchmarking against industry peers. AI governance integrated into business processes. Competence treated as a competitive advantage, not just a compliance obligation.

The TwinLadder Framework: From Compliance to CompetenceFour levels mapping organisational AI maturityCompliance Floor0AI LiteracyMandated (Article 4)Staff understand AI fundamentals, risks, limitations. Documented training.In force since Feb 2, 20251Professional TwinImplied ("context" requirement)Professionals use AI in their domain with verification protocols and judgment preservation.Required for effective compliance2Operational TwinAnticipatoryAI integrated into workflows with governance, measurement, and competence maintenance.Strategic advantage3Ecosystem TwinBeyond regulationOrganisation builds AI ecosystem with clients, partners, industry. Innovation through connected competence.Market leadershipCompetence MissionArticle 4 Requirement: In force since Feb 2, 2025. Levels 1–3 build what Article 4 cannot mandate — but the market demands.
Click to expand

Article 4 sets the floor. Competitive advantage requires going further.

© TwinLadder 2026 · CC-BY-SA 4.0

Article 4 — Mapped to the Six Pillars

Every operative phrase in Article 4 maps to one or more of the six pillars. This interactive mapping shows exactly how the standard translates regulatory language into measurable competence.

Providers and deployers of AI systems

PS

shall take measures

SP

to ensure, to their best extent

SPS

a sufficient level of AI literacy

PS

of their staff and other persons dealing with the operation and use of AI systems on their behalf

SPS

taking into account their technical knowledge, experience, education and training

SP

the context in which the AI systems are to be used, and considering the persons or groups of persons on whom the AI systems are to be used

PSSS

The compliance floor: score 52

Based on a line-by-line mapping of Article 4 to our six pillars, the minimum score for a defensible compliance position is approximately 52 — the transition point from Developing to Implementing. Scoring 52 means you can demonstrate that your organisation has taken reasonable measures. Scoring below it means you have identifiable, measurable gaps.

  • All six pillars must score above zero — a single zero-score pillar means a fundamental gap in compliance
  • Policy & Data Protection and Training carry the highest compliance weight — these are where regulators will look first
  • The floor is not a ceiling — organisations scoring 52 are compliant but fragile. A single staff change or new tool deployment could drop you below
  • Enforcement is already active (since February 2025) with penalties up to €15M or 3% of global turnover

How to use the standard

The Twin Ladder Standard supports a complete journey from measuring where you are to proving where you need to be.

1

Assess

Take the AI-powered conversational assessment to get your baseline scores across all six pillars. The free Quick Scan takes 15 minutes. The Executive Report provides a detailed gap analysis with prioritised recommendations.

Start assessment
2

Learn

Enrol in Twin Ladder Academy courses mapped to your weakest pillars. Foundation, Leadership, and Mastery pathways cover everything from Article 4 basics to cross-functional AI governance.

Explore courses
3

Certify

Reassess to measure your progress. Build an evidence portfolio that documents training completed, policies adopted, and competence achieved. Work toward Twin Ladder Certified status.

View pricing

Open Methodology

CC-BY-SA 4.0

The Twin Ladder Standard is published under Creative Commons Attribution-ShareAlike 4.0 International. The methodology is open — anyone can use, adapt, and redistribute it with attribution. The platform, assessment tools, and certification programme are proprietary. The standard is free. The implementation is ours.

Adopt the standard

The Twin Ladder Standard is free to use, adapt, and redistribute under CC BY-SA 4.0. View the full methodology on GitHub, or assess your organisation now to find out where you stand.