The Evolution of Responsible AI

What “responsible” means in the context of AI has changed profoundly in recent years. Once a philosophical discussion about doing good, it has become a practical question of accountability and evaluation. In education, this shift is particularly urgent: by 2025, only 20–40% of institutions are estimated to have formal AI governance structures, while teachers and pupils are already using AI every day.

history responsible AI

Updated 5 October 2025 3-minute read

TL;DR Summary

AI in education is evolving from ethics to evaluation. Only about one in four institutions have governance frameworks, while AI use grows rapidly. The future lies in continuous learning, measurement, and improvement — with humans firmly in control.

Four Phases of Maturity

Table 1 shows the evolution of Responsible AI (RAI) mirrors a broader societal transition from ideals to evaluation.

Table 1. Four Phases of Maturity
Phase Period Core focus Main question
1. Awareness Up to 2018 Ethical reflection What values should guide AI?
2. Regulation 2018–2023 Legal compliance What must we do to comply with the law?
3. Governance Since 2023 Organisational structure Who decides, and who is accountable?
4. Evaluation Now Learning and improvement Does AI work as intended, and how can we improve it?

In 2025, around 80% of education institutions are still in the first two stages – revealing a clear implementation gap.

“AI maturity begins when schools not only design responsibly — but also learn responsibly.”

The Implementation Gap in Education

AI is already embedded in classrooms: around 65% of pupils use generative tools such as ChatGPT, while only about one in four schools have any formal governance framework in place. This gap creates real risks:

Examples are visible across the sector: automated proctoring misidentifying pupils with regional accents, admissions algorithms overvaluing incomplete datasets, or AI screening tools reducing staff diversity. The “trust gap” arises not from bad intent, but from the lack of structured evaluation and accountability.

From Ethics to Evaluation

Responsible AI has evolved from a moral compass to a governance and evaluation framework. Today, responsibility means being able to explain, justify, and adjust. Evaluation must take place on four levels:

  1. Legal: does the system comply with the GDPR and the EU AI Act?
  2. Ethical: does it promote fairness, inclusion, and autonomy?
  3. Educational: does it strengthen learning quality and pedagogy?
  4. Operational: is it transparent, controllable, and correctable?

Ethics remains the compass — but evaluation is the proof that it works.

Emerging Evaluation Frameworks

International initiatives are now helping schools to evaluate AI responsibly:

  • EU AI Act (2025–2026): defines obligations for high-risk AI systems, including human oversight.
  • European Education AI Governance Framework (2025): the first sector-specific model linking compliance, ethics, and educational impact.
  • World Economic Forum Playbook (2024): outlines nine practical actions for assessing and improving AI systems.
  • UNESCO Guidelines for AI in Education (2023): promote transparency, inclusion, and human-in-the-loop governance.

Together, these frameworks mark a shift from reactive compliance to proactive evaluation — designing AI with trust built in from the start.

Towards Responsible Autonomy

The next step in this evolution is responsible autonomy. AI may assist, but it should never decide autonomously. Evaluation must therefore be cyclical: Define > Apply > Assess > Improve.

Schools that embrace this cycle move from compliance to ownership. They cultivate a culture in which critical reflection, human control, and ethical practice become second nature.

“The future of AI in education is not fully automated — it is consciously governed.”
« More Responsible AI