Education technology has long promised transformation and mostly delivered mediocrity. The sector is littered with failed initiatives — interactive whiteboards, tablet-based learning platforms, learning management systems that teachers resent using. Most failed not because the technology was broken, but because it was grafted onto educational models that did not change fundamentally. Expensive software replacing pencils and paper is not transformation.
AI in education presents a genuinely different opportunity because it addresses a fundamental structural problem in education that has no solution without computational capability: personalisation at scale. The ability to tailor instruction to the learning pace, style, and needs of individual students is the gold standard of education. It is also unachievably expensive without technology. A teacher cannot personalise instruction for thirty students simultaneously without external support. An AI system can.

But the opportunity is real only if education technology companies approach it differently than previous waves have. That means understanding educational outcomes, not just technology adoption. Understanding learner psychology, not just feature development. Understanding the legitimate concerns about AI in education, and building systems that address those concerns rather than dismissing them.
Where AI Genuinely Transforms Learning
The most promising AI education applications are those that address specific, measurable problems with clear causal mechanisms.
Intelligent tutoring systems represent the most mature category. These systems adapt their instruction to match the learner’s understanding in real time, providing targeted feedback and adjusting difficulty based on performance. The research on intelligent tutoring systems is consistent: they improve learning outcomes, particularly for students at the extremes — those struggling with foundational concepts and those ready for advanced material. They also provide teachers with real-time insight into student understanding that would be impossible to gather manually.
Automated assessment and feedback is another strong category. AI systems can grade essays, evaluate problem-solving approaches, and provide detailed feedback at a scale that no human teacher could achieve. More importantly, they can do so consistently and without the unconscious bias that influences human grading. The companies building this capability are solving a real problem — teachers spending hours on grading work that could be marked instantly, freeing time for actual teaching.
Adaptive learning pathways represent a third opportunity. By tracking what a student has learned, where they struggled, and how they learn best, AI systems can recommend the optimal sequence of learning activities. Students move at their own pace through material, supported by AI that understands their learning profile. The theory is sound and the early evidence from implementations is promising.
The Risks That Must Be Addressed
But the education space carries risks that technology companies must take seriously. The first is the risk of amplifying inequality. AI education systems are only as good as the training data they are built on. If that data comes primarily from affluent school systems, the systems will learn to optimise for the learning patterns and outcomes that characterise those systems. When deployed in under-resourced schools, they may perform poorly or, worse, reinforce existing inequalities.
The second is the risk of inappropriate automation. Not everything that can be automated should be. The relationship between a teacher and a student — the personalised encouragement, the belief in capability, the modelling of intellectual engagement — cannot be entirely replaced by AI without diminishing what education is. The companies that win in education AI will be those that use AI to augment teaching, not replace it.
The third is the risk to student privacy and data protection. Education systems are generating increasingly detailed data about student learning, cognition, and behaviour. If that data is not carefully protected — if it is sold, misused, or lost — the consequences are genuine and lasting. European companies deploying education AI must navigate GDPR carefully and must be transparent about what data they collect, what they use it for, and who has access to it.
The European Approach
European education systems offer some structural advantages for education AI. First, the regulatory framework around student data and algorithmic decision-making is more rigorous than in other regions. GDPR and emerging AI Act requirements mean that European companies deploying education AI cannot ignore privacy and algorithmic fairness; they must build these requirements into their systems from the start. This creates competitive advantage when operating internationally.
Second, European education systems still emphasise critical thinking, self-directed learning, and intellectual development over pure test optimisation. This creates demand for education AI that genuinely enhances learning rather than merely improving test scores. European ed-tech companies are more likely to be built with genuine educational outcomes as the success metric.
Third, European teachers’ unions and education stakeholders are actively engaged in conversations about how AI should be deployed in education. This has the potential to create friction, but it also creates an opportunity for companies that build tools with educators rather than for educators.
What European Ed-Tech Companies Must Do
At NexaTech Ventures, we are backing education AI companies that approach the space with appropriate intellectual humility. They understand that technology is a tool for education, not a replacement for it. They are rigorous about measuring educational outcomes, not just technology adoption. They are transparent about their limitations and about the risks their systems carry.
Specifically, the companies we back are doing three things. First, working directly with education institutions to understand real problems rather than imposing technology-led solutions. Second, investing in longitudinal outcome measurement — tracking what happens to students who use their systems over years, not weeks. Third, building transparency and explainability into their systems so that educators understand what the AI is doing and why.
The education AI market is substantial and genuinely important. The companies that approach it with scientific rigour and educational integrity will build defensible businesses and will make a genuine difference to learning outcomes.
Scott Dylan is the Founder of NexaTech Ventures. He writes on AI, education, and technology investment. Read more at scottdylan.com.


