The fundamentals of artificial intelligence and education in 2026
This blog is part of a short series of articles I will be writing to welcome the start of 2026. All opinions expressed are my own and do not represent the views of any individuals or organisations mentioned in the articles. Further, I am a UK-based academic scientist and these articles reflect the UK context.
In the previous part of this blog series, I discussed how I became co-Director of the Ellison Institute of Technology Fundamentals of AI Centre for Doctoral Training (EIT FoAI CDT), and how I think about building PhD programmes more generally. In this part, I want to focus on a more specific and arguably more difficult question: what exactly are the fundamentals of AI? And, closely related, how should we think about training doctoral students in this area as we move into 2026?
What are the fundamentals of AI?
If you ask ten people what the fundamentals of AI are, you will almost certainly get ten different answers. That is not a failure of the field so much as a reflection of its breadth, its pace of change, and its unresolved intellectual questions.
For some, the fundamentals of AI are grounded in long-standing theoretical foundations. This view emphasises areas such as probability theory, statistics, optimisation, learning theory, information theory, logic and algorithms. From this perspective, progress in AI rests on understanding generalisation, uncertainty, sample efficiency, and the limits of computation. These foundations have underpinned decades of work in machine learning and remain essential for reasoning rigorously about models and methods.
For others, particularly given developments over the last decade, the fundamentals of AI are increasingly associated with understanding modern, large-scale systems whose behaviour was not explicitly programmed. This includes work aimed at explaining the emergent and often poorly understood properties of contemporary AI systems: large language models, foundation models, and other architectures whose capabilities appear to arise from scale, data, and training dynamics rather than carefully specified rules.
In this view, the fundamentals are less about closed-form theory and more about empirical understanding. Questions about representation, emergence, alignment, robustness, and interpretability become central. Why do these systems work as well as they do? Where do they fail? How stable are their behaviours across contexts, data distributions, and deployment settings?
There is also a third, comtemporary perspective that situates AI fundamentals in interaction with the world. From this angle, the fundamentals are not purely mathematical or computational, but socio-technical. They include how AI systems interact with people, institutions, and environments; how incentives shape deployment; how values, norms, and power structures are encoded; and how systems behave once embedded in real-world settings.
Building a CDT in FoAI
These perspectives are not mutually exclusive, but they do pull in different directions when it comes to training doctoral students. A curriculum built solely around classical theory risks being disconnected from the systems that currently dominate practice. One focused only on contemporary models risks producing researchers without the conceptual tools to reason about why those systems behave as they do, or how the field might evolve beyond them.
The challenge, then, is not to identify a single definitive set of fundamentals, but to decide which foundations matter most for developing researchers who can contribute meaningfully to the field over the long term - researchers who can adapt as the technology changes, rather than being trained narrowly around the current state of the art.
For this reason, in designing the FoAI CDT we organised the intellectual space around three non-mutually exclusive domains. These domains are not intended to be rigid tracks or silos, but overlapping perspectives through which students can frame and evolve their research over time.
(1) Theory and Foundations
This domain encompasses the deeply theoretical and foundational aspects of mathematics and computer science that underpin AI. It includes areas such as learning theory, optimisation, probability, statistics, information theory, algorithms, and the mathematical properties of representations.
Research in this space often focuses on understanding the conditions under which AI methods work, when they fail, and what guarantees - if any - can be provided about their behaviour. If you are interested in the analysis of optimisation algorithms, generalisation bounds, or the relative efficiency of different representational frameworks, your work would naturally sit here.
While research in this domain can appear distant from immediate practical application, its value lies in producing rigorous, generalisable insights. These foundations provide the conceptual scaffolding that allows the field to move beyond empirical trial-and-error and to reason more systematically about the capabilities and limitations of AI systems.
(2) Applied Fundamentals
Applied Fundamentals encompasses research driven by scientific and real-world applications, where empirical observation plays a central role in shaping new AI concepts and methods. Researchers working in this domain often start from practical problems—whether in science, engineering, or other domains—and use these challenges to interrogate existing theory and methods.
Work in this space frequently involves developing novel variations of established approaches, often breaking the simplifying assumptions that made the original theory mathematically tractable. In doing so, it tests the robustness of foundational ideas and extends them to settings that better reflect the complexity of real-world data and systems.
This domain sits at the interface between theory and practice. It recognises that many advances in AI have emerged not from theory alone, but from the tension between theoretical expectations and empirical behaviour. Applied Fundamentals provides a structured way to engage with that tension rather than treating it as an inconvenience.
(3) Systems & Engineering
The Systems & Engineering domain focuses on the study of modern large-scale AI systems and the principles that govern their construction, efficiency, and behaviour. This includes work on architectures, training pipelines, infrastructure, scalability, robustness, and the emergent properties of complex systems.
Researchers in this domain often operate at a higher level of abstraction, making use of commonly adopted building blocks - such as transformers - rather than re-deriving methods from first principles. The goal is not simply to build bigger or faster systems, but to understand how system-level choices interact with learning dynamics, data, and deployment contexts. For example, this might include the study of agent-based systems and reinforcement learning, where behaviour emerges through interaction with environments, tools, and other agents over time. In these settings, questions about coordination, credit assignment, stability, and long-horizon behaviour become central, further reinforcing the need for system-level abstractions that go beyond static model performance.
A key challenge in this space is reconciling empirical observations of system behaviour with approximate or simplified models that allow analysis, comparison, and generalisation. Systems & Engineering research is therefore critical for translating theoretical insight into practice at scale, and for ensuring that AI systems behave in predictable and reliable ways when deployed.
Maths, Physics and Engineering
In my mind, these domains are analogous to the way in which pure mathematics, theoretical physics, and engineering have historically been linked and have interacted with one another.
Pure mathematics develops abstract structures and results without immediate concern for application, yet over time it has repeatedly provided the language and tools that make new theories possible. Theoretical physics has often drawn directly on these mathematical foundations to formulate models of the natural world, while simultaneously posing new problems that push mathematics in unexpected directions. Engineering, in turn, takes physical laws and theoretical insight and turns them into working systems - bridges, engines, circuits, and communication networks - that operate in the real world.
Crucially, engineering does not attempt to understand these systems at the most fundamental, atomistic level. Instead, it relies on abstraction, approximation, and modelling at appropriate scales. No engineer designs a jet engine by simulating individual particles; they work with higher-level representations that capture the behaviour that matters for performance, safety, and reliability.
AI increasingly occupies a similar position. Theoretical work provides insights into learning, optimisation, and representation. Applied research tests and stretches these ideas in complex, often messy real-world settings. Systems and engineering research builds artefacts - large, interconnected models and infrastructures - whose behaviour emerges from interactions across many layers and cannot be fully understood by inspecting individual components in isolation.
As with engineering, the challenge is to develop the right abstractions: models that are simple enough to reason about, yet rich enough to capture essential behaviour. Understanding modern AI systems is therefore not just a matter of more data or more computation, but of finding conceptual frameworks that allow us to reason about complexity without being overwhelmed by it.
Seen this way, training doctoral students in the fundamentals of AI is not about choosing between theory and practice. It is about enabling movement between levels of abstraction, and about equipping researchers with the intellectual tools to contribute at the points where these domains intersect. Historically, many of the most significant advances have occurred precisely at those intersections, and there is good reason to believe the same will be true for AI.
Universities staying relevant
Taken together, these three domains provide a flexible framework for the FoAI CDT. Rather than forcing a narrow or fashionable definition of AI, they create an inclusive umbrella under which a broad range of students and academic supervisors can meaningfully contribute. This structure also allows us to design a teaching curriculum that moves between levels of abstraction, evolves as the field changes, and situates technical content within a wider intellectual landscape. Instead of prescribing a single path, the aim is to equip students with foundations that allow them to navigate a field that remains very much in flux.
That flexibility is not just desirable; it is necessary.
There is an increasing gulf between academic AI and industry-led AI, driven by differences in scale, pace, and incentives. Many of the most visible advances in AI now emerge from environments with access to vast computational resources, proprietary datasets, and engineering teams operating on timelines that are difficult to match within universities. This is not a criticism of academia, but a practical reality.
Even at doctoral level, UK higher education cannot generally aspire to sit at the absolute cutting edge of large-scale AI development in the way that a small number of well-resourced industrial labs can (and increasingly even small-to-medium enterprises, SMEs). Attempting to compete directly on those terms is neither realistic nor, in my view, desirable. Doing so risks distorting academic priorities and undervaluing the distinctive contributions that universities are uniquely well placed to make.
Universities have a long-standing and central role as educators and stewards of good research practice. This goes beyond the development of practical skills—such as software engineering or familiarity with particular toolchains—to cultivating a deeper understanding of how and why methods work.
Anyone who has attempted to publish in a major machine learning conference in recent years will recognise the increasingly formulaic nature of both writing and reviewing. A new model is proposed, it is evaluated against a set of benchmarks, followed by a sequence of ablation studies. Performance improves, components are removed, curves are plotted. The implicit question is always: does it work better than what came before?
But the more important question—often left unanswered—is why.
Why does a particular architectural choice matter? Why does performance improve or degrade under certain conditions? Why should we expect behaviour observed on a benchmark to persist across different datasets, tasks, or deployment contexts? Without engaging seriously with these questions, it becomes difficult to distinguish genuine insight from incremental optimisation.
Academic environments are uniquely positioned to slow things down enough to interrogate assumptions, to prioritise understanding over leaderboard performance, and to reward explanation rather than just results. In a field moving as quickly as AI, that role is not antiquated—it is essential.
Developing researchers who are comfortable asking why, and who can reason carefully about evidence and uncertainty, is ultimately more valuable than training students to optimise for the conventions of the moment. Tools, benchmarks, and dominant paradigms will change. The ability to think critically and adapt is what endures.
Partnership with Ellison Institute of Technology
These realities are exactly why the partnership with the Ellison Institute of Technology (EIT) is so important for the FoAI CDT.
The collaboration brings together genuinely complementary strengths. The University of Oxford provides deep academic expertise, a culture of critical inquiry, and a long tradition of foundational research and doctoral training. It is an environment well suited to asking difficult questions, interrogating assumptions, and prioritising understanding over short-term performance.
EIT, by contrast, operates closer to an industrial research model: problem-driven, well-resourced, and oriented towards ambitious scientific applications of AI. This brings exposure to large-scale systems, modern tooling, and the kinds of empirical challenges that increasingly define contemporary AI research. Importantly, this is not industry in the narrow sense of product optimisation or rapid commercialisation, but an environment where real-world constraints and scale are treated as first-class research concerns.
What makes this partnership work is that these perspectives are aligned rather than in tension. Many researchers at EIT place a high value on fundamental understanding and are motivated by questions that cut across theory, application, and system-level behaviour. This creates space for doctoral students to engage with modern AI systems without being trained narrowly around benchmarks, leaderboards, or transient trends.
In practice, this means students can be exposed to different research cultures and ways of working while maintaining a shared commitment to rigour, explanation, and long-term impact. They can see how theoretical ideas break — or need to be adapted - when confronted with real systems, and how empirical observations can in turn motivate new foundational questions.
The FoAI CDT is therefore not an attempt to replicate large industrial AI labs within a university setting, nor to chase the cutting edge for its own sake. Instead, it is about creating an environment in which students learn to move fluently between abstraction and implementation, between explanation and experimentation, and between academic and applied perspectives.
In a field evolving as rapidly as AI, this ability to navigate between worlds may be more valuable than proximity to any single technical frontier. It is also, in my view, where doctoral training can make its most distinctive and lasting contribution.
In that sense, the question is not whether universities can keep up with AI, but whether they can continue to shape the kinds of researchers the field will need in the long term.
Coming next: In the next blog in the series I will talk about health data - what is its value?