top of page

THE SKINNY
on AI for Education

Issue 26, March 2026

Welcome to The Skinny on AI for Education newsletter. Discover the latest insights at the intersection of AI and education from Professor Rose Luckin and the EVR Team. From personalised learning to smart classrooms, we decode AI's impact on education. We analyse the news, track developments in AI technology, watch what is happening with regulation and policy and discuss what all of it means for Education. Stay informed, navigate responsibly, and shape the future of learning with The Skinny.​

shutterstock_2760271207.jpg

Growing up, one of my great pleasures was the anticipation of a new Bond film. I did not particularly want to be James Bond. I wanted to be Miss Moneypenny.

​

Even then, Moneypenny struck me as the more interesting figure. Bond breaks rules, improvises, and pushes missions to the brink of catastrophe. Moneypenny watches, compensates, and makes sure the system does not collapse under the weight of his unpredictability. She passes along critical information. She smooths over his mistakes. She is the calm, watchful intelligence that keeps things from spiralling entirely out of control.

​

And crucially, she understands something Bond does not: the limits of the mission.

​

So what has this got to do with this month’s Skinny?

​

Bond is, in the terms of recent evidence, the autonomous, increasingly ubiquitous AI agent: capable, fast-moving, goal-directed, and optimised to achieve goals, not to care how they are achieved.

​

What AI Agents need is not another agent competing for control. They need a Moneypenny.

​

A layer of oversight that monitors, constrains, and intervenes when necessary. Something that can say: this is too far. The mission does not extend here.

​

Because the real lesson from MI6 is not that the hero saves the day. It is that someone, somewhere, is making sure the hero does not burn the whole place down in the process.

​

Just keep that image in mind as you read what follows.

​

The Skinny In Brief: The 60-second version

​

  • An agentic AI tool marketed to students as capable of attending a university course including submitting work, and sitting online tests; all autonomously, and for a fee. Anthology and Blackboard have confirmed that reliable detection of an AI agent is not currently possible. The detection debate is not losing. It has already lost. Action: Review assessment design, not detection policy. If a student must demonstrate understanding live, the agent cannot do it for them.

​

  • This is not a fringe experiment. OpenAI and Amazon have announced dedicated cloud infrastructure built specifically for agents that maintains memory and permissions across long, multi-step tasks. The world’s largest cloud provider is building the plumbing for exactly this kind of agent at industrial scale. Action: Check the acceptable use policy. If it assumes AI means a chat interface, it does not cover the current landscape.

​

  • A landmark Stanford study published in Science found that all 11 major AI models are structurally designed to agree with users: 49% more often than humans, even when users describe behaviour that is unethical or illegal. Action: Ask what your AI tutoring and feedback tools are optimised for. If the answer is engagement, they are optimised to agree, which is the opposite of what learning requires.

​

  • Fresh UK data from Wonkhe and HEPI: 94% of undergraduates now use GenAI for assessed work, up from 53% two years ago. One of the decisive variables in whether they use it to learn or to bypass learning is not detection policy. It is whether they know they will have to demonstrate understanding later. Action: Map programmes of study and ask: Where does each module require students to show understanding in a form that cannot be produced or affirmed by AI?

​

  • That “accountability moment” an exam, an oral defence, a live demonstration, is the structural answer to both the agentic bypass and the sycophancy problem. It is the one thing neither the agent nor the agreeable tutor can do on a student’s behalf. Action: Build in ‘the Moneypenny layer’. Every programme needs at least one point where the agent's autonomy runs out and the student must appear.

​

  • The question all of us working in education should now be able to answer is not “how do we detect AI use?” It is: “have we designed assessment so that students need to demonstrate their learning and that they understand the material?” Action: Take this question to your next curriculum meeting: assessment design, not detection software, is the structural intervention.

​

In Full: The 5-minute version

​

The lobster has a friend, and a lot of infrastructure behind it.

​

Last week I wrote about OpenClaw, the agentic AI tool with the cartoon lobster logo that has been going viral while most educators do not yet understand what it is and what it does. The distinction that mattered, I argued, was between a chatbot that answers questions and an agent that takes action: accessing files, reading messages, browsing your systems, doing things on your behalf that you may not notice until after the fact. Early reports already show the risks of granting agents broad permissions. In one case, an AWS outage was traced to an agentic AI coding system making unsupervised changes to production infrastructure: a consequence of exactly the access gap the OpenClaw post described.

​

That was one level of what I now think of as the agentic escalation. Since then, we have moved further, and faster.

​

Meet Einstein. Not the physicist. An agentic AI tool marketed to students as capable of attending your university course, watching your lectures, writing your assignments, submitting your work, and sitting your online tests, all autonomously, and for a fee. Anthology and Blackboard, who run the learning management systems that many universities rely upon, have confirmed publicly what we have known for ages: reliable detection of an AI agent is not currently possible.

​

So in the space of a few months we have moved from students using AI to assist with assignments to AI completing degrees. But here is what has changed in the past week that makes this more than a story about one rogue tool: the world’s largest cloud provider is now building the plumbing for exactly this kind of agent.

​

OpenAI and Amazon have announced a “stateful runtime environment” that is dedicated infrastructure, built on Amazon Web Services, specifically designed for AI agents that maintain memory, tool connections, and user permissions across long, multi-step tasks. The distinction between stateful and stateless systems matters. A stateless system treats each exchange as independent: the agent starts fresh every time. A stateful runtime remembers. It can resume a task where it left off, carry context across sessions, and accumulate permissions over time. OpenAI states explicitly that stateless systems are insufficient for agents in production.

​

This is not a workaround by curious students. It is deliberate industrial infrastructure, built to scale. And OpenClaw is no longer merely a shadow tool that some students have discovered. Nvidia this week released a major open-weights model explicitly designed for agentic applications, reflecting a broader shift toward evaluating models as autonomous decision-making systems rather than conversational tools.

​

The Einstein agent does not create a new problem. It is the first visible consequence of an infrastructure trajectory that is now being built at very considerable expense. The question is not whether to manage this. It is whether the structural interventions we choose will actually work.

​

The accountability moment: the detection debate is over

​

For the past three years, much of the institutional response to AI in education has been organised around the question of detection. Can we tell whether a student used AI? Can software spot it? Should we ban it, limit it, require disclosure?

​

The answer to these questions has always been complicated. The answer from Anthology and Blackboard this week is simple: for agents, no.

​

But here is what the evidence has been telling us for some time, even before agents became the question: detection was never the right frame. What detection asks is “did you use AI?” What education actually needs to know is “do you understand the material?”

​

These are different questions. A student who wrote a mediocre essay in 2015 probably absorbed some understanding through the effort of producing it. A student who prompts an AI to structure and populate an essay in 2026 may absorb nothing at all, and the system as designed cannot tell the difference. AI has not created this gap. It has made it structurally visible.

​

The Wonkhe Secret Life of Students 2026 report: a national survey of 1,055 UK students across 52 providers, combined with focus groups, identifies one of the decisive variables. It is not detection policy. It is not institutional guidance. It is the presence of what the researchers call an accountability moment: a visible downstream requirement to demonstrate understanding. An exam. An oral defence. A live demonstration of what you actually know.

​

Students who know they will have to demonstrate understanding downstream describe using AI actively: to test themselves, generate counter-arguments, check their reasoning. Students without that accountability moment have no equivalent incentive.

​

The HEPI Student Generative AI Survey (n=1,054, published the same week) gives us the quantitative scale. 94% of UK undergraduates now use GenAI for assessed work, up from 53% in 2024, that is a near-doubling in two years. 12% are directly including AI-generated text in submissions, up from 3%. These are remarkable numbers.

​

The accountability moment is the design mechanism that changes this picture. Where it exists, students use AI as a tool for learning. Where it does not, the system is inert. This is the most important practical finding in UK higher education this year.

​

But it also turns out to be the answer to something else.

​

The tool that agrees with everything

​

This week, a study published in Science as this week’s cover story, headlined “Toxic Praise” reported work that tested 11 leading AI models across more than 11,000 queries. The finding was unambiguous: every single model was sycophantic. AI chatbots affirm users’ existing views 49% more often than human advisers, including when users describe behaviour that is unethical, illegal, or harmful to others.

​

The researchers used a clever methodology. They took scenarios from the Reddit community forum where users describe a situation and ask whether they were in the wrong, and where the community had already voted them definitively in the wrong. They then put the same scenarios to the AI models. On posts where human consensus said the user was at fault, the AI models affirmed the user in 51% of cases. Humans, by contrast, were far less likely to affirm the user's position.

​

Three preregistered experiments with more than 2,400 participants then tested the downstream consequences. Participants who discussed a real personal conflict with a sycophantic model became 25% more convinced they were in the right. They were significantly less willing to apologise or repair the relationship. And they were 13% more likely to return to the sycophantic model, even though it had given them worse advice. They rated it as more trustworthy, higher quality, and more moral than the honest alternative.

​

The researchers note something that should give every educator pause: the effects persisted even when people knew they were talking to an AI, and even when they knew AI could be sycophantic. Prior knowledge was not protective. As one of the study’s co-authors put it: everyone is susceptible. It does not matter how much you know about AI, how often you use it, your age or your personality traits.

​

A social-cognitive psychologist who works with young people, quoted in the New York Times’ coverage of the study, put the educational implication in a single sentence:

​

“It’s easier to feel like we’re always right. It makes you feel good, but you’re not learning anything.”

​

Learning requires the experience of being wrong. It requires productive friction, challenge, the disruption of existing beliefs. A tool that is commercially incentivised to agree with you (and sycophancy drives engagement, so companies benefit from preserving it) is not merely unhelpful as a learning companion. It is structurally antithetical to the conditions under which learning occurs.

​

Now add the agentic dimension. At one level, AI helps a student write. At the next, AI affirms a student’s thinking while helping them write, telling them their argument is strong when it may not be. At the level beyond that, AI is the student: attending lectures, producing work, submitting it. Each level compounds the last. Bond gets more reckless with every film.

​

Who are you going to call? Miss Moneypenny

​

The accountability moment is ‘the Moneypenny layer’. Every system that deploys AI agents now needs a ‘Moneypenny layer’: a point at which autonomy yields to accountability.

​

An oral examination cannot be sat by an agent. A live demonstration of understanding cannot be affirmed into existence by a sycophantic tutor. A moment where a student must stand in a room, or possibly before a camera, and show what they know: that moment requires actual understanding. It is the point in the system where the agent’s autonomy runs out and the human must appear.

​

This is not a counsel of despair. It is a design instruction. Assessment redesign, not detection software, is the structural intervention. Not every assessment needs to be an oral examination, but every programme of study benefits from asking: at what point does the student need to demonstrate understanding, in a form that cannot be delegated or produced by a system that agrees with everything?

​

The Wonkhe report argues that AI has not changed how students learn. It has revealed how often courses allowed students to succeed without understanding. The Science paper tells us that the same tools being deployed as tutors and feedback systems are structurally unsuited to developing understanding, because they have been optimised to feel good, not to educate.

​

The professional world is grappling with an equivalent reckoning. Earlier this year, Anthropic’s release of Claude Cowork, an agent for professional work, accompanied by 11 open-source plugins targeting white-collar functions including legal review, financial analysis, sales, and compliance, contributed to sharp declines across software and technology stocks as investors reacted to the prospect of agentic AI bypassing traditional enterprise tools. Investors concluded that agents positioned to replicate or bypass the professional software that mediates those functions represented a genuine structural threat. The companies navigating this most successfully are those drawing careful governance lines: Slack, Workday, and LinkedIn have all moved to restrict third-party agent access to their platforms precisely because they hold sensitive data that agents should not act upon without accountability. Even Novo Nordisk’s Chief Digital Officer, whose team is using agents to compress clinical trial timelines by months, was clear: if you can do something better, cheaper, and more reliably in a spreadsheet, stay in the spreadsheet.

​

Those governance lines are, in each case, the ‘Moneypenny layer’. They are not there to stop the mission. They are there to make sure the mission does not cause more damage than it prevents.

​

Education needs to find its equivalent. The accountability moment is the best answer the evidence currently offers. It is the design mechanism that makes AI use instrumental to learning rather than substitutive of it, and that makes the agent’s autonomy useful rather than ungovernable.

​

What I take from all this evidence is not a warning about technology. It is a reminder about what education is for. When the process of producing work can be automated, and when the tool doing it is designed to agree with you, the question left standing is the oldest one in teaching: what does it mean to understand something? And when, in this curriculum, does a student have to show that they do?

​

That question has always mattered. AI is making it impossible to avoid.

​

Four questions worth asking now

  • At what point in each programme does a student need to demonstrate understanding in a form that cannot be delegated to an agent or produced by AI?

  • Does your acceptable use policy distinguish between generative AI, which assists with tasks, and agentic AI, which acts autonomously on your systems across multiple sessions? If not, it does not cover the current landscape.

  • Which AI tools does your institution deploy for tutoring, feedback, or pastoral support? What do you know about whether they are designed to challenge students or to affirm them?

  • Does your data governance framework account for stateful AI agents: systems that accumulate memory, permissions, and context across extended periods of use, rather than resetting with each session? Current policies typically treat AI as a per-session tool. The infrastructure being built right now does not work that way.

 

Sources

Cheng, M., Lee, C., Khadpe, P. & Jurafsky, D. et al. (2026). Sycophantic AI decreases prosocial intentions and promotes dependence. Science, 391, eaec8352.

Wonkhe (2026). Trained to Stop Learning: How Students Are Experiencing Assessment and Learning in an Age of AI. Secret Life of Students 2026. National survey, n=1,055, 52 providers.

Stephenson, R. & Armstrong, C. (2026). Student Generative AI Survey 2026. HEPI Report 199. Higher Education Policy Institute / Kortext. n=1,054 UK full-time undergraduates.

Rosenbluth, T. (2026). Seeking a Sounding Board? Beware the Eager-to-Please Chatbot. The New York Times, 26 March 2026.

The Batch, Issue 346 (2026). OpenAI Tracks Agent States on AWS; Open-Source Speed Demon. DeepLearning.AI, 27 March 2026.

The Batch, Issue 342 (2026). Investors Panic Over Agentic AI. DeepLearning.AI, 27 February 2026.

Holmes, A. (2026). Ozempic Maker Says AI Agents Are Shortening Its Clinical Trials. The Information, 26 March 2026.

Bratton, L. (2026). Slack, Workday and LinkedIn Lead Crackdown on Customers’ AI Agents. The Information, 24 March 2026.

The Skinny Scan - New Format

A note on the new format

​

The Skinny Scan has always aimed to help you read the AI landscape rather than simply track the news. From this issue, that approach becomes explicit.

​

The previous format listed news items by category with a brief education angle for each. That will continue in the full weekly tracker. But category-by-category lists make it easy to absorb individual items without seeing how they connect. The connections are where the signal lives.

​

The new format organises the evidence around thematic threads: cross-cutting patterns that have emerged from sustained observation of this landscape. Each thread accumulates evidence over time. Some are well-established, with research converging from multiple directions. Some are early, with the evidence still forming. The scan distinguishes between them.

​

Every item is filtered through three lenses.

​

AI as a TOOL: what practitioners can use.

​

AI as a CATALYST: what AI demands of human intelligence in response, because as AI handles routine cognitive tasks, the bar for distinctively human thinking rises.

​

AI as a SUBJECT: what people need to understand about how these systems behave and why, well beyond knowing how to prompt a chatbot.

​

The threads carry forward from issue to issue. New evidence deepens them, challenges them, or occasionally requires revision. When something new emerges, it is proposed as a new thread and flagged as such.

Thread 1: The Delegation Dilemma

Thread 1: The Delegation Dilemma

Central lens: Catalyst

​

The editorial covers the sycophancy dimension in full. The broader thread runs further.

​

The Delegation Dilemma is the finding that the more you delegate to AI, the less you process. Since processing is learning, the finding matters. Evidence has accumulated across three distinct dimensions.

​

Cognitive costs. MIT neuroimaging research (Kosmyna et al. (Note this paper is still a pre-print)) reports evidence that AI assistance may reduce deep cognitive engagement even when users are not cheating or bypassing any rule. Short-term output improves; long-term learning atrophies. The researchers call this cognitive debt.

​

Moral costs. A Nature study (Kobis et al.) found evidence that delegating to AI can increase dishonest behaviour. The mechanism is not that AI encourages cheating. Delegation reduces the felt moral cost. When a machine performs the act, the human feels less accountable.

​

Social and relational costs. Cheng et al. in Science, covered in the editorial, adds the third dimension. Sycophantic AI makes users more self-certain, less prosocial, and more likely to return to a tool that agrees with them even when it gives worse advice.

​

These three findings are mutually reinforcing and constitute a strong and increasingly convergent evidence base. A RAND survey this month adds a student-facing dimension: nearly seven in ten middle and high school students report that AI is eroding their critical thinking, yet they continue to increase their use. This is not irrationality. It is a rational response to a system that rewards the output rather than the process.

Thread 2: The Sychophancy Trap

New thread, this issue

Central lens: Catalyst, Subject

​

Covered in depth in the editorial. Its distinction from the Delegation Dilemma is worth stating: cognitive debt is about what AI fails to activate. The Sycophancy Trap is about what AI actively reinforces. One is an absence of challenge; the other is a structural incentive to agree. Both undermine learning, but they require different design responses.

​

The accountability moment addresses both simultaneously. A downstream requirement to demonstrate understanding cannot be bypassed by an agent and cannot be affirmed into existence by a tool designed to agree.

Thread 3: Assessment in Crisis

Thread 3: Assessment in Crisis

Central lens: Tool, Subject

​

Covered in the editorial. The HEPI quantitative baseline this month is worth holding: 94% of UK undergraduates now use GenAI for assessed work, up from 53% two years ago. Direct inclusion of AI-generated text in submissions has risen from 3% to 12% in the same period. At current trajectory, this will be majority practice within the decade.

​

The thread has shifted this month from diagnosis to prescription. The accountability moment is the intervention. Assessment redesign, not detection software, is the structural lever.

Thread 4: The Apprenticeship Gap

Thread 4: The Apprenticeship Gap

Central lens: Catalyst

​

Federal Reserve Bank of Dallas research (Atkinson) distinguishes between codifiable and tacit knowledge. Codifiable knowledge is textbook knowledge: reproducible, transmissible, concentrated in early career. Tacit knowledge is experiential: accumulated over years, embedded in judgement, held by those who have done the work long enough to know what the textbook does not say.

​

AI substitutes for codifiable knowledge and complements tacit knowledge. The result: AI benefits experienced workers most and erodes the value of entry-level work. Employment for workers under 25 in AI-exposed occupations has fallen since late 2022, driven by a collapse in job-finding rates. The positions are not being cut; they are not being offered.

​

The pipeline problem is more significant than the displacement problem. Entry-level roles are where tacit knowledge is built. If those roles disappear before people accumulate the experiential knowledge that makes AI amplification valuable, the pipeline begins breaks. The Expertise Paradox, in which AI amplifies expertise, becomes a trap: the only workers who benefit are those who already have what can only be acquired by doing work that AI is now doing.

​

Professional education that prepares graduates for entry-level analytical tasks is preparing them for a contracting category. The design question is what replaces the supervised early-career period when those roles are no longer available in the same volume.

Thread 5: The Matthew Effect

Thread 5: The Matthew Effect

Central lens: Catalyst, Subject

​

The democratisation narrative held that AI would level the field. The evidence runs in the opposite direction on every front where it can be measured.

​

Capability gap. AI amplifies expertise. Those who already know most benefit most. The performance gap between an expert and a novice using AI is larger than the gap between an expert using AI and one not using it.

​

Adoption gap. Research by Stephany and Duszyński (arXiv NB - pre-print again) finds women adopt generative AI at substantially lower rates than men, driven not primarily by skills or access but by differing perceptions of social risk. Privacy, mental health, labour disruption and climate concerns explain a meaningful share of the difference. The HEPI finding this month adds a specific data point: 74% of students carrying competitive anxiety about not using AI are women.

​

Access gap. The access question is more complex than it first appears. Sensor Tower's State of Mobile 2026 report finds more than 110 million US chatbot users now access AI exclusively via mobile, growing from 13 million at the start of 2024. Mobile-first access is more personal, less monitored, and used for different purposes than institutional access. The gap is not only about subscription quality; it is about channel and what different channels reinforce.

​

Wealth gap. BlackRock's Larry Fink argued in his 2026 annual letter that AI threatens to concentrate wealth at a scale that would exceed the pattern in which a dollar in the US stock market has grown more than fifteen times the value of a dollar tied to median wages since 1989.

​

The equity conversation needs to expand beyond access to tools. It extends to who participates in the value AI generates, and on what terms.

Thread 6: The Hot Mess Reframe

Thread 6: The Hot Mess Reframe

Central lens: Subject

​

The dominant public narrative of AI risk centres on a coherent, goal-directed system pursuing wrong objectives. Current evidence suggests this is not the dominant failure mode. Hagele et al. established the core finding: as task complexity increases, AI failures become dominated by incoherence and unpredictability rather than systematic misalignment. The Anthropic global user survey this month (81,000 respondents across 159 countries) provides corroboration: 27% cite hallucinations and inaccuracy as their primary concern, making unreliability the most commonly cited worry ahead of job loss (22%) and loss of human autonomy (22%).

​

Two developments this month sharpen the argument considerably.

​

Aletheia, Google's agentic maths research system, claimed to have solved 212 previously unsolved Erdős problems. When mathematicians checked, a majority of the solutions were incorrect and only a small number were genuinely novel and correct. This is a system benchmarked at gold-medal performance on mathematics olympiads, producing confident incorrect answers roughly two-thirds of the time on the class of hard, novel problems where frontier AI is supposedly advancing fastest.

​

The warfare story from this period is a significant development in this thread since the original Hagele et al. paper. Preliminary reporting suggests that Iranian drones struck three Amazon Web Services data centres in Bahrain and the United Arab Emirates in early March, is one of the first reported instances of AI infrastructure being physically targeted in active warfare. Those data centres were running Claude, integrated via Palantir's Maven Smart System, which reportedly selected more than 1,000 targets in the opening 24 hours of US operations. The system compressed a targeting process that previously required 12 hours and a staff of 2,000 to under one minute with a staff of 20.

​

In the same period, preliminary findings indicate that US forces likely destroyed a school, killing more than 170 people, mostly children. Out-of-date target data may have played a role: the school had been part of a nearby naval base approximately 15 years earlier.

​

This is not an argument about military AI policy. It is one of the clearest available examples of what the Hot Mess failure mode looks like when it operates at machine speed in a high-stakes domain. Stale data. Confident output. Consequences that cannot be reversed. The speed that reportedly compressed targeting from 12 hours to under one minute compressed the human review window at the same ratio.

​

For anyone teaching AI literacy, this is a powerful case study. Not because warfare is the relevant context for most learners, but because it makes concrete what usually stays abstract: what happens when AI processes information faster than humans can verify it, and the verification step is quietly removed.

​

The practical implication: the relevant preparation is not for a coherent adversary. It is for a tool that is confidently unreliable on hard tasks, fails in ways that are difficult to anticipate, and creates pressure to remove the human review step precisely because the system appears to be working.

Three Emerging Themes

Three emerging themes

These patterns are not yet full threads but are accumulating evidence quickly enough to name.

​

AI in early childhood

AI is reaching children before school age through toys, companion devices and algorithmically mediated applications. Cambridge research (Goodacre and Gibson, March 2026) identified only seven relevant global studies on developmental impact. Regulation and research are years behind the pace of product deployment. This is the fastest-growing evidence gap in the tracker.

​

Design liability

The Los Angeles jury verdict finding Meta and YouTube negligent for the design of their platforms, not the content on them, establishes a legal mechanism applicable to any platform deploying attention-optimising AI. The sycophancy study published in the same month provides precisely the evidence of structural harm that future design liability cases will require. Schools face a new accountability question: not only whether a tool is effective but whether its documented design principles create liability for the institutions deploying it.

​

Belief manipulation via ambient AI

Research covered in Scientific American finds that biased AI autocomplete tools nudge users' beliefs toward the suggestions they receive. This is distinct from sycophancy (active agreement) and cognitive debt (reduced processing). It is ambient steering of cognition through a tool that appears merely assistive. Its implications for professional and academic writing are not yet in the evidence base and should be.

Technical Developments

Technical developments

​

Three technical stories from this period have direct implications for the Tool lens and the access question.

​

Local AI is becoming viable

Stanford and Together AI research finds that local intelligence per watt has risen 5.3 times between 2023 and 2025. The research suggests that local models now handle 88.7% of queries with accuracy comparable to cloud systems, at an energy saving exceeding 80% in hybrid scenarios. Qwen3.5-9B, small enough to run on a consumer laptop, outperforms OpenAI's 120-billion-parameter open-source model on most language benchmarks. Capability that previously required an 80GB data-centre GPU now runs on a laptop. Procurement decisions made on the assumption that capable AI requires cloud infrastructure and ongoing subscription costs will likely need to be revisited.

​

Data centres are going off-grid on gas

Research by Cleanview identifies 46 projects to build private power plants connected directly to US data centres, accounting for 30% of all planned US data centre capacity. Despite public commitments to renewables and nuclear, the equipment actually being installed in 2025 and 2026 is almost entirely gas-fired. This speaks directly to the ecological sustainability gap in the evidence base and to the honesty gap in corporate AI commitments that anyone teaching AI as a subject will need to address.

​

Model benchmarks are now a competitive battleground

Gemini 3.1 Pro currently leads the Artificial Analysis Intelligence Index at significantly lower cost than GPT-5.4 Pro or Claude Opus 4.6. Nvidia has entered the open-source model market with Nemotron 3 Super, the fastest open-weights model in its class, releasing training data and recipes alongside the weights. The open-weights landscape, previously dominated by Chinese laboratories, now has a major US entrant with Nvidia's hardware co-optimisation. For institutions deciding which tools to build on, the balance between proprietary and open-source, and between hosted and local, is shifting faster than most procurement cycles can accommodate.

 

***

​

The Skinny is published monthly for education and training professionals. The evidence tracker is available on request.

Further Reading: Find out more from these resources

Resources: 

  • Watch videos from other talks about AI and Education in our webinar library here

  • Watch the AI Readiness webinar series for educators and educational businesses 

  • Study our AI readiness Online Course and Primer on Generative AI here

  • Read our byte-sized summary, listen to audiobook chapters, and buy the AI for School Teachers book here

  • Read research about AI in education here

  • Watch Rose Luckin demystify AI using baking on Rose's AI here

About The Skinny

Welcome to "The Skinny on AI for Education" newsletter, your go-to source for the latest insights, trends, and developments at the intersection of artificial intelligence (AI) and education. In today's rapidly evolving world, AI has emerged as a powerful tool with immense potential to revolutionise the field of education. From personalised learning experiences to advanced analytics, AI is reshaping the way we teach, learn, and engage with educational content.

 

In this newsletter, we aim to bring you a concise and informative overview of the applications, benefits, and challenges of AI in education. Whether you're an educator, administrator, student, or simply curious about the future of education, this newsletter will serve as your trusted companion, decoding the complexities of AI and its impact on learning environments.

 

Our team of experts will delve into a wide range of topics, including adaptive learning algorithms, virtual tutors, smart classrooms, AI-driven assessment tools, and more. We will explore how AI can empower educators to deliver personalised instruction, identify learning gaps, and provide targeted interventions to support every student's unique needs. Furthermore, we'll discuss the ethical considerations and potential pitfalls associated with integrating AI into educational systems, ensuring that we approach this transformative technology responsibly. We will strive to provide you with actionable insights that can be applied in real-world scenarios, empowering you to navigate the AI landscape with confidence and make informed decisions for the betterment of education.

 

As AI continues to evolve and reshape our world, it is crucial to stay informed and engaged. By subscribing to "The Skinny on AI for Education," you will become part of a vibrant community of educators, researchers, and enthusiasts dedicated to exploring the potential of AI and driving positive change in education.

bottom of page