Zum Inhalt springen
Open book with an AI circuit coming out of it
Ole Lindgren, Stina Boge and Lise Carlsen Mar 14, 20255 min read

AI in the Classroom: Friend, Foe, or Frenemy?

Übersetzung, erstellt von einem KI Übersetzungsdienst

AI in the Classroom: Friend, Foe, or Frenemy?
4:20

61-percent

Picture this: A classroom where 61% of European students are secretly asking AI to help with their homework. No, this isn't science fiction, it's Monday morning in 2025.

Welcome to education's wild frontier, where a generation that has never known a world without smartphones is now handling AI tools like a seasoned cowboy expertly twirling a lasso. They're the most tech-savvy generation in human history, and they've got Large Language Models (LLMs) doing their bidding before they've even had breakfast. 

Teenage boy sat with laptop and surrounded by little robots

AI in education: It's complicated 

The battle lines over AI in education have been drawn, and they're about as nuanced as a Twitter argument. 

In one corner: the optimists, sporting "AI Will Save Education" t-shirts and dreaming of personalised learning utopias where every student receives tailored instruction, freeing up time for teachers to focus on actual teaching.   

In the other corner: the critics, clutching well-worn copies of "1984" and warning that AI will turn students into uncritical zombie plagiarists who can't think for themselves and whose data is being harvested by corporate overlords. 

As ICT Norway points out, this debate often "lacks nuance and is driven more by emotions and fears than facts." It's like watching people argue about whether fire is good or bad, rather than discussing how to use it safely to cook dinner. 

When AI goes rogue: The misinformation menace 

The World Economic Forum ranks misinformation among the top global risks between 2024 and 2026. Large Language Models (LLMs) can generate misleading or biased content, sometimes due to innocent hallucinations, other times as a tool for manipulation. In his 2025 research paper, LLMs, Truth, and Democracy: An Overview of Risks, Mark Coeckelbergh warns that political groups can weaponise LLMs to spread misinformation. Yet, the responsibility for addressing these risks largely remains in the hands of Big Tech.
 
Meanwhile, Big Tech flexes its political muscles. President Trump has already rolled back AI oversight, instructing the Office of Management and Budget to loosen guidelines for AI use in federal agencies. The American Civil Liberties Union argues that his message is clear: full speed ahead, safety features optional.

Lost in translation: The bias beneath the surface 

Researchers Aleksandra Urman and Mykola Makhortykh discovered something fascinating: ask an AI about Russia's invasion of Ukraine in different languages, and you might wonder if you're talking about the same event. AI is wearing different political hats depending on which language you're speaking. The research also revealed that the levels of disparity between languages varied across different LLMs.

Imagine a high school history class asked three different AI tools about World War 2, and getting interpretations so different they might as well have been describing entirely different timelines. It's like asking three eyewitnesses to describe the same car accident and hearing stories involving a sedan, a spaceship, and a horse-drawn carriage. 

Europe's digital compass: Values before virality 

As Europeans, we're approaching these challenges with our characteristic blend of innovation and humanitarian values. 

Our vision for digital learning extends beyond flashy technology to creating spaces where:

  • Student privacy is guarded
  • Teachers can verify information without needing a degree in computer science
  • Technology serves education, not shareholders
  • Digital tools enhance human connection rather than replace it 

The teacher's new toolbox: Beyond banning 

"We believe very strongly that AI literacy, actually teaching kids in developmentally appropriate ways what AI is and isn't and how to use those tools responsibly, is a better approach than just outright AI banning," says Amanda Bickerstaff, CEO of AI For Education and former teacher. 

Banning AI in schools is not the right approach. It just means students will use them anyway, but without any guidance. Instead, savvy educators are teaching students to: 

  • Compare AI-generated content with verified sources 
  • Craft prompts that extract useful information without falling into bias traps
  • Recognise when AI is appropriate (research starting points) and when it's not (forming personal opinions on complex ethical issues) 

A teacher in front of a divided classroom, one side shows students happily engaging with AI while the other side shows apathic students looking at their screens

Transparency as a guiding principle 

In Europe, we believe educational AI should be transparent. Algorithms affecting teaching must remain open to scrutiny, with source criticism built into systems rather than treated as optional extras. 

This means EdTech developers need to explain their AI recommendations. Schools and parents deserve to see inside the "black box" of educational AI, ensuring that AI serves students rather than shaping their perspectives without accountability. You can read more about how we in itslearning work with AI here.

Tomorrow's classroom: Human-centred tech 

The European educational technology sector isn't just creating tools; we're crafting a digital future where democratic values and educational excellence walk hand in hand. 

Tomorrow's education isn't simply about adding more screens and blinking lights. It's about preserving our humanity in an increasingly digital world. It's about creating environments where everyone can develop into their best selves, supported by technology that reinforces our educational traditions and European values. 

Join the conversation: Everyone has a role

How can you contribute? 

  • If you're an educator: Start conversations about AI literacy in your school. Share your experiences of integrating AI in your classroom. 
  • If you're a policymaker: Advocate for thoughtful AI guidelines that balance innovation with appropriate safeguards.  
  • If you're a parent: Talk with your children about how they encounter AI in their learning. Encourage them to question digital information with healthy skepticism. 
  • If you're an EdTech developer: Design with transparency, and involve actual educators in your development process.

By equipping students with the skills to navigate our AI-driven world with both confidence and critical thinking, we're not just preparing them for the future. We're building a digital learning landscape worthy of our highest educational aspirations. This isn't just an opportunity; it's our responsibility. 

The question isn't whether AI belongs in education. It's already there. The real question is whether we'll shape it purposefully or simply react to whatever Silicon Valley sends our way. Europe has long balanced innovation with ethical responsibility. This moment is no different.

ZUGEHÖRIGE ARTIKEL