Why teachers should stop calling AI’s mistakes ‘hallucinations’


Although an essay can be generated by artificial intelligence tools, the result may have several factual errors. It is crucial for schools and districts to communicate to both teachers and students the necessity of inspection in a manner that does not reinforce preexisting biases within these systems. —Image by freepik

Artificial intelligence tools can produce an essay on the migratory patterns of waterfowl or US President Barack Obama's K-12 education agenda in seconds – but the work might be riddled with factual errors.

Those inaccuracies are commonly known as "hallucinations" in computer science speak – but education technology experts are trying to steer away from that term.

"We know that industry tends to use the term 'hallucinations' to allude to errors that are made by [AI] systems and tools," said Pati Ruiz, a senior director of ed tech and emerging technologies for Digital Promise, a nonprofit that works on technology and equity issues, during an Education Week webinar earlier this month.

But researchers who think about how to talk about AI recommend using another name for those errors – such as "mistake," Ruiz said.

First off, the word "hallucinations," Ruiz said, "make(s) light of mental health issues."

And she added that using that word for AI's errors "might give students a false sense of this tool having humanlike qualities. And that's something that we advocate against, right? We advocate for folks to understand these tools as just that, tools that will support us as humans."

‘AI systems and tools make lots of mistakes’

Ruiz noted that she and another expert who spoke during the webinar, Kip Glazer, the principal of Mountain View High School in California, wrote about this issue earlier this year.

What's more, students need to understand that they shouldn't take any information that they get from ChatGPT and similar tools at face value, Ruiz said.

"Generative AI systems and tools make lots of mistakes," she said. "We need to have expertise across content areas so that we can review the outputs of generative AI. And we recommend always questioning the outputs of generative AI systems and tools."

Schools and districts need to make that need for scrutiny clear to teachers and students. "Guidance is really important so that we can all use (AI) effectively and appropriately and in a way that doesn't perpetuate the biases that already exist in these systems," Ruiz added. – Education Week, Bethesda, Md./Tribune News Service

Follow us on our official WhatsApp channel for breaking news alerts and key updates!

Next In Tech News

New app helps you sit up straight while at your computer
Dispose of CDs, DVDs while protecting your data and the environment
'Just the Browser' strips AI and other features from your browser
How do I reduce my child's screen time?
Anthropic buys Super Bowl ads to slap OpenAI for selling ads in ChatGPT
Chatbot Chucky: Parents told to keep kids away from talking AI dolls
South Korean crypto firm accidentally sends $44 billion in bitcoins to users
Opinion: Chinese AI videos used to look fake. Now they look like money
Anthropic mocks ChatGPT ads in Super Bowl spot, vows Claude will stay ad-free
Tesla 2.0: What customers think of Model S demise, Optimus robot rise

Others Also Read