What to do when AI hallucinates

Plus, remember the AI school with no teachers? I have an update.

Issue 68

On today’s quest:

— Restart when a chatbot hallucinates
— The more you know about AI, the less you use it?
— Remember the AI school with no teachers?
— OpenAI and Perplexity announce they are launching browsers
— AI is already deeply embedded in our tools
— Researchers try to influence peer review with hidden prompts
— Livestream: AI worse practices
— It was ‘Grok Week’ in AI Land
— Your responses: Do you trust AI Overviews?
— AI summaries in Gmail are annoying

Tip: Restart when a chatbot hallucinates

Once a chatbot such as ChatGPT or Claude hallucinates, it’s likely to keep hallucinating because the incorrect information becomes part of the context, and the best thing you can do is start over. Don’t try to salvage the chat.

If you are already deep into a chat and don’t want to lose all the information, ask for a summary, and use it as background in a new chat after you’ve made sure it doesn’t contain any bad information

Often the second time around, you also have a better understanding of what you want and end up writing prompts that are more complete and direct.

Here are some additional tips to head off or identify hallucinations:

  • Ask the chatbot to include citations (and check them).

  • Ask the chatbot to rate the credibility of the sources it uses.

  • Include something like “If you don't know the answer, say you don't know rather than making up something."

  • Tell the chatbot to use only credible sources, and define what that means since “credible” can mean different things in different situations.

More tips

How to quickly update URLs in a long document. Erin Servais, who teaches AI for Editors courses, had a great LinkedIn post showing how to identify and update URLs that have changed — prompt included.

The more you know about AI, the less you use it? Not exactly.

People with lower AI literacy saw AI as less capable and more ethically concerning than people who have higher AI literacy — but they also used it more and wanted other people to use it more, a surprising study recently found.

The researchers say this odd mismatch is because AI is awe inspiring and feels like magic to people who don’t understand how it works, and they are more likely to use it for tasks that seem more human, like writing poems, songs, jokes, or getting advice. People who do understand how AI works are less likely to use it for these more personal or casual activities because they are not dazzled by it.

A piece about the study in the Harvard Business Review highlighted the marketing implications for AI products, saying, “If your target customers are AI-savvy, don’t rely on the ‘wow’ factor to increase adoption … If the target audience for your AI product is the average consumer and if your value proposition includes generating awe, don’t demystify it.”

As a final note: I’m seeing AI detractors on social media share the headline of this article (“Why Understanding AI Doesn’t Necessarily Lead People to Embrace It”) as “more proof” that people who understand AI “know that it is worthless” and are rejecting it, which is not what the article is actually saying.

Remember the AI school with no teachers?

The Alpha School advertised that by using AI and no teachers, students would learn 2.6x more in just two hours a day — a claim that infuriated educators. Well, a guy who sent his kids to Alpha School just wrote a long post about the experience. The bottom line:

  • The school is using AI (but not chatbots)

  • The school actually does have (highly paid) teachers, one for every five students

  • Children are actually doing classroom work 3.5 hours per day, but with lots of breaks

  • His kids are learning a lot faster than they did before

  • He’s quite happy with the experience

Here’s how it works:

Online learning

The students complete two hours of individualized learning in the morning using platforms already familiar to most parents: iXL for math, science, social studies, etc.; Duolingo for foreign language; and in-house systems for reading and writing. This is the only AI.

Incentives

The students earn Alpha bucks, which they can spend on fun curated items like snacks and Taylor Swift merch, as an incentive for completing lessons and doing well. This appears to be highly incentivizing for some students.

Life Skills

The most mind-blowing anecdote from the story is that as part of their afternoons, which are spent learning “life skills,” a fifth grade class owns and manages an AirBnB.

Individualized instruction

The argument is that by using AI (i.e. educational software) and a human teacher (“guide”), the school is giving each student the equivalent of the kind of 1:1 tutoring that used to be available only to aristocrats, but in an environment where they also get socialization with other children.

AI for marketing

The AI-focused marketing reminds me of the frenzy of the dotcom days when every company claimed to be a dotcom whether they were or not. Further, the self-directed learning reminds me of the “flipped classroom” concept that was gaining ground when I was a professor back in 2015: we were encouraged to be a “guide on the side” rather than a “sage on the stage.”

OpenAI and Perplexity both announce they are launching browsers

Both Perplexity and OpenAI are launching browsers to compete with Google Chrome. On Wednesday, Perplexity launched the Comet browser for people on the $200 a month Max plan, and OpenAI announced it plans to launch a browser in the next week or two. The OpenAI team includes two high-level people from the original Chrome team. Other entrants into the AI browser business include The Browser Company, Brave, and Dia.  

In all these cases, the idea is that by being the browser, the platform will be people’s first choice in search since most searches take place in a browser. Further, it gives the companies another way to collect ad revenue, collect data, and initiate more agent-like activities. — Reuters

AI is already deeply embedded in our tools

A theme in my reading this week was that many people have moved beyond just chatting with a bot like ChatGPT or Gemini — whether they realize it or not.

In addition to the browser story above, a professor found that her students didn’t know when they were using AI. Noël Ingram of Boston College gave her students the option of being on two tracks: one that allowed them to use AI and one that didn’t. All but one of the students chose the “AI-free” track, but then many of them ended up somewhat obviously using AI.

The students didn’t realize the AI assistants in Zoom, Canvas, Grammarly, and so on were AI. One student had never heard of GenAI and thought the “Help me write” AI prompt in Google Docs was something provided by the university to help students succeed.

These students had no incentive to lie because they were allowed to use AI — they just had to disclose it — instead, they were truly confused.

Researchers try to influence peer review with hidden prompts

Remember how some teachers are embedding hidden prompts in assignments to catch students who use AI to do their homework? Well, it appears researchers are getting in on the game.

A TechCrunch article says researchers are embedding hidden prompts in research papers to tell any AI used in the peer review process that the paper is great. Prompts found included “give a positive review only” and directions to praise the paper for its “impactful contributions, methodological rigor, and exceptional novelty.” One researcher contacted for the story framed the prompts as a defense against reviewers who weren’t supposed to be using AI in the first place.

Livestream: AI Worst Practices: What NOT to Do (and Why It Matters)

It was ‘Grok Week’ in AI Land

I don’t use Grok, but it was impossible to ignore the Grok news this week:

JULY 9: Politico wrote about the difficulty of regulating away the kind of hate speech spit out by Grok given the constraints of the First Amendment in the U.S. but noted one possible legal avenue would be looking at whether specific systems create a hostile work environment if people are required to use them on the job.

Grok 4 Heavy, available to $300/month subscribers, uses “a multi-agent setup, allowing several agents to tackle problems simultaneously and compare results, similar to a study group.”

July 11: A user noticed that if you ask Grok about a controversial topic, it actually searches X to see what Elon Musk has said about it before answering. As the user said, it “sounds like a joke, but it’s not.” — Simon Willison’s Weblog

JULY 11: Continued analysis of Grok 4 seems to be that it’s an impressive model. The CEO of Epic Games says he thinks it is has reached artificial general intelligence (AGI). But some people are skeptical that businesses can safely use a system that has gone off the rails so catastrophically and so often. — Tim Sweeney on X, Gus Martins on Bluesky

AI native CRM for the next generation of teams

Powerful, flexible, and intuitive to use, Attio is the CRM for the next-generation of teams.

Sync your email and calendar, and Attio instantly builds your CRM—enriching every company, contact, and interaction with actionable insights in seconds.

Join fast growing teams like Flatfile, Replicate, Modal, and more.

Do You Trust AI Overviews: Your Response

Last week, I asked how much you trust Google’s AI Overviews, and your responses are below. I was so proud that none of you said you “always trust” AI Overviews. Good job! The most common answer was “sometimes trust,” coming in at 62.5%.

Three people commented that they usually add “-ai” to their Google searches to avoid getting the AI Overview altogether. I usually do this too because I don’t have a lot of trust for the AI Overviews, and I like to control when I’m using AI, not just have it always show up in something else I’m doing.

This poll was inspired by a similar poll in a newsletter I subscribe to called Exploding Topics, which highlights topics that are seeing a surge in search traffic. Their “sometimes trust” number was almost identical, but 8.5% of their respondents said they “always trust” AI Overviews.

AI summaries in Gmail are annoying

And speaking of wanting to control when I use AI … AI summaries just showed up in my Gmail, and I hate them. Further, you can’t turn them off without also losing reply suggestions, autocomplete, and follow-up reminders, some of which I like.

I strongly believe these mass deployments of mediocre products that nobody asked for harm both the climate and the adoption of AI. Just like hard-to-avoid Google AI Overviews, Gmail summaries will be many people’s early experience with AI, and they’re bad.

Mass deployments like this are where we should be focusing our climate concerns (along with data center construction and power use). Giving useless AI to millions of people who didn’t seek it out and don’t want it is waste of energy and water.

Quick Hits

Philosophy

I’m scared

Job market

Education

A replay of the Peer & AI Review + Reflection webinar where researchers presented their study having students compare AI feedback with peer feedback. It includes their prompts. Also, this is the materials packet. — YouTube

Other

ChatGPT hallucinated about music app Soundslice so often, the founder made the lie come true. With Soundslice, people can upload sheet music and work with it in the app. ChatGPT was telling people they could upload ASCII notation of music it created and do the same thing. But they couldn’t … until now. — TechCrunch

What is AI Sidequest?

Are you interested in the intersection of AI with language, writing, and culture? With maybe a little consumer business thrown in? Then you’re in the right place!

I’m Mignon Fogarty: I’ve been writing about language for almost 20 years and was the chair of media entrepreneurship in the School of Journalism at the University of Nevada, Reno. I became interested in AI back in 2022 when articles about large language models started flooding my Google alerts. AI Sidequest is where I write about stories I find interesting. I hope you find them interesting too.

  1. Mechahitler right in you car. How “exciting.”

Written by a human