Axios AI+

June 04, 2025
Looking forward to seeing you later today for AI+ Summit New York, where you'll hear conversations with New York Gov. Kathy Hochul, actor-filmmaker-entrepreneur Joseph Gordon-Levitt, Runway co-founder and CEO Cristóbal Valenzuela, WndrCo founding partner Jeffrey Katzenberg and more.
If you aren't going to be there in person, you can watch the livestream here starting at 2:30pm ET. Today's AI+ is 1,188 words, a 4.5-minute read.
1 big thing: Why AI is still making things up
AI makers could do more to limit chatbots' penchant for "hallucinating," or making stuff up — but they're prioritizing speed and scale instead.
Why it matters: High-profile AI-induced gaffes keep embarrassing users, and the technology's unreliability continues to cloud its adoption.
The big picture: Hallucinations aren't quirks — they're a foundational feature of generative AI that some researchers say will never be fully fixed.
- AI models predict the next word based on patterns in their training data and the prompts a user provides. They're built to try to satisfy users, and if they don't "know" the answer they guess.
- Chatbots' fabrications were a problem when OpenAI CEO Sam Altman introduced ChatGPT in 2022, and the industry continues to remind users that they can't trust every fact a chatbot asserts.
Every week brings painful new evidence that users are not listening to these warnings.
- Last week it was a report from Robert F. Kennedy Jr.'s Health and Human Services Department citing studies that don't exist. Experts found evidence suggesting OpenAI's tools were involved.
- A week earlier, the Chicago Sun-Times published a print supplement with a summer reading list full of real authors, but hallucinated book titles.
- AI legal expert Damien Charlotin tracks legal decisions in which lawyers have used evidence that featured AI hallucinations. His database details more than 30 instances in May 2025. Legal observers fear the total number is likely much higher.
Yes, but: AI makers are locked in fierce competition to top benchmarks, capture users and wow the media. They'd love to tamp down hallucinations, but not at the expense of speed.
- Chatbots could note how confident the language model is in the accuracy of the response, but they have no real incentive for that, Tim Sanders, executive fellow at Harvard Business School and VP at software marketplace G2, told Axios. "That's the dirty little secret. Accuracy costs money. Being helpful drives adoption."
Between the lines: AI companies are making efforts to reduce hallucinations, mainly by trying to fill in gaps in training data.
- Retrieval augmentation generation (RAG) is one process for grounding answers in contextually relevant documents or data.
- RAG connects the model to trusted data so it can retrieve relevant information before generating a response, producing more accurate answers.
AWS offers Amazon Bedrock, a cloud service that allows customers to use various AI providers and responsible AI capabilities (including reducing hallucinations) to build generative AI applications.
- AWS says its Bedrock Guardrails can filter over 75% of hallucinated responses.
- Researchers from Google's DeepMind, along with Stanford University and the University of Illinois at Urbana-Champaign, are working on a Search-Augmented Factuality Evaluator (SAFE), which uses AI to fact-check AI.
- Anthropic offers a guide to help developers limit hallucinations — including allowing the model to answer, "I don't know."
- OpenAI's developer guide includes a section on how much accuracy is "good enough" for production, from both business and technical perspectives.
Researchers inside AI companies have raised alarms about hallucinations, but those warnings aren't always front of mind in organizations hell-bent on a quest for "superintelligence."
- Raising money for building the next big model means companies have to keep making bigger promises about chatbots replacing search engines and AI agents replacing workers. Focusing on the technology's unreliability only undercuts those efforts.
The other side: Some AI researchers insist that the hallucination problem is overblown or at least misunderstood, and that it shouldn't discourage speedy adoption of AI to boost productivity.
- "We should be using [genAI] twice as much as we're using it right now," Sanders told Axios.
- Sanders disputes a recent New York Times article suggesting that as AI models get smarter, their hallucination rates get worse.
- OpenAI's o3 and similar reasoning models are designed to solve more complex problems than regular chatbots. "It reasons. It iterates. It guesses over and over until it gets a satisfactory answer based on the prompt or goal," Sanders told Axios. "It's going to hallucinate more because, frankly, it's going to take more swings at the plate."
- Hallucinations would be less of a concern, Sanders says, if users understood that genAI is designed to make predictions, not verify facts. He urges users to take a "trust, but verify" approach.
The bottom line: Hallucination-free AI may be a forever problem, but for now users can choose the best model for the job and keep a human somewhere in the loop.
2. California chatbot bill defies moratorium push
Efforts in D.C. to stop state-level AI legislation are clashing with a new California chatbot bill.
Why it matters: All over the country, states are advancing and implementing AI safeguards while Congress considers a national moratorium in the reconciliation bill.
- In California, one measure aims to counter tech marketing that suggests chatbots alleviate loneliness and mental health struggles.
- California state Sen. Steve Padilla, the bill's author, told Axios he "can't think of a more irresponsible or stupid thing to do at this juncture" than for Congress to pass the AI moratorium and stop state lawmakers from regulating the tech.
Zoom in: SB-243 would regulate interactions between chatbots and users by prohibiting operators from using the bot to boost user engagement and requiring a clear label to state that the bot is not human.
- The bill also requires an annual report to the Office of Suicide Prevention detailing the number of times suicidal ideation was detected. The office would then post that data on its website.
What they're saying: "We missed an opportunity with the advent of social media and we have a window of opportunity here to act that we should take advantage of," Padilla said.
- Asked if states should sue if the moratorium passes, Padilla said "every option is on the table."
- Sen. Alex Padilla told Axios in a statement that "the proposed 10-year moratorium on artificial intelligence regulation does not belong in a budget bill, and it's ludicrous that Republicans would choose to jam in this provision instead of working with Democrats on AI policy."
The big picture: California has historically led on tech regulation while Congress gets stuck in debate mode — first on privacy and now on AI.
- The state has enacted a wide range of bills addressing everything from digital replicas of individuals to election ads, and has more legislation in the works.
- Yesterday, 260 state lawmakers from both sides of the aisle across all 50 states sent a letter to Congress urging lawmakers to oppose the moratorium.
If you need smart, quick intel on federal tech policy for your job, get Axios Pro Policy.
3. Training data
- Some researchers suspect DeepSeek may have used Google's Gemini, in part, to train its new reasoning model. (TechCrunch)
- Epic Games is incorporating MetaHuman, an AI-powered character creator, into the latest version of its Unreal Engine. (The Verge)
- Here's a smart-brevity guide from Axios' Jim VandeHei and Mike Allen on boosting your AI skills to survive the mass disruptions coming to the job market. (Axios)
4. + This
As y'all know, I'm pretty excited that my Edmonton Oilers are in the National Hockey League finals, but even non-sports lovers can get excited about Friday's Stanley Pup competition. It features adorable shelter dogs representing each team, with awesome stage names, such as Sidney Pawsby and Tyler Barktuzzi.
Thanks to Scott Rosenberg and Megan Morrone for editing this newsletter and Matt Piper for copy editing.
Sign up for Axios AI+



/2025/05/30/1748638076427.gif?w=3840)


