med-mastodon.com is one of the many independent Mastodon servers you can use to participate in the fediverse.
Medical community on Mastodon

Administered by:

Server stats:

363
active users

#aihallucinations

0 posts0 participants0 posts today

Business Insider: Protesters accuse Google of breaking its promises on AI safety: ‘AI companies are less regulated than sandwich shops’. “The participants of this faux-production were protesters from PauseAI, an activist group concerned that tech companies are racing into AI with little regard for safety. On Monday, the group congregated near King’s Cross station to demand that Google be more […]

https://rbfirehose.com/2025/07/02/protesters-accuse-google-of-breaking-its-promises-on-ai-safety-ai-companies-are-less-regulated-than-sandwich-shops-business-insider/

ResearchBuzz: Firehose | Individual posts from ResearchBuzz · Protesters accuse Google of breaking its promises on AI safety: ‘AI companies are less regulated than sandwich shops’ (Business Insider) | ResearchBuzz: Firehose
More from ResearchBuzz: Firehose

Ars Technica: Unlicensed law clerk fired after ChatGPT hallucinations found in filing. “Last month, a recent law school graduate lost his job after using ChatGPT to help draft a court filing that ended up being riddled with errors. The consequences arrived after a court in Utah ordered sanctions after the filing included the first fake citation ever discovered in the state hallucinated by […]

https://rbfirehose.com/2025/06/05/ars-technica-unlicensed-law-clerk-fired-after-chatgpt-hallucinations-found-in-filing/

ResearchBuzz: Firehose | Individual posts from ResearchBuzz · Ars Technica: Unlicensed law clerk fired after ChatGPT hallucinations found in filing | ResearchBuzz: Firehose
More from ResearchBuzz: Firehose

@Remittancegirl

#ChicagoSunTimes hit by #AIHallucinations generating articles for its content

Via @TheGuardian

"Others on social media have pointed out that 👉the use of AI appears to be found throughout the pages of the Chicago Sun-Times summer 2025 section.👈

👉"Chicago Sun-Times confirms AI was used to create reading list of books that don’t exist👈

Outlet calls story, created by freelancer working with one of the newpaper’s content partner[s], a ‘learning moment’"
theguardian.com/us-news/2025/m

The Guardian · Chicago Sun-Times confirms AI was used to create reading list of books that don’t existBy Marina Dunbar

People complain about #LLMs for "hallucinating." In this context, a hallucination is something that looks like a fact but is actually completely fictitious. (It's a terrible name.)

But here's the thing: every day I talk to humans. The vast majority of humans that I interact with say things that look like facts but are actually completely fictitious.

FWIW, I get many more daily hallucinations from *people* than I do machines.

¯⁠\⁠_⁠(⁠ツ⁠)⁠_⁠/⁠¯

Financial Times: Insurers launch cover for losses caused by AI chatbot errors. “Insurers at Lloyd’s of London have launched a product to cover companies for losses caused by malfunctioning artificial intelligence tools, as the sector aims to profit from concerns about the risk of costly hallucinations and errors by chatbots.”

https://rbfirehose.com/2025/05/15/financial-times-insurers-launch-cover-for-losses-caused-by-ai-chatbot-errors/

ResearchBuzz: Firehose | Individual posts from ResearchBuzz · Financial Times: Insurers launch cover for losses caused by AI chatbot errors | ResearchBuzz: Firehose
More from ResearchBuzz: Firehose

🧠 AI hallucinations are not just technical flaws but a governance risk.

Lloyd’s of London now offers insurance against AI-generated errors, including hallucinations, via a policy developed by Armilla AI.

As AI becomes central to decision-making, legal and ethical safeguards are essential.

🌐 Source: business-standard.com/industry

I tried #JanAI (an app that lets you run #LLM models on local computers so you can have a private #AI assistant).

Was thinking it could be interesting to experiment with this for prose text summaries/analysis.

However, Jan's document parsing is "experimental". Can't attach plain text files (why). Could drag in PDFs. Fine, whatever. "Can you do chapter by chapter summaries of this manuscript of mine" basically did some really weird #aihallucinations based on things that may have been referenced in the text. The overall plot summary was interesting, too bad it was had little to no resemblance to the novel I wrote.

I couldn't resist seeing if, since this is a locally running AI, it could talk about naughty sexy things. "Can't generate anything explicit", the default Llama said. Fine, whatever. Downloaded some "NSFW" model. It generated janky garbage and said "f*ck" a lot, and I do mean literally "f*ck", which messed up the Markdown formatting completely.

This technology is improving but still has a way to go

Ars Technica: Why do LLMs make stuff up? New research peers under the hood.. “From a human perspective, it can be hard to understand why these models don’t simply say ‘I don’t know’ instead of making up some plausible-sounding nonsense. Now, new research from Anthropic is exposing at least some of the inner neural network “circuitry” that helps an LLM decide when to take a stab at a (perhaps […]

https://rbfirehose.com/2025/03/29/ars-technica-why-do-llms-make-stuff-up-new-research-peers-under-the-hood/

Columbia Journalism Review: We Compared Eight AI Search Engines. They’re All Bad at Citing News.. “Overall, the chatbots often failed to retrieve the correct articles. Collectively, they provided incorrect answers to more than 60 percent of queries. Across different platforms, the level of inaccuracy varied, with Perplexity answering 37 percent of the queries incorrectly, while Grok 3 had a […]

https://rbfirehose.com/2025/03/06/columbia-journalism-review-we-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news/

ResearchBuzz: Firehose | Individual posts from ResearchBuzz · Columbia Journalism Review: We Compared Eight AI Search Engines. They’re All Bad at Citing News. | ResearchBuzz: Firehose
More from ResearchBuzz: Firehose

Interesting report this morning of a BC couple that tried to use AI to assist in finding precedents for a civil case regarding their condo.

The tribunal found almost all the citations were false.

> The Geismayrs referenced 10 decisions as part of their argument to allow the unit changes and listed Microsoft Copilot as the source of the cases.

> Nine of the cases were fake, according to tribunal member Peter Mennie.

> "I find it likely that these cases are 'hallucinations' where artificial intelligence generates false or misleading results," Mennie wrote in his Feb. 14 ruling.

cbc.ca/news/canada/british-col

#AI #ArtificialIntelligence #AIHallucinations
#generativeAI #ChatBots #JusticeSystem
#BC #Canada #CoPilot #Microsoft

www.cbc.caB.C. couple referenced non-existent, AI-generated court rulings in condo dispute, tribunal finds

BBC: AI chatbots unable to accurately summarise news, BBC finds. “Four major artificial intelligence (AI) chatbots are inaccurately summarising news stories, according to research carried out by the BBC. The BBC gave OpenAI’s ChatGPT, Microsoft’s Copilot, Google’s Gemini and Perplexity AI content from the BBC website then asked them questions about the news.”

https://rbfirehose.com/2025/02/12/bbc-ai-chatbots-unable-to-accurately-summarise-news-bbc-finds/

Replied in thread

@OliverUv

basically every comp scientist who doesn't work for AI corps hate [AI being pushed by management as a panacea].

At an IBM tech conference for the purpose of learning about AI for the company (and giving a Zowe presentation), one of the first panels I attended was on AI hallucination.

Boy howdy! This was like the films they showed in health class about people not wearing a seatbelt and getting into an accident!

I learned about the disaster of using AI without curation, and about using AI as a very expensive component that did have applications. Yes, you can ethically source your LLM for $$$$. As for the energy cost... [coughs.]

Personally, I dislike how lay people misunderstand the tool. There is definitely a lot of misleading hard sell going on in the industry. I work when I can to disabuse them of their belief. I can't say I hate AI. What I hate is it is the current "hammer" and that I seem to be one of the "nails."

@freeagent @vextaur