this post was submitted on 30 May 2025
227 points (99.1% liked)

Hacker News

1446 readers
329 users here now

Posts from the RSS Feed of HackerNews.

The feed sometimes contains ads and posts that have been removed by the mod team at HN.

founded 8 months ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] vatlark@lemmy.world 15 points 4 days ago (7 children)

Other than websites, I can never get a citation that has any connection to reality from an LLM. They are either poisoning the training data or it's just too much detail for the model to capture.

[–] Trimatrix@lemmy.world -4 points 4 days ago (5 children)

My theory? Its the abstract that trips it up. I don’t know much about neural networks or RAG searching, but the abstract is information dense in a short amount of space. My guess, an LLM scans through a document, finds areas of the document that has the largest quantity of information relevant to the prompt and then branches out until its confident enough in answering the question. While great for most written text, I think the LLM starts with the abstract, gets way too much confidence that it knows enough and then begins spouting bullshit. In my opinion it explains why it creates fictitious citations. Probably sees a citation in an abstract and then conjures up a title from the information in the abstract.

My only evidence is that I try using LLMs to find information quickly in very dense datasheets for electronic components and most times it will rattle off information found in the component overview section rather than telling me about information found in the application, or electronic characteristics section.

[–] AlecSadler@sh.itjust.works 6 points 4 days ago (4 children)

Are you running your own programmatic LLMs? There is a thing called temperature, and it is typically more lenient for public facing LLMs. But leverage that same LLM via APIs and you can adjust the temperature and reduce or eliminate hallucinations.

Ultimately, a little variance (creativity) is somewhat good and passing it through levels of agentic validations can help catch hallucinations and dial in the final results.

That said, I doubt the WH did this, they probably just dumped shit into some crappy public-facing ChatGPT model.

[–] vatlark@lemmy.world 1 points 4 days ago

Interesting, I haven't played much with the APIs, I just started messing with running some locally with ollama

load more comments (3 replies)
load more comments (3 replies)
load more comments (4 replies)