this post was submitted on 27 May 2024
1100 points (98.0% liked)
Technology
59646 readers
2677 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The snake ate it's tail before it's fully grown. The AI inbreeding might be already too far integrated, causing all sorts of Mumbo-Jumbo. Also they have layers of censorship, which effect the results. The same that happened to chatgpt, the more filters they added, the more it confused the result. We don't even know if the hallucinations are fixable, AI is just guessing after all, who knows if AI will ever understand 1+1=2, by calculating, instead of going by probability.
Hallucinations aren't fixable, as LLMs don't have any actual "intelligence". They can't test/evaluate things to determine if what they say is true, so there is no way to correct it. At the end of the day, they are intermixing all the data they "know" to give the best answer, without being able to test their answers LLMs can't vet what they say.
Even saying they're guessing is wrong, as that implies intention. LLMs aren't trying to give an answer, let alone a correct answer. They just put words together.