this post was submitted on 10 Apr 2024
1285 points (99.0% liked)
Programmer Humor
19804 readers
227 users here now
Welcome to Programmer Humor!
This is a place where you can post jokes, memes, humor, etc. related to programming!
For sharing awful code theres also Programming Horror.
Rules
- Keep content in english
- No advertisements
- Posts must be related to programming or programmer topics
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Yeah I read one of the papers that talked about this. Essentially putting AGI data into a training set will pollute it, and cause it to just fall apart. Most LLMs especially are going to be a ton of fun as there were absolutely no rules about what to do, and bots and spammers immediately used it everywhere on the internet. And the only solution is to.... write a model to detect it. Which then they'll make models that bypass that, and there will just be no way to keep the dataset clean.
The hype of AI is warranted - but also way overblown. Hype from actual developers and seeing what it can do when it's tasked with doing something appropriate? Blown away. Just honestly blown away. However hearing what businesses want to do with it, the crazy shit like "We'll fire everyone and just let AI do it!" Impossible. At least with the current generation of models. Those people remind me of the crypto bros saying it's going to revolutionize everything. It might, but you need to actually understand the tech and it's limitations first.
Building my own training set is something I would certainly want to do eventually. Ive been messing with Mistral Instruct using GPT4ALL and its genuinely impressive how quick my 2060 can hallucinate relatively accurate information, but its also evident of limitations. IE I tell it I do not want to use AWS or another cloud hosting service, it will just return a list of suggested services not including AWS. Most certainly a limit of its training data but still impressive.
Anyone suggesting to use LLMs to manage people or resources are better off flipping a coin on every thought, more than likely companies who are insistent on it will go belly up soon enough
You're describing an arms race, which makes me wonder if that's part of the path to AGI. Ultimately the only way to truly detect a fake is to compare it to reality, and the only way to train a model to understand whether it is looking at reality or a generated image is to teach it to understand context and meaning, and that's basically the ballgame at that point. That's a qualitative shift, and in that scenario we get there with opposing groups each pursuing their own ends, not with a single group intentionally making AGI.