this post was submitted on 12 Jul 2023
274 points (97.6% liked)

Technology

58011 readers
3117 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Users of OpenAI's GPT-4 are complaining that the AI model is performing worse lately. Industry insiders say a redesign of GPT-4 could be to blame.

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 86 points 1 year ago* (last edited 1 year ago) (5 children)

The model has become inbred because it’s now impossible to scrape the web without AI content getting ingested, which is full of “hallucinations” and other weird artifacts. The last opportunity to get “uncontaminated” training data was sometime in mid 2022.

Not to say that it’s causing this particular problem, but this issue will emerge eventually. Garbage in = garbage out. Eventually GPT-19 will grow a mighty Habsburg chin.

[–] [email protected] 28 points 1 year ago* (last edited 1 year ago) (3 children)

Maybe not yet, but...

  • Spez will turn Reddit into a bot farm and sell this as training data
  • Musk turns Twitter into a bigoted cesspool and will sell this as training data, which will subsequently be flagged for low quality (also: a botfarm)
  • Threads is a corporate ad dashboard (and we already know how easy it is to GPT copy) and Zuck will sell this as training data
  • Facebook is either dead or only good for boomers and Poles
  • blogs are dead
  • Fediverse is out there waiting to be scraped but possibly too small to sustain a big model

We'te getting there, hopefully.

[–] [email protected] 7 points 1 year ago (1 children)
[–] [email protected] 8 points 1 year ago (1 children)
[–] damnYouSun 10 points 1 year ago

Also We'te, which I believe is a Klingon name.

load more comments (2 replies)
[–] [email protected] 4 points 1 year ago

That hasn't happened yet. Most likely they quantized GPT-4 more. It's still based on the same training data.

[–] [email protected] 4 points 1 year ago (1 children)

I suspect future models are going to have to put some more focus on learning using techniques more like what humans use, and on cognition.

Like, compared to a human these language models need very large quantities of text input. When humans are first learning language they get lots of visual input along with language input, and can test their understanding with trial-and-error feedback from other intelligent actors. I wonder if perhaps those factors greatly increase the rate at which understanding develops.

Also, humans tend to cogitate on inputs while ingesting them during learning. So if the information in new inputs disagrees with current understanding, those inputs are less likely to affect current understanding (there's a whole 'how to change your mind' thing here that is necessary for people to use, but if we're training a model on curated data that's probably less important for early model training).

I don't know details of how model training works, but it would be interesting to know if anyone is using a progressive learning technique where the model that is being trained is used to judge new training data before it is used as a training input to update the model's weights. That would be kind of like how children learn by starting with very simple words and syntax and building up conceptual understanding gradually. I'd assume so, since it's an obvious idea, but I haven't heard about it.

load more comments (1 replies)
[–] [email protected] 33 points 1 year ago* (last edited 1 year ago) (1 children)

The lobotomies will continue. Free models will keep getting better.

[–] [email protected] 11 points 1 year ago* (last edited 1 year ago)

The chatgpt people are really paranoid. Gpt-3 is so good at not halucinating that it often cant, even if it needs to do so to accomplish a task. Fearing the ai will confidently give the wrong answer.

[–] [email protected] 29 points 1 year ago (2 children)

Not the first time OpenAI has done this. DALLE2 used to be the best AI art program in the world. Then OpenAI decided that they didn't want to get sued by celebrities, so they made it so that if a face came out that resembled a celebrity, it would be distorted. But every face kind of looks like someone famous. Ta da! Now DALLE2 can't do faces.

Want a crane shot areal image of a teen couple in a corvette driving off into the sunset? Well, you are now banned for life from the DALLE2 service, because DALLE2 produced an image of a 'shot teen' and that violates it's terms of service.

load more comments (2 replies)
[–] hoshikarakitaridia 26 points 1 year ago (1 children)

The only way in mind this dumbing down happens is by fumbling with the model. So that's the one thing we can be sure: the AI is most definitely changed while publicly staying "ChatGPT 4". I assume they are either using clipping or token limitations to split the server load but fucking up the result, or they are purposely dumbing it down to capitalise on it later by introducing other pay models like ppl already mentioned.

Either way they are shooting themselves in the foot because a bunch of ppl will unsubscribe either out of spite for the change or because it's just not worth it anymore for them.

[–] [email protected] 25 points 1 year ago

AI taking a running leap at enshittification.

[–] [email protected] 20 points 1 year ago (1 children)

Some people have been saying that since the beginning while some haven’t noticed this “decline”. It seems very subjective.

[–] [email protected] 17 points 1 year ago

Honestly as a daily user I think it's a combination of it getting worse at understanding vague prompts and people bumbing up against edge cases more. I would suspect the former is due to things like prompt hardening but can only speculate, while the latter isn't hard to imagine just from frequent use.

[–] [email protected] 9 points 1 year ago (1 children)
[–] [email protected] 15 points 1 year ago (2 children)

You mean "I was right* or "i wrote*"?

[–] [email protected] 13 points 1 year ago

No no, he used to work as a wright. Built ships and shit.

load more comments (1 replies)
[–] [email protected] 9 points 1 year ago

You know how we have pre-bomb steel? We'll have pre-GPT data sets.

[–] [email protected] 6 points 1 year ago

Yeah, when I first started using GPT-4, I didn't notice any hallucinations. Now I'm getting them all the time. Disappointing.

[–] [email protected] 4 points 1 year ago

Just like most people after they achieve success.

load more comments
view more: next ›