this post was submitted on 21 May 2025
298 points (96.6% liked)

Technology

70415 readers
3143 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

Absolutely needed: to get high efficiency for this beast ... as it gets better, we'll become too dependent.

"all of this growth is for a new technology that’s still finding its footing, and in many applications—education, medical advice, legal analysis—might be the wrong tool for the job,,,"

you are viewing a single comment's thread
view the rest of the comments
[–] mindbleach 2 points 1 week ago (1 children)

Local models aren't going anywhere. You can hoard them if you're worried. People freely tweak them to do some weird shit. There's whole independent websites for sharing these tweaks.

The only infrastructure that's at risk is the part where corporations burn billions of dollars hoping the really really big models will turn a profit somehow. The power and capability of small models is basically whatever the big boys did a year ago - and a ton of money has been spent on improving the efficiency of training. Tightening the belt is not going to make things worse. It's certainly not going to make this whole new kind of software un-happen.

[–] [email protected] 0 points 1 week ago (1 children)

Local models aren't what's driving the current AI Summer. No billionaire will be quoted in the New York Times saying how great local models are.

[–] mindbleach 1 points 1 week ago

Why on Earth is that your metric?

The money addicts desperately want big models to win because that means less competition and more subscription revenue. But that's three delusions in a row. Mainframes are basically never a sensible business model. Models costing mere tens of millions already challenge billion-dollar projects. And as you say, infrastructure is expensive, so most people want software that runs within arm's reach. There's every reason these cloud services bleed money.

Local models are what's making all the weird shit you see online. Certainly the porn is not made by typing stuff into Facebook. And again: whatever cutting-edge advances happen in moonshot-priced, datacenter-sized models, they're soon replicated on any recent desktop. If that same blank-check experimentation had only bothered with small models, do you think they'd be further behind or further ahead?

What's driving this AI Summer is a horde of ultranerds, funded by these rich idiots lying to each other. A lot of them will still be doing this if the money dries up. A lot of them already do it for free. See aforementioned tweaking and sharing. If that has to expand to training new models from scratch, expect them to get even smaller, and rely on experimentation more than brute force. Like BitNet reducing all weights to trinary. If it still takes a mountain of GPUs... well, SETI@home couldn't promise people interactive anime waifus.