this post was submitted on 21 May 2025
298 points (96.6% liked)

Technology

71799 readers
4115 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

Absolutely needed: to get high efficiency for this beast ... as it gets better, we'll become too dependent.

"all of this growth is for a new technology that’s still finding its footing, and in many applications—education, medical advice, legal analysis—might be the wrong tool for the job,,,"

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 3 points 1 month ago* (last edited 1 month ago) (2 children)

The issue this time around is infrastructure. The current AI Summer depends on massive datacenters with equally massive electrical needs. If companies can't monetize that enough, they'll pull the plug and none of this will be available to general public anymore.

This system can go backwards. Yes, the R&D will still be there after the AI Winter cycle hits, but none of the infrastructure.

[–] [email protected] 3 points 1 month ago

We'll still have models like deepseek, and (hopefully) discount used server hardware

[–] mindbleach 2 points 1 month ago (1 children)

Local models aren't going anywhere. You can hoard them if you're worried. People freely tweak them to do some weird shit. There's whole independent websites for sharing these tweaks.

The only infrastructure that's at risk is the part where corporations burn billions of dollars hoping the really really big models will turn a profit somehow. The power and capability of small models is basically whatever the big boys did a year ago - and a ton of money has been spent on improving the efficiency of training. Tightening the belt is not going to make things worse. It's certainly not going to make this whole new kind of software un-happen.

[–] [email protected] 0 points 1 month ago (1 children)

Local models aren't what's driving the current AI Summer. No billionaire will be quoted in the New York Times saying how great local models are.

[–] mindbleach 1 points 1 month ago

Why on Earth is that your metric?

The money addicts desperately want big models to win because that means less competition and more subscription revenue. But that's three delusions in a row. Mainframes are basically never a sensible business model. Models costing mere tens of millions already challenge billion-dollar projects. And as you say, infrastructure is expensive, so most people want software that runs within arm's reach. There's every reason these cloud services bleed money.

Local models are what's making all the weird shit you see online. Certainly the porn is not made by typing stuff into Facebook. And again: whatever cutting-edge advances happen in moonshot-priced, datacenter-sized models, they're soon replicated on any recent desktop. If that same blank-check experimentation had only bothered with small models, do you think they'd be further behind or further ahead?

What's driving this AI Summer is a horde of ultranerds, funded by these rich idiots lying to each other. A lot of them will still be doing this if the money dries up. A lot of them already do it for free. See aforementioned tweaking and sharing. If that has to expand to training new models from scratch, expect them to get even smaller, and rely on experimentation more than brute force. Like BitNet reducing all weights to trinary. If it still takes a mountain of GPUs... well, SETI@home couldn't promise people interactive anime waifus.