this post was submitted on 22 Dec 2024
523 points (96.1% liked)

Technology

60292 readers
3330 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 77 points 2 weeks ago* (last edited 2 weeks ago) (7 children)

There is this seeming need to discredit AI from some people that goes overboard. Some friends and family who have never really used LLMs outside of Google search feel compelled to tell me how bad it is.

But generative AIs are really good at tasks I wouldn't have imagined a computer doing just a few year ago. Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow. It's not just hype.

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

[–] [email protected] 71 points 2 weeks ago (1 children)

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

This is literally the hype. This is the hype that is dying and needs to die. Because generative AI is a tool with fairly specific uses. But it is being marketed by literally everyone who has it as General AI that can "DO ALL THE THINGS!" which it's not and never will be.

[–] [email protected] 11 points 2 weeks ago (1 children)

The obsession with replacing workers with AI isn't going to die. It's too late. The large financial company that I work for has been obsessively tracking hours saved in developer time with GitHub Copilot. I'm an older developer and I was warned this week that my job will be eliminated soon.

[–] [email protected] 4 points 2 weeks ago* (last edited 2 weeks ago) (1 children)

The large financial company that I work for

So the company that is obsessed with money that you work for has discovered a way to (they think) make more money by getting rid of you and you're surprised by this?

At least you've been forewarned. Take the opportunity to abandon ship. Don't be the last one standing when the music stops.

[–] [email protected] 5 points 2 weeks ago* (last edited 2 weeks ago)

I never said that I was surprised. I just wanted to point out that many companies like my own are already making significant changes to how they hire and fire. They need to justify their large investment in AI even though we know the tech isn't there yet.

[–] [email protected] 39 points 2 weeks ago (10 children)

Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow

Like which one? Because it's now 2 years we have chatGPT and already quite a lot of (good?) models. Which shakeup do you think is happening or going to happen?

[–] [email protected] 6 points 2 weeks ago (5 children)

Computer programming has radically changed. Huge help having llm auto complete and chat built in. IDEs like Cursor and Windsurf.

I’ve been a developer for 35 years. This is shaking it up as much as the internet did.

[–] [email protected] 35 points 2 weeks ago* (last edited 2 weeks ago) (5 children)

I quit my previous job in part because I couldn't deal with the influx of terrible, unreliable, dangerous, bloated, nonsensical, not even working code that was suddenly pushed into one of the projects I was working on. That project is now completely dead, they froze it on some arbitrary version.
When junior dev makes a mistake, you can explain it to them and they will not make it again. When they use llm to make a mistake, there is nothing to explain to anyone.
I compare this shake more to an earthquake than to anything positive you can associate with shaking.

[–] [email protected] 2 points 2 weeks ago (1 children)

And so, the problem wasn't the ai/llm, it was the person who said "looks good" without even looking at the generated code, and then the person who read that pull request and said, again without reading the code, "lgtm".

If you have good policies then it doesn't matter how many bad practice's are used, it still won't be merged.

The only overhead is that you have to read all the requests but if it's an internal project then telling everyone to read and understand their code shouldn't be the issue.

[–] [email protected] 12 points 2 weeks ago

The problem here is that a lot of the time looking for hidden problem is harder than writing good code from scratch. And you will always be at a danger that llm snuck some sneaky undefined behaviour past you. There is a whole plethora of standards, conventions, and good practices that help humans to avoid it, which llm can ignore at any random point.
So you're either not spending enough time on review or missing whole lot of bullshit. In my experience, in my field, right now, this review time is more time consuming and more painful than avoiding it in the first place.
Don't underestimate how degrading and energy sucking it is for a professional to spend most of the working time sitting through autogenerated garbage, and how inefficient it is.

load more comments (4 replies)
[–] [email protected] 34 points 2 weeks ago (15 children)

I hardly see it changed to be honest. I work in the field too and I can imagine LLMs being good at producing decent boilerplate straight out of documentation, but nothing more complex than that.

I often use LLMs to work on my personal projects and - for example - often Claude or ChatGPT 4o spit out programs that don't compile, use inexistent functions, are bloated etc. Possibly for languages with more training (like Python) they do better, but I can't see it as a "radical change" and more like a well configured snippet plugin and auto complete feature.

LLMs can't count, can't analyze novel problems (by definition) and provide innovative solutions...why would they radically change programming?

load more comments (15 replies)
[–] [email protected] 2 points 1 week ago (1 children)

@[email protected] 1 year. Let me know about the seachange of new 10x transform based programmers that have automated me out of a job.

[–] [email protected] 2 points 1 week ago

@horse_battery_staple Ok, I will remind you on Friday Dec 26, 2025 at 7:49 AM PST.

load more comments (2 replies)
load more comments (9 replies)
[–] [email protected] 24 points 2 weeks ago (23 children)

Computers have always been good at pattern recognition. This isn't new. LLM are not a type of actual AI. They are programs capable of recognizing patterns and Loosely reproducing them in semi randomized ways. The reason these so-called generative AI Solutions have trouble generating the right number of fingers. Is not only because they have no idea how many fingers a person is supposed to have. They have no idea what a finger is.

The same goes for code completion. They will just generate something that fills the pattern they're told to look for. It doesn't matter if it's right or wrong. Because they have no concept of what is right or wrong Beyond fitting the pattern. Not to mention that we've had code completion software for over a decade at this point. Llms do it less efficiently and less reliably. The only upside of them is that sometimes they can recognize and suggest a pattern that those programming the other coding helpers might have missed. Outside of that. Such as generating act like whole blocks of code or even entire programs. You can't even get an llm to reliably spit out a hello world program.

[–] [email protected] 5 points 2 weeks ago (2 children)

I never know what to think when I come across a comment like this one—which does describe, even if only at a surface level, how an LLM works—with 50% downvotes. Like, are people angry at reality, is that it?

[–] [email protected] 15 points 2 weeks ago (1 children)

With as much misinformation that's being spread about regarding LLMs. It would only lose more people's comprehension to go into anything more than a generalization.

The problem is people are being sold AGI. But chat GPT and all these other tools don't even remotely qualify for that. They're really nothing more than a glorified Alice chatbot system on steroids. The one neat new trick to all this is that they've automated the training a bit. But these llms have no more comprehension of their output or the input they were given than something like the old Alice chatbot.

These tools have been described as artificial intelligence to layman for decades at this point. It makes it really hard to change that calcified opinion. People would rather believe that it's some magical thing not just probability and maths.

[–] [email protected] 8 points 2 weeks ago

They are bullshit machines, trained to output something that users think is the right output.

[–] Naz 2 points 2 weeks ago (1 children)

Downvoting someone on the Internet is easier than tangentially modifying reality in a measurable way

load more comments (1 replies)
[–] [email protected] 3 points 2 weeks ago (3 children)

"It's part of the history of the field of artificial intelligence that every time somebody figured out how to make a computer do something—play good checkers, solve simple but relatively informal problems—there was a chorus of critics to say, 'that's not thinking'"
-Pamela McCorduck

"AI is whatever hasn't been done yet."
- Larry Tesler

That's the curse of the AI Effect.
Nothing will ever be "an actual AI" until we cross the barrier to an actual human-like general artificial intelligence like Cortana from Halo, and even then people will claim it isn't actually intelligent.

[–] [email protected] 8 points 2 weeks ago (6 children)

Well at least until those who study intelligence and self-awareness actually come up with a comprehensive definition for it. Something we don't even have currently. Which makes the situation even more silly. The people selling LLMs and AGNs as artificial intelligence are the PT Barnum of the modern era. This way to the egress folks come see the magnificent egress!

load more comments (6 replies)
[–] [email protected] 8 points 2 weeks ago (1 children)

I mean, I think intelligence requires the ability to integrate new information into one's knowledge base. LLMs can't do that, they have to be trained on a fixed corpus.

Also, LLMs have a pretty shit-tastic track record of being able to differentiate correct data from bullshit, which is a pretty essential facet of intelligence IMO

[–] [email protected] 8 points 2 weeks ago

LLMs have a perfect track record of doing exactly what they were designed to, take an input and create a plausible output that looks like it was written by a human. They just completely lack the part in the middle that properly understands what it gets as the input and makes sure the output is factually correct, because if it did have that then it wouldn't be an LLM any more, it would be an AGI.
The "artificial" in AI does also stand for the meaning of "fake" - something that looks and feels like it is intelligent, but actually isn't.

load more comments (1 replies)
load more comments (21 replies)
[–] [email protected] 17 points 2 weeks ago

This is easy to say about the output of AIs.... if you don't check their work.

Alas, checking for accuracy these days seems to be considered old fogey stuff.

[–] [email protected] 11 points 2 weeks ago* (last edited 2 weeks ago) (1 children)

Goldman Sachs, quote from the article:

“AI technology is exceptionally expensive, and to justify those costs, the technology must be able to solve complex problems, which it isn’t designed to do.”

Generative AI can indeed do impressive things from a technical standpoint, but not enough revenue has been generated so far to offset the enormous costs. Like for other technologies, It might just take time (remember how many billions Amazon burned before turning into a cash-generating machine? And Uber has also just started turning some profit) + a great deal of enshittification once more people and companies are dependent. Or it might just be a bubble.

As humans we're not great at predicting these things including of course me. My personal prediction? A few companies will make money, especially the ones that start selling AI as a service at increasingly high costs, many others will fail and both AI enthusiasts and detractors will claim they were right all along.

load more comments (1 replies)
[–] [email protected] 3 points 2 weeks ago (4 children)

See now, I would prefer AI in my toaster. It should be able to learn to adjust the cook time to what I want no matter what type of bread I put in it. Though is that realky AI? It could be. Same with my fridge. Learn what gets used and what doesn't. Then give my wife the numbers on that damn clear box of salad she buys at costco everytime, which take up a ton of space and always goes bad before she eats even 5% of it. These would be practical benefits to the crap that is day to day life. And far more impactful then search results I can't trust.

[–] [email protected] 10 points 2 weeks ago* (last edited 2 weeks ago) (1 children)

There's a good point here that like about 80% of what we're calling AI right now... isn't even AI or even LLM. It's just.... algorithm, code, plain old math. I'm pretty sure someone is going to refer to a calculator as AI soon. "Wow, it knows math! Just like a person! Amazing technology!"

(That's putting aside the very question of whether LLMs should even qualify as AIs at all.)

[–] [email protected] 4 points 2 weeks ago (7 children)

In my professional experience, AI seems to be just a faster way to generate an algorithm that is really hard to debug. Though I am dev-ops/sre so I am not as deep in it as the devs.

[–] [email protected] 2 points 1 week ago

I remined of the time researchers used an evolutionary algorithm to devise a circuit that would emit a tone on certain audio inputs and not on others. They examined the resulting circuit and found an extra vestigial bit, but when they cut it off, the chip stopped working. So they re-enabled it. Then they wanted to show off their research at a panel, and at the panel it completely failed. Dismayed they brought it back to their lab to figure out why it stopped working, and it suddenly started working fine.

After a LOT of troubleshooting they eventually discovered that the circuit was generating the tone by using the extra vestigial bit as an antenna that picked up emissions from a CRT in the lab and downconverted it to the desired tone frequency. Turn of the antenna, no signal. Take the chip away from that CRT, no signal.

That's what I expect LLMs will make. Complex, arcane spaghetti stuff that works but if you look at it funny it won't work anymore, and nobody knows how it works at all.

load more comments (6 replies)
[–] [email protected] 6 points 2 weeks ago (3 children)

You better believe that AI-powered toaster would only accept authorized bread from a bakery that paid top dollar to the company that makes them. To ensure the best quality possible and save you from inferior toast, of course.

[–] [email protected] 3 points 2 weeks ago

Lol, enshitification should at least take a few months... I hope.

[–] [email protected] 2 points 2 weeks ago

or you go to make some toast and it spends 15 minutes downloading "updates" before you can use it

load more comments (1 replies)
[–] [email protected] 3 points 2 weeks ago (1 children)
[–] [email protected] 3 points 2 weeks ago

I was so hoping that was toasty the toaster! Waffles? How about a bagel?

[–] [email protected] 3 points 2 weeks ago* (last edited 2 weeks ago) (2 children)

I agree with your wife: there’s always an aspirational salad in the fridge. For most foods, I’m pretty good at not buying stuff we won’t eat, but we always should eat more veggies. I don’t know how to persuade us to eat more veggies, but step 1 is availability. Like that Reddit meme

  1. Availability
  2. ???
  3. Profit by improved health
load more comments (2 replies)
[–] [email protected] 2 points 2 weeks ago

Like what outcome?

I have seen gains on cell detection, but it's "just" a bit better.