this post was submitted on 26 Feb 2025
481 points (97.1% liked)
Technology
63375 readers
4568 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I'm not. Apologies if I was unclear, but I was specifically referencing the fact that you were saying AI was going to accelerate to the point that it replaces human labor, and I was simply stating that I would prefer a world in which human labor is not required for humans to survive, and we can simply pursue other passions, if such a world where to exist, as a result of what you claim is happening with AI. You claimed AI will get so good it replaces all the jobs. Cool, I would enjoy that, because I don't believe that jobs are what gives human lives meaning, and thus am fine if people are free to do other things with their lives.
The automation of labor is not even remotely comparable to the creation of a technology who's explicit, sole purpose is to cause the largest amount of destruction possible.
Could there hypothetically be an AI model far in the future, once we secure enough computing power, and develop the right architecture, that technically meets the definition of AGI, (however subjective it may be) that then decides to do something to harm humans? I suppose, but that's simply not looking to be likely in any way, (and I'd love if you could actually show any data/evidence proving otherwise instead of saying "it just is" when claiming it's more dangerous) and anyone claiming we're getting close (e.g. Sam Altman) just simply has a vested financial interest in saying that AI development is moving quicker and at a higher scale than it actually is.
It's called having a disagreement and refuting your points. Just because someone doesn't instantly agree with you doesn't mean that I'm automatically mistaken. You're not the sole arbiter of truth. Judging from how you, three times now, have assumed that I must be secretly suppressing the fact that AI is actually going to do more damage than nuclear bombs, just because I disagree with you, it's clear that you are the one making post-hoc justifications here.
You are automatically assuming that because I disagree, I actually don't disagree, and must secretly believe the same thing as you, but am just covering it up. Do not approach arguments from the assumption that the other person involved is just feigning disagreement, or you will never be capable of even considering a view other than the one you currently hold.
The fact you'd even consider me possibly using AI to write a comment is ridiculous. Why would I do that? What would I gain? I'm here to articulate my views, not my views but only kind of, without any of my personal context, run through a statistical probability machine.