this post was submitted on 13 Aug 2023
1072 points (96.0% liked)
Technology
60123 readers
2725 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I've started going down this rabbit hole. The takeaway is that if we define intelligence as "ability to solve problems", we've already created artificial intelligence. It's not flawless, but it's remarkable.
There's the concept of Artificial General Intelligence (AGI) or Artificial Consciousness which people are somewhat obsessed with, that we'll create an artificial mind that thinks like a human mind does.
But that's not really how we do things. Think about how we walk, and then look at a bicycle. A car. A train. A plane. The things we make look and work nothing like we do, and they do the things we do significantly better than we do them.
I expect AI to be a very similar monster.
If you're curious about this kind of conversation I'd highly recommend looking for books or podcasts by Joscha Bach, he did 3 amazing episodes with Lex.
Current "AI" doesn't solve problems. It doesn't understand context. It can't see fingers and say "those are fingers, make sure there's only five". It can't tell the difference between a truth and a lie. It can't say "well that can't be right!" It just regurgitates an amalgamation of things humans have showed it or said, with zero understanding. "Consciousness" and certainly "sapience" aren't really relevant factors here.
You’re confusing AI with AGI. AGI is the ultimate goal of AI research. AI are all the steps along the way. Step by step, AI researchers figure out how to make computers replicate human capabilities. AGI is when we have an AI that has basically replicated all human capabilities. That’s when it’s no longer bounded by a particular problem.
You can use the more specific terms “weak AI” or “narrow AI” if you prefer.
Generative AI is just another step in the way. Just like how the emergence of deep learning was one step some years ago. It can clearly produce stuff that previously only humans could make, which in this case is convincing texts and pictures from arbitrary prompts. It’s accurate to call it AI (or weak AI).
Yeah, well, "AGI" is not the end result of this generative crap. You're gonna have to start over with something different one way or another. This simply is not the way.
So...it acts like a human?
No? There's a whole lot more to being human than being able to separate one object from another and identify it, recognize that object, and say "my database says that there should only be two of these in this context". Current "AI" can't even do that much-- especially not with art.
Do you know what "sapience" means, by the way?