this post was submitted on 18 Jun 2023
2 points (100.0% liked)
Singularity | Artificial Intelligence (ai), Technology & Futurology
96 readers
1 users here now
About:
This sublemmy is a place for sharing news and discussions about artificial intelligence, core developments of humanity's technology and societal changes that come with them. Basically futurology sublemmy centered around ai but not limited to ai only.
Rules:
- Posts that don't follow the rules and don't comply with them after being pointed out that they break the rules will be deleted no matter how much engagement they got and then reposted by me in a way that follows the rules. I'm going to wait for max 2 days for the poster to comply with the rules before I decide to do this.
- No Low-quality/Wildly Speculative Posts.
- Keep posts on topic.
- Don't make posts with link/s to paywalled articles as their main focus.
- No posts linking to reddit posts.
- Memes are fine as long they are quality or/and can lead to serious on topic discussions. If we end up having too much memes we will do meme specific singularity sublemmy.
- Titles must include information on how old the source is in this format dd.mm.yyyy (ex. 24.06.2023).
- Please be respectful to each other.
- No summaries made by LLMs. I would like to keep quality of comments as high as possible.
- (Rule implemented 30.06.2023) Don't make posts with link/s to tweets as their main focus. Melon decided that the content on the platform is going to be locked behind login requirement and I'm not going to force everyone to make a twitter account just so they can see some news.
- No ai generated images/videos unless their role is to represent new advancements in generative technology which are not older that 1 month.
- If the title of the post isn't an original title of the article or paper then the first thing in the body of the post should be an original title written in this format "Original title: {title here}".
- Please be respectful to each other.
Related sublemmies:
[email protected] (Our community focuses on programming-oriented, hype-free discussion of Artificial Intelligence (AI) topics. We aim to curate content that truly contributes to the understanding and practical application of AI, making it, as the name suggests, “actually useful” for developers and enthusiasts alike.)
Note:
My posts on this sub are currently VERY reliant on getting info from r/singularity and other subreddits on reddit. I'm planning to at some point make a list of sites that write/aggregate news that this subreddit is about so we could get news faster and not rely on reddit as much. If you know any good sites please dm me.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Watch pinned post on this community called "sparks of AGI", it should help you broaden your understanding on intelligence of LLM's and potentially ai's using different architecture.
How much time have you yourself spent trying to construct a system which applies LLMs to AGI problems? For me, it's a few full work days' worth of tinkering -- not much, but enough that I feel like I have some real-world perspective on what's involved. Please be careful about talking down to me about my need to broaden my understanding.
So I did watch part of Bubeck's talk quite a while ago when it first came out, because this is a deeply important topic to me. I don't fully agree with it; so let me take an excerpt from his paper to explain what I disagree with:
To me, this is a pretty clear statement of the core of what Bubeck is saying, both in the paper and the talk: He goes through a very accurate list of the unbelievably impressive things that GPT-4 can do. Then, he says more or less that because it can do those things, it must be intelligent (or at least have the first sparks of real intelligence). To me, I simply don't agree with that. Computerized systems could already do extremely impressive computational things; extending that into the domain of language is a huge leap forward, maybe towards AGI. But, if you're going to say that because it can mimic the language of reasoning, it must be able to reason, because there's in practice no difference between those things, then I don't agree with that.
Could we use LLMs as building blocks for a real AGI system? Yes, absolutely; like I say, I've spent a slight but nonzero amount of time actually experimenting with that myself. Are LLMs impressive? Fuck yes. Are they necessarily intelligent because they can do these impressive language-related tasks? To me, no. That seems like a non sequitur. To me, it's still clear interacting even with GPT-4 that it doesn't have real understanding of the underlying concepts, and these models are just getting better and better at moving symbols around. Again, I'd actually contrast that specifically against things like AlphaZero, where it does have a deep understanding of the underlying concepts, to the point that it can easily arrive at novel ideas on its own, beyond and superior to what it was programmed with.
Bruh, what? I just recommended you some videos becuse you wrote about how you were surprised about capabilities of AlphaZero. I'm not talking down to you, I was just trying to help you learn more and that's it...
Edit: And you overestimate human intelligence from what I can tell, we are not that special