this post was submitted on 22 Sep 2023
15 points (89.5% liked)

LocalLLaMA

2179 readers
1 users here now

Community to discuss about LLaMA, the large language model created by Meta AI.

This is intended to be a replacement for r/LocalLLaMA on Reddit.

founded 1 year ago
MODERATORS
 

Reversal knowledge in this case being, if the LLM knows that A is B, does it also know that B is A, and apparently the answer is pretty resoundingly no! I'd be curious to see if some CoT affected the results at all

you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 4 points 11 months ago (2 children)

While I'm not totally caught up in LLM magic, as far as I am aware, all LLMs are doing is (heavily simplified) very fancy auto correct/text prediction. LLMs don't "know" anything. They aren't equating anything, the aren't "learning" the way people learn by association or relating a word to an object or idea. So from my understanding your assumption that.

LLM knows that A is B

already doesn't make sense in the context of current LLMs. Lots of people have made posts about A is B. So text prediction is saying when A, high probability that then B. Its not pulling from a base of knowledge, then constructing an answer to a question based on that pool of knowledge. It's finding relationships between character groups. Similar to Libre Office, it can check your grammar based on known patterns, but a LLM can use the greater context of its training data to find larger and larger patterns of character groups.

I'm not a computer scientist, but from my understanding LLMs are widely misunderstood. People talk so often about how they "hallucinate", or that they are "inaccurate", but I think those discussions are totally irrelevant in the long term. Have you ever considered that your phone's text completion is lying? What does that even mean, for auto-correct to lie? It doesn't know anything its just guessing the next letters/words given the words written so far. That's all LLMs are doing too, just significantly more sophisticated. So I have never once ever considered anything produced by a LLM as true or false, because it cannot possibly do that.

[โ€“] noneabove1182 5 points 11 months ago

To start, everything you're saying is entirely correct

However, the existence of emergent behaviours like chain of thought reasoning shows that there's more to this than pure text predictions, it picks up patterns that were never explicitly trained, so it's entirely feasible to ponder if they're able to recognize reverse patterns

Hallucinations are a vital part of understanding the models, they might not be long term problems but getting them to understand what they actually know to be true is extremely important in the growth and adoption of LLMs

I think there's a lot more to the training and generation of text than you're giving it credit, the simplest way to explain it is that it's text prediction, but there's way too much depth to the training and model to say that's all it is

At the end of the day it's just a fun thought inducing post :) but when Andrej karparthy says he doesn't have a great intuition on how LLM knowledge works (though in fairness he theorizes the same as you, directional learning) I think we can at least agree none of us know for sure what is correct!

load more comments (1 replies)