this post was submitted on 31 Jul 2024
285 points (96.4% liked)

Technology

59646 readers
2656 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Meta "programmed it to simply not answer questions," but it did anyway.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 122 points 3 months ago (2 children)

Hallucinating is a fancy term for BEING WRONG.

Unreliable bullshit generator is still unreliable. Imagine that!

[–] [email protected] 52 points 3 months ago* (last edited 3 months ago) (2 children)

AI doesn't know what's wrong or correct. It hallucinates every answer. It's up to the supervisor to determine whether it's wrong or correct.

Mathematically verifying the correctness of these algorithms is a hard problem. It's intentional and the trade-off for the incredible efficiency.

Besides, it can only "know" what it has been trained on. It shouldn't be suprising that it cannot answer about the Trump shooting. Anyone who thinks otherwise simply doesn't know how to use these models.

[–] [email protected] 12 points 3 months ago (3 children)

It is impossible to mathematically determine if something is correct. Literally impossible.

At best the most popular answer, even if it is narrowed down to reliable sources, is what it can spit out. Even that isn't the same thing is consensus, because AI is not intelligent.

If the 'supervisor' has to determine if it is right and wrong, what is the point of AI as a source of knowledge?

[–] [email protected] 3 points 3 months ago* (last edited 3 months ago) (4 children)

It is impossible to mathematically determine if something is correct. Literally impossible.

No, you're wrong. You can indeed prove the correctness of a neural network. You can also prove the correctness of many things. It's the most integral part of mathematics and computer-science.

For example a very simple proof: with the conjecture that an even number is 2k of a number k, then you can prove that the addition of two even numbers is again an even number (and that prove is definite): 2a+2b=2(a+b), since a+b=k for some k.

Obviously, proving more complex mathematical problems like AI is more involved. But that's why we have scientists that work on that.

At best the most popular answer, even if it is narrowed down to reliable sources, is what it can spit out. Even that isn't the same thing is consensus, because AI is not intelligent.

That is correct. But it's not a limitation. It's by design. It's the tradeoff for the efficiency of the models. It's like lossy JPG compression. You accept some artifacts but in return you get much smaller images and much faster loading times.

But there are indeed "AI"s and neural networks that have been proven correct. This is mostly applied to safety critical applications like airplane collision avoidance systems or DAS. But a language model is not safety critical; so we take full advantage.

If the 'supervisor' has to determine if it is right and wrong, what is the point of AI as a source of knowledge?

You're completely misunderstanding the whole thing. The only reason why it's so incredibly good in many applications is because it's bad in others. It's intentionally designed that way. There are exact algorithms and there approximation algorithms. The latter tend to be much more efficient and usable in practice.

[–] [email protected] 10 points 3 months ago* (last edited 3 months ago) (3 children)

You can prove some things are correct, like math problems (assuming the axioms they are based on are also correct).

You can't prove that things like events having happened are correct. That's even a philosophical issue with human memory. We can't prove anything in the past actually happened. We can hope that our memory of events is accurate and reliable and work from there, but it can't actually be proven. In theory everything before could have just been implanted into our minds. This is incredibly unlikely (as well as not useful at best), but it can't be ruled out.

If we could prove events in the past are true we wouldn't have so many pseudo-historians making up crazy things about the pyramids, or whatever else. We can collect evidence and make inferences, but we can't prove it because it is no longer happening. There's a chance that we miss something or some information can't be recovered.

LLMs are algorithms that use large amounts of data to identify correlations. You can tune them to give more unique answers or more consistent answers (and other conditions) but they aren't intelligent. They are, at best, correlation finders. If you give it bad data (internet conversations) or incomplete data then it at best will (usually confidently) give back bad information. People who don't understand how they work assume they're actually intelligent and can do more than this. This is dangerous and should be dispelled quickly, or they believe any garbage it spits out, like the example from this post.

[–] [email protected] 3 points 3 months ago

You can’t prove that things like events having happened are correct.

You can't so solidly that this shouldn't even be discussed.

What should be is whether you can make a machine capable of reasoning.

There's symbolic logic, so you can maybe some day make a machine that makes correct syllogisms, detects incorrect syllogisms and such.

People who don’t understand how they work assume they’re actually intelligent and can do more than this. This is dangerous and should be dispelled quickly, or they believe any garbage it spits out, like the example from this post.

Sadly there's that archetype of "the narrow-minded not cool scientist against the cool brave inventor" which means that actively dispelling that may do harm. People who don't understand will match the situation with that archetype and it will reinforce their belief.

[–] [email protected] -3 points 3 months ago

Well but this kind of correctness applies to everything. By thag logic, you can't believe anything. I'm talking about an entirely different correctness. Like resistance against certain adversarial attacks. Of course, proving that the model is always correct, is as complicated as modelling the entire reality. That's infeasible. But it's also infeasible for every other software.

[–] otp -4 points 3 months ago (2 children)

This sounds like an overly pedantic view of "prove"

[–] conciselyverbose 8 points 3 months ago (2 children)

It's not pedantic. You can mathematically prove math.

You can't mathematically/algorithmically prove an event happened or did not happen.

[–] otp 0 points 3 months ago (1 children)

Adding "mathematically/algorithmically" in front of the word "prove" as if it were always implicitly there, and suggesting that it's the only way we should be using the word "prove" seems pretty darned pedantic to me.

[–] conciselyverbose 3 points 3 months ago* (last edited 3 months ago)

We're describing the behavior of software. It must be implicitly included. Software cannot do anything that isn't algorithmic.

[–] [email protected] -1 points 3 months ago (1 children)

You can prove mathematical logic and you can (not 1-to-1) tie that to symbolic logic, but since it's not 1-to-1, because of ambiguity of symbols, there will be much more complexity. I personally think that the future of various machine assistants lies there, and what LLM's now do is going to be used in auxiliary roles for that.

[–] conciselyverbose 6 points 3 months ago (2 children)

The problem is that mathematical proofs rely on the basic premise that the underlying assumptions are rock solid, and that the rules of the math are rock solid. It's rigorous logic rules, applied mathematically.

The real world is Bayesian. Even our hard sciences like physics are only "mostly" true, which is why stuff like relativity could throw a wrench in it. There's inherent uncertainty for everything, because it's all measurement based, with errors, and more importantly, the relationships all have uncertainty. There is no "we know a^2 and b^2, so c^2 must be this". It's "we think this news source is generally reliable and we think the sentiment of the article is that this crime was committed, so our logical assumption is that the crime was probably committed". But no link in the chain is 100%. "Rock solid" sources get corrupted, generally with a time lag before it's recognizable. Your interpretation of a simple article may be damn near 100%, but someone is still going to misread it, and a computer definitely can.

Uncertainty is central to reality, down to the fact that even quantum phenomena have to be talked about probabilistically because uncertainty is built in all the way down.

[–] [email protected] 1 points 3 months ago

This is why many philosophers came to criticize metaphysical logic in the 1800s, viewing it as dealing with absolutes when reality does not actually exist in absolutes, stating that we need some other logical system which could deal with the "fuzziness" of reality more accurately. That was the origin of the notion of dialectical logic from philosophers like Hegel and Engels, which caught on with some popularity in the east but then was mostly forgotten in the west outside of some fringe sections of academia. Even long prior to Bell's theorem, the physicist Dmitry Blokhintsev, who adhered to this dialectical materialist mode of thought, wrote a whole book on quantum mechanics where the first part he discusses the need to abandon the false illusion of the rigidity and concreteness of reality and shows how this is an illusion even in the classical sciences where everything has uncertainty, all predictions eventually break down, nothing is never possible to actually fully separate something from its environment. These kinds of views heavily influenced the contemporary physicist Carlo Rovelli as well.

[–] [email protected] -1 points 3 months ago (1 children)

You are describing LLMs, yes. But not what I'm describing.

I'm talking about machine finding syllogisms and checking their correctness. This can't be rock solid because of interpretation of the statement in natural language with its fuzzy semantics, but everything after that can be made rock solid. While in LLMs even it isn't.

That's what I'm talking about.

Humans make mistakes, but not such as LLM-generated texts contain.

I mean that one can build a reasoning machine which an LLM isn't.

[–] conciselyverbose 3 points 3 months ago* (last edited 3 months ago) (1 children)

I'm not describing LLMs. LLMs are completely irrelevant, and my examples had nothing to do with LLMs.

Formal logic requires propositions be Boolean in nature. They're true, or they're false.

That's not the real world. There are no booleans in the real world. In the real world, everything, down to the fundamental particles, is inherently probabilistic.

Our "certainty" is at most 99. a lot of 9s. It's never 100%. You can't say "the New York Times said X", and "the New York Times is perfectly reliable", so "X must be true". It's "given that the NYT said X and the NYT has a history of reporting facts with reasonably high accuracy, the probability X is true is...". If they get caught being shady, the estimates of previous information learned from them is retroactively changed. But there is no "proof", because there is no certainty anywhere in the chain. The world and human understanding of it has to be Bayesian. Again, down to the Uncertainty Principle about low level particles. Uncertainty is fundamental to reality. There is no certainty.

[–] [email protected] -3 points 3 months ago (1 children)

Why are you writing this to me?

Do you know what a syllogism is?

It doesn't require being certain of the information we're building it on. Only of existence of such categories.

Naturally people in Antiquity and Middle Ages who used symbolic logic were even less certain of the actual truths and lies in the world than we are.

It allows the truth to be subjective, but not the logical constructions. This is a very important trait both then and now.

The difference between the filter and the data going through it.

Of course you can't just feed all the data of all the PoVs and similar cases on something, integrate it into a model and expect your PoV to not clash with its output.

It's philosophically the same as why using dialectics is bad for science.

[–] conciselyverbose 2 points 3 months ago (1 children)

A syllogism is a tool for theoretical reasoning that doesn't actually apply in the real world, because it relies on Boolean possibility spaces. There is never an "all articles by X are correct", and there is no theoretical possibility that "all articles by X are correct" in the real world. The connections in the real world are literally always probabilistic. In every case. Every time.

You can't use formal logic for any real world use case because there are no valid starting assumptions. The only thing logic can ever prove is internal consistency, not fact.

[–] [email protected] -1 points 3 months ago

The only thing logic can ever prove is internal consistency, not fact.

Yes, and being able to build structures with internal consistency would be an advantage.

Nobody says you can prevent any "AI" oracle from saying things that aren't true.

But a tool which would generate a tree of possible logical conclusions from something given in language and then divided into statements on objects with statistical dependencies could be useful.

[–] [email protected] -5 points 3 months ago (1 children)

No. It's just pure math and logic. And LLMs are nothing more than billions of additions and multiplications. Literally. You can prove certain things on it just like you can prove theorems in mathematics. It's an ongoing ressearch field.

[–] [email protected] 6 points 3 months ago (1 children)

It's just pure math and logic. And LLMs are nothing more than billions of additions and multiplications.

Okay: using additions and multiplications prove the assassination attempt on Donald Trump happened

[–] [email protected] -1 points 3 months ago (1 children)

How would you even prove something like that outside of LLMs? What is your point? That you cannot prove anything except "I think therefore I am"?

Either you haven't read my comments or you're intentionally trying to be provocative.

[–] [email protected] 3 points 3 months ago (1 children)

My point is what OPs point was (which you veered away from in order to try to show off that You Are Very Smart): it is literally impossible for a computer system to prove a historical event has happened.

[–] [email protected] -1 points 3 months ago (1 children)

I'm having a hard time keeping track of all of the threads and replies evolving here. Forgive me. But I assume you mean the followong one?

It is impossible to mathematically determine if something is correct. Literally impossible.

This is simply a wrong statement. You can indeed prove certain properties on these models. That implies of course that you're able to formulate that property fully.

I don't know why the discussion went this far off track. The main point though is that everyone including OP is trying to discredit AI by bringing up things it was never supposed to be good at. By design, it's not good at knowledge retrieval. But everyone is hating it because it's hallucinating fake news. It's beyond me why people argue like that.

[–] [email protected] 2 points 3 months ago* (last edited 3 months ago)

You can indeed prove certain properties on these models.

Okay, how does the model prove the assassination attempt happened? Because that is what OP was talking about.
It was clear from the context that OP was saying "It is impossible to mathematically determine if something [historical] is correct." They omitted one word and instead of using context clues you went into a long unnecessary post on how we prove even numbers are divisible by 2. If you tried Iron Manning their post instead of trying to show off with an "Um Actually...." You wouldn't be getting lost in the replies as we'd be staying on the original topic.

The main point though is that everyone including OP is trying to discredit AI by bringing up things it was never supposed to be good at.

We're missing the context again. It's not people trying to discredit AI. People are trying to discredit companies insisting on using AI for things it is bad at.

It sounds like you actually agree with OP: AI should not be used for this purpose. Instead of saying "I agree, this is a bad use of AI, it should only be used for X, Y, and Z" you felt the need to White Knight for AI. The problem right now isn't AI being attacked, it's companies treating AI like a miracle that can do everything.

[–] [email protected] 6 points 3 months ago (1 children)

Your proof example is a proof from your discrete structures class. That’s very different than “proving” something like “the Trump assassination attempt was a conspiracy.”

Otherwise we could have gotten rid of courts a long time ago.

[–] [email protected] 0 points 3 months ago* (last edited 3 months ago)

Well obviously. But that was not at all what I said or claimed. I just said that you can prove certain properties of neural networks because others said that you can't. And others also misunderstood LLMs in general. They believe it's an information retrival service, which is wrong.

Besides, your argument, as you've written it, applies to everything. Literally. From Wikipedia, to News, even up to your eyesight. What can you actually prove? I don't understand the point you're making and how that is related to LLMs.

[–] [email protected] 0 points 3 months ago

Just like us. Sometimes it's better to have bullshit predictions than none.

[–] [email protected] -2 points 3 months ago (1 children)

The only reason why it’s so incredibly good in many applications is because it’s bad in others. It’s intentionally designed that way.

lolwut

[–] [email protected] -1 points 3 months ago (3 children)

It's designed in a ways that'll make it inherently incorrect. Even on a physical basis (due to numeric issues). It's not a problem of the algorithm because it has been designed that way. The problem is that you don't know how to correctly use it.

I can't explain it any differently without getting overly technical. You wouldn't understand it anyways, judging by your comment "lolwut". If you want to learn how LLMs work specifically, there are plenty of ressources on the internet.

[–] [email protected] 3 points 3 months ago* (last edited 3 months ago) (1 children)

It’s designed in a ways that’ll make it inherently incorrect. Even on a physical basis (due to numeric issues). It’s not a problem of the algorithm because it has been designed that way. The problem is that you don’t know how to correctly use it.

"It doesn't make a good source of knowledge."

"Yeah, but it is designed to be inherently wrong"

How does that make any sense when trying to use something for knowledge? Being inherently wrong is the opposite of helpful for knowledge.

AI is great at pattern recognition, but knowledge isn't pattern recognition. Needing to know when it gives false information requires the "supervisor" to already have that knowledge. That makes the AI less useful than a simple reference because at least the reference can come from a trusted source.

If people stopped trying to jam AI into situations where being correct is important it wouldn't be a problem. But excusing that because it is designed to be inherently wrong deserves another LOLWUT.

[–] [email protected] -3 points 3 months ago* (last edited 3 months ago)

How does that make any sense when trying to use something for knowledge? Being inherently wrong is the opposite of helpful for knowledge.

It was never designed to reproduce knowledge. It was designed to do reasoning and natural language processing and generation. You're using it wrong.

LULWUT

If you don't know what you're talking about and don't have any capacity to learn something new, it's sometimes best to stop talking. Especially when you're starting to get rude to knowlegable people that try to explain it to you.

[–] [email protected] 2 points 3 months ago

It's designed in a ways that'll make it inherently incorrect. Even on a physical basis (due to numeric issues). It's not a problem of the algorithm because it has been designed that way. The problem is that you don't know how to correctly use it.

So it is bad at things like giving or finding factual information. I agree, companies need to stop cramming it into everything (like search engines) for tasks that it is specifically bad at because it is not designed for it.

[–] [email protected] 1 points 3 months ago (1 children)

Can you recommend any for resource to start with? (If I can be picky, then something I can consume after a whole day of being a patent because there is no energy for much else.)

[–] [email protected] 0 points 3 months ago (1 children)

That is, unless you define correct in mathematical terms. Which no one has done yet.

[–] [email protected] -1 points 3 months ago

It also wouldn't be a source of knowledge. It would be a shitty calculator.

[–] [email protected] -3 points 3 months ago (1 children)

We should understand that 99.9% of what wee say and think and believe is what feels good to us and we then rationalize using very faulty reasoning, and that's only when really challenged! You know how I came up with these words? I hallucinated them. It's just a guided hallucination. People with certain mental illnesses are less guided by their senses. We aren't magic and I don't get why it is so hard for humans to accept how any individual is nearly useless for figuring anything out. We have to work as agents too, so why do we expect an early days LLM to be perfect? It's so odd to me. Computer is trying to understand our made up bullshit. A logic machine trying to comprehend bullshit. It is amazing it even appears to understand anything at all.

[–] [email protected] 2 points 3 months ago* (last edited 3 months ago)

You know how I came up with these words? I hallucinated them. It’s just a guided hallucination.

The the word hallucination means literally anything you want it to. Cool, cool. Very valiant of you.

[–] [email protected] 0 points 3 months ago (1 children)

Uhm. Have you ever talked to a human being.

[–] [email protected] 2 points 3 months ago* (last edited 3 months ago)

Human beings are not infallible either.

[–] [email protected] 1 points 3 months ago (2 children)

That's like saying car crash is just a fancy word for accident, or cat is just a fancy term for animal.

Hallucination is a technical term for this type of AI, and it's inherent to how it works at it's core.

And now I'll let you get back to your hating.

[–] [email protected] 9 points 3 months ago

Hallucination is also wildly misleading. The AI does not believe something that isn't real, it was incorrect in the words it guessed would be appropriate.

[–] [email protected] -1 points 3 months ago

The funny thing is we hallucinate all our answers too. I don't know where these words are coming from and I am not reasoning about them other than construction of a grammatically correct sentence. Why did I type this? I don't have a fucking clue. 😂

We map our meanings onto whatever words we see fit. If I had a dollar for every time I've heard a Republican call Obama a Marxist still blows my mind.

Thank you for saying something too. Better than I could do. I've been thinking about AI since I was a little kid. I've watched it go from at best some heuristic pathfinding in video games all the way to what we have now. Most people just weren't ever paying attention. It's been incredible to see that any of this was even possible.

I watched Two Minute Papers from back when he was mostly doing light transport simulation (raytracing). It's incredible where we are, but baffling people can't see the tech as separate from good old capitalism and the owner class. It just so happens it takes a fuckton of money to build stuff like this, especially at first. This is super early.