this post was submitted on 29 Jan 2024
1019 points (99.1% liked)

Technology

59105 readers
3851 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 7 points 9 months ago (3 children)
[–] [email protected] 6 points 9 months ago

Here is an alternative Piped link(s):

AI does not exist, but it will ruin everything anyway.

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.

[–] [email protected] 2 points 9 months ago (2 children)

Loads of good points in that video, thanks for posting. The only argument I don't really agree with is about bias. She's implying here that a human decision maker would be less biased than the AI model. I'm not convinced by that because the training data is just a statistical record of human bias. So as long as the training data is well selected for your problem, it should be a good predictior for the likelihood of bias in your human decision maker.

[–] [email protected] 3 points 9 months ago (1 children)

I think with a human operator, we can be proactive. A person can be informed of bias, learn to recognize it, and even attempt to compensate for their own.

An AI model is working off of aggregate past data that we already know is biased. There is currently no proactive anti bias training that can be done to a AI model without massively altering the dataset, which, at some level of alteration, loses its value as true to life data.

Secondly, AI is a black box. we can’t see inner the workings of the model and determine what types of associations it is making to come to its result. So we don’t even know what part of the dataset would need to be altered to address the bias.

Lastly, the default assumption by end users will be, unless there are glaring defects, that any individual result is correct and unbiased, because “AI was made by smart people and data, and data doesn’t lie.” And because interrogating and validating the result defeats the whole purpose of using AI to cut out those steps of the process.

[–] [email protected] 1 points 9 months ago

I think with a human operator, we can be proactive. A person can be informed of bias, learn to recognize it, and even attempt to compensate for their own.

I think you're being very optimistic here. I hope very much that you'd be right about the humans. I have a feeling that a lot of these type of decisions are also resulting from implicit biases in humans that these humans themselves might not even recognize or acknowledge. Few sexists or racists will admit to being racists or sexists.

I agree about your point about the "computer says no" issue. That's also addressed in the video and fits well into her wider point that large parts of the population not understanding how so-called AI works is a huge problem.

[–] [email protected] 2 points 9 months ago (1 children)

the training data is just a statistical record of human bias.

It's not. It's a record of online conversations, which tend to be more polarized and extreme than real people.

[–] [email protected] 1 points 9 months ago

That's why I said

So as long as the training data is well selected for your problem...

It's clear that in the training data for LLMs, 4chan, reddit, etc. are over-represented, so that explains why chatgpt might be more awful than an average person. Having an LLM decide on, e.g., college admission would be like having a Twitter poll to decide on who should be its next CEO. Like that's obviously stupid, nobody would ever do that, right?

The problem is that for the college admission example, the models were trained on previous admissions, taken by college employees , and these models are still biased.

[–] [email protected] 2 points 9 months ago

acollierastro is a treasure.