this post was submitted on 30 Aug 2024
39 points (100.0% liked)

TechTakes

1416 readers
202 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 1 year ago
MODERATORS
 

Got the pointer to this from Allison Parrish who says it better than I could:

it's a very compelling paper, with a super clever methodology, and (i'm paraphrasing/extrapolating) shows that "alignment" strategies like RLHF only work to ensure that it never seems like a white person is saying something overtly racist, rather than addressing the actual prejudice baked into the model.

top 6 comments
sorted by: hot top controversial new old
[–] [email protected] 15 points 2 months ago (1 children)

I haven't been able to put this into words; but I've been surprised at how little thought is given to AI bias both by industry and the general public.

Before everyone went goo-goo for chatbots and shitty "art" there was a lot of talk in industry of algorithmic bias, articles about algorithmic bias, and companies who messed it up got shamed-- at least if they were particularly egregious about it.

But now it seems that "algorithm" has been rebranded to "AI". Because the black-box is a different sort of black box that people aren't familiar with, it's like they're willing to give it the benefit of the doubt all over again.

Even what controversy there has been (the Google image generation diversity prompting thing) has been more seen as humans influencing an AI rather than bias from the AI itself. And indeed Twitter thinks that if only they tell their chatbot to not be "woke" that will make it cool and edgy but also neutral.

But I just don't get it, this stuff is obviously and blatantly biased. It's fundamentally biased in that it reflects the internet without even a surface level understanding of what it's saying. I feel like I entered some sort of bizarro dimension where everyone forgot all the lessons of the past 15 years.

[–] [email protected] 5 points 2 months ago* (last edited 2 months ago)

They started calling it AI and people immediately started asking the kinds of questions that science fiction had primed us for, whether that's Skynet, Data, or Marvin. Apparently even among the supposedly intelligent folks the fact never quite landed that these stories (or at least the good ones) were ultimately trying to comment on people rather than creating fanciful situations to be dissected for their own sake.

[–] eestileib 14 points 2 months ago (1 children)

Yes, this is what they are designed to do when used in hiring and criminal justice contexts. They would not be getting used if they did anything else.

Nicely demonstrated by the researchers, but can anybody say they are surprised?

[–] [email protected] 14 points 2 months ago

I don’t think anyone is surprised, but brace yourself for the next round of OpenAI and peers claiming to fix this issue.

[–] L0rdMathias 3 points 2 months ago

Interesting results, interesting insights, neat takeaways and argumentation.

It's unfortunate they only tested models that were trained on SAE and they didn't have a control group of language models in other dialects. Seems like a huge oversight.

I wonder how this would play out with a model that has been trained on AAE, another non-SAE dialect, or even one trained in English but optimized for a non-english language.

[–] [email protected] -3 points 2 months ago

Humans do this too