Got the pointer to this from Allison Parrish who says it better than I could:
it’s a very compelling paper, with a super clever methodology, and (i’m paraphrasing/extrapolating) shows that “alignment” strategies like RLHF only work to ensure that it never seems like a white person is saying something overtly racist, rather than addressing the actual prejudice baked into the model.
I haven’t been able to put this into words; but I’ve been surprised at how little thought is given to AI bias both by industry and the general public.
Before everyone went goo-goo for chatbots and shitty “art” there was a lot of talk in industry of algorithmic bias, articles about algorithmic bias, and companies who messed it up got shamed-- at least if they were particularly egregious about it.
But now it seems that “algorithm” has been rebranded to “AI”. Because the black-box is a different sort of black box that people aren’t familiar with, it’s like they’re willing to give it the benefit of the doubt all over again.
Even what controversy there has been (the Google image generation diversity prompting thing) has been more seen as humans influencing an AI rather than bias from the AI itself. And indeed Twitter thinks that if only they tell their chatbot to not be “woke” that will make it cool and edgy but also neutral.
But I just don’t get it, this stuff is obviously and blatantly biased. It’s fundamentally biased in that it reflects the internet without even a surface level understanding of what it’s saying. I feel like I entered some sort of bizarro dimension where everyone forgot all the lessons of the past 15 years.
They started calling it AI and people immediately started asking the kinds of questions that science fiction had primed us for, whether that’s Skynet, Data, or Marvin. Apparently even among the supposedly intelligent folks the fact that these stories (or at least the good ones) were ultimately trying to comment on people rather than creating fanciful situations to be dissected for their own sake.