WhatsApp’s AI shows gun-wielding children when prompted with ‘Palestine’::By contrast, prompts for ‘Israeli’ do not generate images of people wielding guns, even in response to a prompt for ‘Israel army’

    • theyoyomaster@lemmy.world
      link
      fedilink
      English
      arrow-up
      29
      arrow-down
      11
      ·
      1 year ago

      This isn’t anything they actively did though. The literal point of AI is that it learns on its own and comes up with its own response absent human interaction. Meta very likely specifically added code to try and prevent this, but it just fell short of overcoming the bias found in the overwhelming majority of content that led to the model associating Hamas with Palestine.

      • Valmond@lemmy.mindoki.com
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        2
        ·
        1 year ago

        It’s not about “adding code” or any other bullshit.

        AI today is trained on datasets (that’s about it), the choice of datasets can be complicated, but that’s where you moderate and select. There is nothing “AI learns of its own” sci-fi dream going on.

        Sigh.

        • Serdan@lemm.ee
          link
          fedilink
          English
          arrow-up
          6
          ·
          1 year ago

          It’s reasonable to refer to unsupervised learning as “learning on its own”.

        • Torvum@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          3
          ·
          1 year ago

          Really wish the term virtual intelligence was used (literally what it is)

          • ichbinjasokreativ@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            2
            ·
            1 year ago

            One of the many great things about the mass effect franchise is its separation of AI and VI, the latter being non-conscious and simple and the former being actually ‘awake’

        • theyoyomaster@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          It is about adding code. No dataset will be 100% free of undesirable results. No matter what marketing departments wish, AI isn’t anything close to human “intelligence,” it is just a function of learned correlations. When it comes to complex and sensitive topics, the difference between correlation and causation is huge and AI doesn’t distinguish. As a result, they absolutely hard code AI models to avoid certain correlations. Look at the “[character] doing 9/11” meme trend. At the fundamental level it is impossible to restrict undesirable outcomes by avoiding them in training models because there are an infinite combinations of innocent things that become sensitive when linked in nuanced ways. The only way to combat this is to manually delink certain concepts; they merely failed to predict it correctly for this specific instance.

    • pete_the_cat@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      14
      ·
      edit-2
      1 year ago

      I forget if it was on here or Reddit, but I remember seeing an article a week or so ago where the translation feature on Facebook ended up calling Palestinians terrorists “accidentally”. I cited the fact that Mark is Jewish, and probably so are a lot of the people that work there. The US is also largely pro-Israel, so it was probably less of an accidental bug and more of an intentional “fuck Palestine”. I got downvoted to hell and called a conspiracy theorist. I think this confirms I had the right idea.