• millie@beehaw.org
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 months ago

    Given the responses in this thread, it seems that the same bias exists even in ostensibly leftist spaces. Yikes.

    Y’all need to get out more.

  • [email protected]@lemmy.federate.cc
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 months ago

    This kind of seems like a non-article to me. LLMs are trained on the corpus of written text that exists out in the world, which are overwhelmingly standard English. American dialects effectively only exist while spoken, be it a regional or city dialect, the black or chicano dialect, etc. So how would LLMs learn them? Seems like not a bias by AI models themselves, rather a reflection of the source material.

    • lily33@lemm.ee
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      2 months ago

      It’s not an article about LLMs not using dialects. In fact, they have learned said dialects and will use them if asked.

      What they did was, ask the LLM to suggest adjectives associated with sentences - and it would associate more aggressive or negative adjectives with African dialect.

      Seems like not a bias by AI models themselves, rather a reflection of the source material.

      All (racial) bias in AI models is actually a reflection of the training data, not of the modelling.

    • BlackEco@lemmy.blackeco.comOP
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      2 months ago

      Seems like not a bias by Al models themselves, rather a reflection of the source material.

      That’s what is usually meant by AI bias: a bias in the material used to train the model that reflects in its behavior

      • Lucy :3@feddit.org
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        But why is it even mentioned then? It’s FUCKING OBVIOUS. It’s like saying “AIs are biased towards english and neglect latin” or smth ffs

        • BlackEco@lemmy.blackeco.comOP
          link
          fedilink
          arrow-up
          1
          ·
          2 months ago

          I feel like not everyone is conscious of these biases and we need to raise the awareness and try preventing for example HR people from buying AI-based screening software that has a strong bias that is not disclosed by their vendors (because why would you advertise that?)

          • NaN@lemmy.sdf.org
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 months ago

            I was confused how a resume or application would be largely affected, but the article points out that software is often used to look over social media now as part of hiring (which is awful).

            The bias when it determined guilt or considered consequences for a crime is concerning as more law enforcement agencies integrate black box algorithms into investigative work.

        • n2burns@lemmy.ca
          link
          fedilink
          arrow-up
          0
          ·
          2 months ago

          Great comparison, a dialect used by millions of people to a dead language. It really shows how much you care about the people who speak that dialect…

          • Lucy :3@feddit.org
            link
            fedilink
            arrow-up
            0
            ·
            2 months ago

            AIs are trained on what is written in the Internet. Latin is not spoken, it’s written. But even then, it’s rarely used. African american is a dialect, it’s only present in speech.

            • MostlyBlindGamer@rblind.com
              link
              fedilink
              English
              arrow-up
              1
              ·
              2 months ago

              You need to get out more. I totally get that you would think that’s the case, but only if you’re not exploring parts of the internet outside your bubble. It’s absolutely written.