• phx@lemmy.ca
    link
    fedilink
    arrow-up
    11
    ·
    edit-2
    1 year ago

    Any training data almost certainly has biases. For awhile, if you asked for pictures of people eating waffles or fried chicken they’d very likely be black.

    Most of the pictures I tried of kid-type characters were blue eyed.

    Then people review the output and say "hey this might still racist, so they tweak things to “diversity” the output. This is likely the result of that, where they’ve “fixed” one “problem” and created another.

    Behold, Homer in brownface. D’oh!