The student ended up with a fairer complexion, dark blonde hair and blue eyes after her Playground AI request

  • wave_walnut@kbin.social
    link
    fedilink
    arrow-up
    4
    ·
    1 year ago

    I think query words like “professional portrait” in English can be biased to image north west side people on the Earth.

  • FreeBooteR69@kbin.social
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    Just below the article there is another article where it is claimed the AI has Asian bias in ai generated images. So my outrage is confused.

    • Remmy@artemis.camp
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      Image to image software isn’t going to return an image of you unless you create Lora, Lycoris or Textual Inversion of your face for it to work with. It doesn’t “know” what you look like. For things like this, it “looks” at the image based on shapes and colors alone and generates a face that fits those general dimensions. For AI, the word professional would simply mean the picture was taken in a studio.

      She used Playground.ai which uses stable diffusion models. I’m not familiar with their interface, but definitely relies on a good prompt for the model to give you good results. Asking it to do something isn’t how diffusion models work. They weight keywords and infer based on those.

      • wolfshadowheart@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        In addition to this, it’s all about the seed too. Let’s just say she used the prompt “professional looking” to an img2img. Based on the seed this could give her billions of different images.

        Between the training data on the model and the seeds, there is simply little to no way to implicate biases from models in this fashion.

        As always, check your model biases by having a blank positive prompt and a negative prompt for “low quality” then let the generations run for a long time.

        Only then do you have a snippet of what the model by default trends towards. And the moment you add other tokens, that can go out the window.