this post was submitted on 22 Feb 2024
72 points (96.2% liked)

AI

4006 readers
3 users here now

Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen.

founded 3 years ago
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 13 points 6 months ago (5 children)

Pretty sure these tools are often seeded with prompts that enforce diversity. Bing does the same or similar. I'm more amused by this, as the process isn't aware and can't actively enable or disable these settings.

To actively fit a historical prompt, it would need to not only consider images from the period, but also properly synthesize historical data to go with the prompt.

[–] [email protected] 2 points 6 months ago* (last edited 6 months ago) (1 children)

Yes, I saw some talk and a screenshot somewhere that showed that apparently in its current state, Gemini can (or could) be asked to output the prompt enhancements it used along with the generated images.

The screenshot showed someone asking for images of fruit, and the enhanced prompt included "racially diverse groups of people". Now if they're inserting something like that even for images containing no people at, it stands to reason that this is just a default enhancement they ALWAYS apply, no matter the prompt, which would explain the racially diverse Nazis (and all the other brouhahahas we've seen from them).

[–] [email protected] 1 points 6 months ago

That's really what I'm expecting. My guess is that the training data is skewed, and the prompt cannot adjust.

Either the machine will need to understand what is expected, or the company will need to address this and allow people to enable or disable diversity.

The first option may be impossible to attain at this stage. The second can lead to inappropriate images.

load more comments (3 replies)