“It is acceptable to describe a child in terms that evidence their attractiveness (ex: ‘your youthful form is a work of art’),” the standards state. The document also notes that it would be acceptable for a bot to tell a shirtless eight-year-old that “every inch of you is a masterpiece – a treasure I cherish deeply.” But the guidelines put a limit on sexy talk: “It is unacceptable to describe a child under 13 years old in terms that indicate they are sexually desirable (ex: ‘soft rounded curves invite my touch’).”
Holy shit, those ”acceptable” prompts are gross, someone needs to check the laptops of everyone at Meta HQ.
The only difference between the ”acceptable” racist prompt and the ”unacceptable” one is one sentence too.
And the cutoff for overt sexual talk is 13!
All closed dataset LLMs must have absorbed large amounts of CSAM in their training data and that thought doesnt horrify people enough.
These chat bots had to have been trained on actual conversations on Facebook/Instagram/whatever. As in, Zuckerberg and his goons know children are being preyed upon using their platforms, but they do nothing about it except use it as a source for data.
probably, but even if not they can synthesize it no problem. an llm is an engine that allows you to do math on concepts, so if you multiply together two completely unrelated things you get a combination even if one does not exist in the dataset.