News Score: Score the News, Sort the News, Rewrite the Headlines

X blames users for Grok-generated CSAM; no fixes announced

No one knows how X plans to purge bad prompters While some users are focused on how X can hold users responsible for Grok’s outputs when X is the one training the model, others are questioning how exactly X plans to moderate illegal content that Grok seems capable of generating. X is so far more transparent about how it moderates CSAM posted to the platform. Last September, X Safety reported that it has “a zero tolerance policy towards CSAM content,” the majority of which is “automatically” dete...

Read more at arstechnica.com

© News Score  score the news, sort the news, rewrite the headlines