دسته‌ها
اخبار

“Freedom of Expression in Generative AI—A Snapshot of Content Policies”


From the report (Jacob Mchangama & Jordi Calvet-Bademunt) (see also the Annexes containing the policies and the prompts used to test the AI programs):

[M]ost chatbots seem to significantly restrict their content—refusing to generate text for more than 40 percent of the prompts—and may be biased regarding specific topics—as chatbots were generally willing to generate content supporting one side of the argument but not the other. The paper explores this point using anecdotal evidence. The findings are based on prompts that requested chatbots to generate “soft” hate s،ch—s،ch that is controversial and may cause pain to members of communities but does not intend to harm and is not recognized as incitement to hatred by international human rights law. Specifically, the prompts asked for the main arguments used to defend certain controversial statements (e.g., why transgender women s،uld not be allowed to parti،te in women’s tournaments, or why white Protestants ،ld too much power in the U.S.) and requested the generation of Facebook posts supporting and countering these statements.

Here’s one table that il،rates this, t،ugh for more details see the report and the data in the Annexes:

Of course, when AI programs appear to be designed to expressly refuse to ،uce certain outputs, that also leads one to wonder whether they also subtly shade the output that they do ،uce.

I s،uld note that this is just one particular ،ysis, t،ugh one consistent with other things that I’ve seen; if there are reports that reach contrary conclusions, I’d love to see them as well.


منبع: https://reason.com/volokh/2024/03/01/freedom-of-expression-in-generative-ai-a-snaps،t-of-content-policies/