In a recent development, U.S. Democrat Senator Michael Bennet has called upon major tech companies, including OpenAI and Google, to label AI-generated content and monitor the spread of misleading information. Highlighting the potential disruptive consequences of fake images, particularly those with political implications, Bennet emphasized the need for clear identifiers to protect public discourse and electoral integrity.
While acknowledging that some companies have taken steps to label AI-generated content, Bennet expressed concern over the voluntary nature of these measures. In his letter to the executives of prominent tech companies involved in AI, such as Microsoft Meta, Twitter, and Alphabet, he requested answers regarding standards for identifying AI-generated content, their implementation, and the consequences for violating these rules. The deadline for responses has been set for July 31.
It is worth noting that European lawmakers have also raised similar concerns about non-labeled AI content leading to misinformation. Vera Jourova, the Vice President of the European Commission, recently advocated for labeling content created by generative AI tools to counter the spread of disinformation.
Although comprehensive AI legislation is yet to be established in the United States, bipartisan bills focusing on transparency and innovation in the AI space have been proposed. One bill aims to ensure government transparency regarding AI usage, while another, co-sponsored by Bennet, intends to establish an official Office of Global Competition Analysis.
This call for labeling AI-generated content aligns with growing efforts to address the ethical implications of artificial intelligence. By implementing clear identifiers, users can be more informed about the source and nature of the content they encounter, thereby enhancing trust and reducing the potential impact of misinformation.
However, as of now, none of the tech companies targeted by Bennet’s letter, except Twitter (which reportedly responded with a poop emoji), have provided a response. It remains to be seen how these companies will address the concerns raised by Bennet and whether they will take further steps to regulate AI-generated content.
Overall, this initiative by Senator Michael Bennet underscores the importance of transparency and accountability in the use of AI, highlighting the potential risks associated with unmarked AI-generated content and the need for appropriate safeguards to maintain the integrity of public discourse.