Technology

Senate Committee Criticizes Tech Giants for AI Transparency

Published November 27, 2024

Artificial intelligence chatbots like ChatGPT, Google’s Gemini, and Meta’s Llama have been identified as "high-risk" by a Senate inquiry, which calls for stricter regulations and accountability measures for these technologies. The report highlights that these AI models should undergo mandatory transparency and testing due to their potential risks.

Amid growing popularity in schools and workplaces, the Senate committee noted that original creators, including artists and writers, are at risk as these companies use copyrighted materials without permission to train their AI systems. Labor senator Tony Sheldon, who led the committee, expressed frustration over the evasive responses from tech executives, likening their tactics to a cheap magic trick.

The extensive inquiry spanned over nine months and culminated in 13 key recommendations, advocating for comprehensive legislation akin to EU regulations to ensure safety in high-risk AI applications across various sectors. Executives from major companies such as Amazon, Google, and Meta faced backlash for their lack of clear answers regarding user data collection practices linked to their AI models.

For instance, Amazon’s public policy chief, Matt Levey, was unable to clarify whether data from Alexa devices influenced their AI training, marking a trend of non-responsiveness across all three tech giants. The report indicated that tech companies frequently referenced their privacy policies as justification for using user data, even when the average consumer would need 46 hours a month just to read through them.

During the hearings, an AI chatbot unexpectedly interrupted a Google executive’s testimony, prompting questions about whether AI was involved in preparing her responses. Lucinda Longcroft, the executive in question, denied this suggestion.

Senator Sheldon articulated his dismay by saying, "These tech giants aren’t pioneers; they’re pirates. They are taking from our culture and creativity for their own profit while leaving Australians without support." He urged for legislation that prioritizes the rights of individuals over the interests of Silicon Valley corporations.

The proposed "AI act" would introduce crucial regulations for the technology industry and has received backing from both human rights advocates and media entities. Nevertheless, it faced pushback from some tech companies and major banks, who claimed it could be overly restrictive.

The committee highlighted that creative professionals are particularly vulnerable to the consequences of AI developing without proper oversight. The report encourages greater transparency in how AI developers use copyrighted material, insisting that they should license content and adequately compensate its creators.

Recently, over 36,000 creatives from various fields, including prominent names like actor Julianne Moore and author James Patterson, signed an open letter denouncing the unauthorized use of human art to train AI systems. Evidence presented at the hearings showed that voice actors' contracts might give Amazon rights to use their voices for generating AI audiobooks, risking countless job losses.

Claire Pullen, head of the Australian Writers’ Guild, expressed satisfaction with the inquiry’s findings, asserting that it solidifies the creators' stance against the theft of their work by tech companies. She remarked that it’s frustrating it took a Senate committee to compel tech firms to be accountable to local creators.

Industry Minister Ed Husic is currently evaluating the report’s proposals. Senate members like David Shoebridge commended the recommendations, particularly acknowledging the exploitation faced by creative workers. However, they were disappointed that the report did not fully align Australia with leading countries on AI strategy.

Calls for immediate legislative action have also been made, focusing on enforcing safeguards to ensure responsible AI use, especially in high-risk scenarios. This includes a push for transparency in political advertising to limit AI’s role in electoral processes.

Overall, the inquiry has sparked significant discussions on the ethical responsibilities of tech companies, with a focus on protecting the rights of individuals and creative workers in the age of AI.

AI, Senate, Technology