Tech Firms Criticize New Australian Standards for Generative AI in Online Safety
The Guardian
SKIPPED
Details
- Date Published
- 29 Feb 2024
- Priority Score
- 4
- Australian
- Yes
- Created
- 10 Mar 2025, 10:27 pm
Description
The standards target generative AI’s misuse potential but Microsoft says its ability to flag problematic material could be hurt too
Summary
The article examines concerns from tech firms like Microsoft, Meta, and Stability AI about Australia's proposed online safety standards targeting generative AI. These standards aim to prevent the misuse of AI for generating harmful content, such as synthetic child abuse materials and sexual deepfakes. The tech companies argue that stringent rules could hinder AI's ability to effectively detect and prevent such content, as 'clean' training data might reduce the models' precision. The policy's relevance extends to global AI safety discourse, engaging with complex issues about balancing safety and innovation in AI systems. The outcome of this policy debate will have significant implications for international discussions on AI regulation.
Body
New Australian online safety standards cover a variety of technologies, including generative AI.Photograph: Dominic Lipinski/PAView image in fullscreenNew Australian online safety standards cover a variety of technologies, including generative AI.Photograph: Dominic Lipinski/PAThis article is more than1 year oldTech firms say new Australian standards will make it harder for AI to protect online safetyThis article is more than 1 year oldThe standards target generative AI’s misuse potential but Microsoft says its ability to flag problematic material could be hurt tooFollow our Australia news live blog for latest updatesGet ourmorning and afternoon news emails,free appordaily news podcastTech companies say new Australian safety standards will inadvertently make it harder for generative AI systems to detect and prevent online child abuse and pro-terrorism material.Under two mandatorystandards aimed at child safetyreleased in draft form by the regulator last year, the eSafety commissioner, Julie Inman Grant, proposed providers detect and remove child-abuse material and pro-terrorism material “where technically feasible”, as well as disrupt and deter new material of that nature.The standards cover a variety of technologies, including websites, cloud storage services, text messages and chat apps. They also cover high-impact generative AI services and open-source machine learning models.Australia releases new online safety standards to tackle terror and child sexual abuse contentRead moreIn a submission to the consultation on the standards published on Thursday, WeProtect Global Alliance, a non-profit consortium of more than 100 governments and 70 companies targeting child sexual exploitation and abuse online, highlighted the nature of the problem eSafety is trying to address. It said open-source AI is already being used to produce child abuse material and deepfakes, and the proposed standards capture the right platforms and services.“By focusing on the potential for misuse, the threshold reflects the reality that even machine learning and artificial intelligence models with limited direct exposure to sensitive data or datasets containing illicit data may still be misused to create illegal content, such as ‘synthetic’ child sexual abuse material and sexual deepfakes.”Sign up for Guardian Australia’s free morning and afternoon email newsletters for your daily news roundupBut tech companies includingMicrosoft, Meta and Stability AI saidtheir technologies were being developed with guardrails in place to prevent them being used in such a way.Microsoft warned that the standards, as drafted, could limit the effectiveness of AI safety models being used to detect and flag child abuse or pro-terror material.“To ensure that AI models and safety systems (such as classifiers) can be trained to detect and flag such content requires that the AI is exposed to such content and evaluation processes are put in place to measure and mitigate risks,” Microsoft said.“Entirely ‘clean’ training data may reduce the effectiveness of such tools and reduce the likelihood they operate with precision and nuance.“One of the most promising elements of AI tooling for content moderation is advanced AI’s ability to assess context – without training data that supports such nuanced assessment, we risk losing the benefits of such innovation.”Stability AI similarly warned that AI would play a large role in online moderation, and overly broad definitions could make it harder to determine what must be picked up in order to comply with the proposed standards.Facebook’s parent companyMetasaid while its Llama 2 model had safety tools and responsible use guides, it would be difficult to enforce safeguards when the tool is downloaded.skip past newsletter promotionafter newsletter promotionApple warns Australian proposal to force tech companies to scan cloud services could lead to mass surveillanceRead more“It is not possible for us to suspend provision of Llama 2 once it has been downloaded nor terminate an account, or to deter, disrupt, detect, report or remove content from models that have been downloaded,” the company said.Google recommended that AI not be included in the standards and instead be considered wholly as part of the current government review of the Online Safety Act and the Basic Online Safety Expectations.The tech companies alsoechoed comments made by Applelast week that the standards must explicitly state that proposals to scan cloud and message services “where technically feasible” will not compromise encryption, and technical feasibility will cover more than simply the cost to a company to develop such technology.In a statement, Inman Grant said the standards would not require industry to break or weaken encryption, monitor texts or indiscriminately scan large amounts of personal data, and the commissioner was now considering potential amendments to clarify this point.“Fundamentally, eSafety does not believe industry should be exempt from responsibility for tackling illegal content being hosted and shared freely on their platforms. eSafety notes some large end-to-end encrypted messaging services are already taking steps detect this harmful content,” she said.Final versions of the standards will be tabled in parliament for consideration later this year, Inman Grant said.Explore more on these topicsAustralia newsArtificial intelligence (AI)Online abuseMicrosoftAppleMetaGooglenewsShareReuse this content