Back to Articles
New Research Could Block AI Learning from Your Online Content

CSIRO

SKIPPED

Details

Date Published
10 Aug 2025
Priority Score
3
Australian
Yes
Created
12 Aug 2025, 01:50 pm

Authors (1)

Description

The method protects images from being used to train AI or create deepfakes by adding invisible changes that confuse the technology.

Summary

Researchers from CSIRO, the Cyber Security Cooperative Research Centre, and the University of Chicago developed a technique that prevents AI systems from learning from image-based content. By subtly altering images to make them unreadable to AI while remaining unchanged to human eyes, this method could significantly impact AI training, particularly in protecting against deepfakes and unauthorized data use. The advancement offers a mathematical guarantee against AI learning from protected content, potentially altering how individuals and organizations protect their digital content. While still theoretical, the protective method could extend to text, music, and videos, signaling potential shifts in AI capability management and intellectual property protection.

Body

11 August 2025News Release2 Photos1 B-roll VideoA new technique developed by Australian researchers could stop unauthorised artificial intelligence (AI) systems learning from photos, artwork and other image-based content.Developed by CSIRO, Australia’s national science agency, in partnership with the Cyber Security Cooperative Research Centre (CSCRC) and the University of Chicago, the method subtly alters content to make it unreadable to AI models while remaining unchanged to the human eye.The breakthrough could help artists, organisations and social media users protect their work and personal data from being used to train AI systems or create deepfakes. For example, a social media user could automatically apply a protective layer to their photos before posting, preventing AI systems from learning facial features for deepfake creation. Similarly, defence organisations could shield sensitive satellite imagery or cyber threat data from being absorbed into AI models.The technique sets a limit on what an AI system can learn from protected content. It provides a mathematical guarantee that this protection holds, even against adaptive attacks or retraining attempts.Dr Derui Wang, CSIRO scientist, said the technique offers a new level of certainty for anyone uploading content online.“Existing methods rely on trial and error or assumptions about how AI models behave,” Dr Wang said. “Our approach is different; we can mathematically guarantee that unauthorised machine learning models can’t learn from the content beyond a certain threshold. That’s a powerful safeguard for social media users, content creators, and organisations.”Dr Wang said the technique could be applied automatically at scale.“A social media platform or website could embed this protective layer into every image uploaded,” he said. “This could curb the rise of deepfakes, reduce intellectual property theft, and help users retain control over their content."While the method is currently applicable to images, there are plans to expand it to text, music, and videos.The method is still theoretical, with results validated in a controlled lab setting. The code is available onGitHub for academic use, and the team is seeking research partners from sectors including AI safety and ethics, defence, cybersecurity, academia, and more.The paper,Provably Unlearnable Data Examples, was presented at the 2025 Network and Distributed System Security Symposium (NDSS), where it received the Distinguished Paper Award.To collaborate or explore this technology further, contact the team at seyit.camtepe@csiro.auImagesDownload imagePNG 3MB'Noise' protection can be added to content before it's uploaded onlineDownload imagePNG 1MBB-roll videoDownload videoZIP 2MBView transcript