2 Sources
[1]
Topological approach detects adversarial attacks in multimodal AI systems
New vulnerabilities have emerged with the rapid advancement and adoption of multimodal foundational AI models, significantly expanding the potential for cybersecurity attacks. Researchers at Los Alamos National Laboratory have put forward a novel framework that identifies adversarial threats to foundation models -- artificial intelligence approaches that seamlessly integrate and process text and image data. This work empowers system developers and security experts to better understand model vulnerabilities and reinforce resilience against ever more sophisticated attacks. The study is published on the arXiv preprint server. "As multimodal models grow more prevalent, adversaries can exploit weaknesses through either text or visual channels, or even both simultaneously," said Manish Bhattarai, a computer scientist at Los Alamos. "AI systems face escalating threats from subtle, malicious manipulations that can mislead or corrupt their outputs, and attacks can result in misleading or toxic content that looks like a genuine output for the model. When taking on increasingly complex and difficult-to-detect attacks, our unified, topology-based framework uniquely identifies threats regardless of their origin." Multimodal AI systems excel at integrating diverse data types by embedding text and images into a shared high-dimensional space, aligning image concepts to their textual semantic notion (like the word "circle" with a circular shape). However, this alignment capability also introduces unique vulnerabilities. As these models are increasingly deployed in high-stakes applications, adversaries can exploit them through text or visual inputs -- or both -- using imperceptible perturbations that disrupt alignment and potentially produce misleading or harmful outcomes. Defense strategies for multimodal systems have remained relatively unexplored, even as these models are increasingly used in sensitive domains where they can be applied to complex national security topics and contribute to modeling and simulation. Building on the team's experience developing a purification strategy that neutralizes adversarial noise in attack scenarios on image-centered models, this new approach detects the signature and origin of adversarial attack on today's advanced artificial intelligence models. A novel topological approach The Los Alamos team's solution harnesses topological data analysis, a mathematical discipline focused on the "shape" of data, to uncover these adversarial signatures. When an attack disrupts the geometric alignment of text and image embeddings, it creates a measurable distortion. The researchers developed two pioneering techniques, dubbed "topological-contrastive losses," to quantify these topological differences with precision, effectively pinpointing the presence of adversarial inputs. "Our algorithm accurately uncovers the attack signatures, and when combined with statistical techniques, can detect malicious data tampering with remarkable precision," said Minh Vu, a Los Alamos postdoctoral fellow and lead author on the team's paper. "This research demonstrates the transformative potential of topology-based approaches in securing the next generation of AI systems and sets a strong foundation for future advancements in the field." The framework's effectiveness was rigorously validated using the Venado supercomputer at Los Alamos. Installed in 2024, the machine's chips combine a central processing unit with a graphics processing unit to address high-performance computing and giant-scale artificial intelligence applications. The team tested it against a broad spectrum of known adversarial attack methods across multiple benchmark datasets and models. The results were unequivocal: the topological approach consistently and significantly outperformed existing defenses, offering a more reliable and resilient shield against threats. The team presented the work, "Topological Signatures of Adversaries in Multimodal Alignments," at the International Conference on Machine Learning.
[2]
New Approach Detects Adversarial Attacks in Multimodal AI Systems | Newswise
Newswise -- New vulnerabilities have emerged with the rapid advancement and adoption of multimodal foundational AI models, significantly expanding the potential for cybersecurity attacks. Researchers at Los Alamos National Laboratory have put forward a novel framework that identifies adversarial threats to foundation models -- artificial intelligence approaches that seamlessly integrate and process text and image data. This work empowers system developers and security experts to better understand model vulnerabilities and reinforce resilience against ever more sophisticated attacks. "As multimodal models grow more prevalent, adversaries can exploit weaknesses through either text or visual channels, or even both simultaneously," said Manish Bhattarai, a computer scientist at Los Alamos. "AI systems face escalating threats from subtle, malicious manipulations that can mislead or corrupt their outputs, and attacks can result in misleading or toxic content that looks like a genuine output for the model. When taking on increasingly complex and difficult-to-detect attacks, our unified, topology-based framework uniquely identifies threats regardless of their origin." Multimodal AI systems excel at integrating diverse data types by embedding text and images into a shared high-dimensional space, aligning image concepts to its textual semantic notion (like the word "circle" with a circular shape). However, this alignment capability also introduces unique vulnerabilities. As these models are increasingly deployed in high-stakes applications, adversaries can exploit them through text or visual inputs -- or both -- using imperceptible perturbations that disrupt alignment and potentially produce misleading or harmful outcomes. Defense strategies for multimodal systems have remained relatively unexplored, even as these models are increasingly used in sensitive domains where they can be applied to complex national security topics and contribute to modeling and simulation. Building on the team's experience developing a purification strategy that neutralized adversarial noise in attack scenarios on image-centered models, this new approach detects the signature and origin of adversarial attack on today's advanced artificial intelligence models. The Los Alamos team's solution harnesses topological data analysis, a mathematical discipline focused on the "shape" of data, to uncover these adversarial signatures. When an attack disrupts the geometric alignment of text and image embeddings, it creates a measurable distortion. The researchers developed two pioneering techniques, dubbed "topological-contrastive losses," to quantify these topological differences with precision, effectively pinpointing the presence of adversarial inputs. "Our algorithm accurately uncovers the attack signatures, and when combined with statistical techniques, can detect malicious data tampering with remarkable precision," said Minh Vu, a Los Alamos postdoctoral fellow and lead author on the team's paper. "This research demonstrates the transformative potential of topology-based approaches in securing the next generation of AI systems and sets a strong foundation for future advancements in the field." The framework's effectiveness was rigorously validated using the Venado supercomputer at Los Alamos. Installed in 2024, the machine's chips combine a central processing unit with a graphics processing unit to address high-performance computing and giant-scale artificial intelligence applications. The team tested it against a broad spectrum of known adversarial attack methods across multiple benchmark datasets and models. The results were unequivocal: the topological approach consistently and significantly outperformed existing defenses, offering a more reliable and resilient shield against threats. The team presented the work, "Topological Signatures of Adversaries in Multimodal Alignments," at the International Conference on Machine Learning. Funding: This work was supported by the Laboratory Directed Research and Development program and the Institutional Computing Program at Los Alamos.
Share
Copy Link
Researchers at Los Alamos National Laboratory have created a novel framework using topological data analysis to identify and counter adversarial threats in multimodal AI systems, enhancing cybersecurity for advanced artificial intelligence models.
Researchers at Los Alamos National Laboratory have developed a groundbreaking framework to identify and counter adversarial threats in multimodal AI systems. This novel approach comes at a crucial time when the rapid advancement of foundational AI models has opened up new vulnerabilities to cybersecurity attacks 12.
Multimodal AI systems, which integrate and process both text and image data, have become increasingly prevalent. However, their ability to align diverse data types in a shared high-dimensional space also introduces unique vulnerabilities. Manish Bhattarai, a computer scientist at Los Alamos, explains, "As multimodal models grow more prevalent, adversaries can exploit weaknesses through either text or visual channels, or even both simultaneously" 1.
These vulnerabilities can lead to misleading or toxic content that appears genuine, posing significant risks in high-stakes applications and sensitive domains, including national security 2.
Source: Tech Xplore
The Los Alamos team's solution harnesses topological data analysis, a mathematical discipline focused on the "shape" of data, to uncover adversarial signatures. When an attack disrupts the geometric alignment of text and image embeddings, it creates a measurable distortion 12.
The researchers developed two pioneering techniques called "topological-contrastive losses" to quantify these topological differences with precision. Minh Vu, a Los Alamos postdoctoral fellow and lead author of the study, states, "Our algorithm accurately uncovers the attack signatures, and when combined with statistical techniques, can detect malicious data tampering with remarkable precision" 2.
The framework's effectiveness was rigorously validated using the Venado supercomputer at Los Alamos, installed in 2024. This advanced machine combines CPU and GPU capabilities to address high-performance computing and giant-scale AI applications 1.
The team tested the framework against a broad spectrum of known adversarial attack methods across multiple benchmark datasets and models. The results were unequivocal: the topological approach consistently and significantly outperformed existing defenses, offering a more reliable and resilient shield against threats 12.
This research demonstrates the transformative potential of topology-based approaches in securing the next generation of AI systems. It sets a strong foundation for future advancements in the field, empowering system developers and security experts to better understand model vulnerabilities and reinforce resilience against increasingly sophisticated attacks 2.
The team presented their work, titled "Topological Signatures of Adversaries in Multimodal Alignments," at the International Conference on Machine Learning, marking a significant step forward in AI security research 12.
NVIDIA announces significant upgrades to its GeForce NOW cloud gaming service, including RTX 5080-class performance, improved streaming quality, and an expanded game library, set to launch in September 2025.
9 Sources
Technology
5 hrs ago
9 Sources
Technology
5 hrs ago
As nations compete for dominance in space, the risk of satellite hijacking and space-based weapons escalates, transforming outer space into a potential battlefield with far-reaching consequences for global security and economy.
7 Sources
Technology
21 hrs ago
7 Sources
Technology
21 hrs ago
OpenAI updates GPT-5 to make it more approachable following user feedback, sparking debate about AI personality and user preferences.
6 Sources
Technology
13 hrs ago
6 Sources
Technology
13 hrs ago
A pro-Russian propaganda group, Storm-1679, is using AI-generated content and impersonating legitimate news outlets to spread disinformation, raising concerns about the growing threat of AI-powered fake news.
2 Sources
Technology
21 hrs ago
2 Sources
Technology
21 hrs ago
A study reveals patients' increasing reliance on AI for medical advice, often trusting it over doctors. This trend is reshaping doctor-patient dynamics and raising concerns about AI's limitations in healthcare.
3 Sources
Health
13 hrs ago
3 Sources
Health
13 hrs ago