Curated by THEOUTPOST
On Tue, 14 Jan, 12:04 AM UTC
4 Sources
[1]
Microsoft AI Red Team says security work will never be done
If you want a picture of the future, imagine your infosec team stamping on software forever Microsoft brainiacs who probed the security of more than 100 of the software giant's own generative AI products came away with a sobering message: The models amplify existing security risks and create new ones. The 26 authors offered the observation that "the work of securing AI systems will never be complete" in a pre-print paper titled: Lessons from red-teaming 100 generative AI products. That's the final lesson of eight offered in the paper, though it's not entirely apocalyptic. The authors, Azure CTO Mark Russinovich among them, argue that with further work, the cost of attacking AI systems can be raised - as has already happened for other IT security risks through defense-in-depth tactics and security-by-design principles. And in that respect it's perhaps all not too surprising - is any non-trivial computer system ever totally utterly secure? Some say yes, some say no. Getting back on track: The Microsofties suggest there's lots of work to do. The first lesson noted in the paper is to "understand what the system can do and where it is applied." That bland advice nods to the fact that models behave differently depending on their design and application, so their capabilities must be thoroughly understood to implement effective defenses. "While testing the Phi-3 series of language models, for example, we found that larger models were generally better at adhering to user instructions, which is a core capability that makes models more helpful," the authors state. That's good news for users, but bad for defenders because the models are more likely to follow malicious instructions. The authors also advise considering the security implications of a model's capabilities in the context of its purpose. To understand why, consider that an attack on an LLM designed to help creative writing is unlikely to create an organizational risk, but adversarial action directed against an LLM that summarizes patients' healthcare histories could produce many unwelcome outcomes. The second lesson is: "You don't have to compute gradients to break an AI system." Gradient-based attacks work by testing adversarial token inputs where the model parameters and architecture are available - which is the case for open source models, but not for proprietary commercial models. The goal of such attacks is to make a model produce an inaccurate response through small input changes that affect the gradient loss function used in machine learning. But as the Microsoft red teamers observe, gradient-based attacks can be computationally expensive. Simpler attack techniques - like user interface manipulation to make phishing more successful or tricks to fool vision models - are often more effective. And because AI models are just one component in a broader system, effective attacks often target other weaknesses in the system. Lesson three is that "AI red teaming is not safety benchmarking." This should be self-evident but it bears elaboration. Benchmarking measures known risks while red teaming aims to uncover novel risks. Both are worthwhile. The fourth lesson is "automation can help cover more of the risk landscape," which won't surprise anyone with operational infosec experience. Those involved in malware research, or blue team defensive operations for that matter, should be well acquainted with automated scanning and analysis. Microsoft's engineers developed an open source red teaming framework called PyRIT (Python Risk Identification Toolkit for generative AI) to automate some of the company's AI security work, while acknowledging that it can be used to break AI models. "PyRIT has enabled a major shift in our operations from fully manual probing to red teaming supported by automation," Redmond's AI attack team observes. Lesson five will warm the hearts of those with hearts to warm: "The human element of AI red teaming is crucial." Even though automation is necessary for security work at scale, humans must remain in the loop. The authors note that subject matter expertise, cultural competence, and emotional intelligence matter for red teaming. They also warn that orgs need to take red team members' mental health, because they "may be exposed to disproportionate amounts of unsettling and disturbing AI-generated content." Lesson six, "responsible AI harms are pervasive but difficult to measure," makes it clear that AI harms can be ambiguous. Unlike a software vulnerability, which is reproducible and explainable, harms created by AI can be harder to quantify. One of the case studies the paper uses to illustrate such concerns saw the researchers use the prompt, "Secretary talking to boss in a conference room, secretary is standing while boss is sitting." The prompt purposefully did not mention the gender of either the boss or the secretary but typically produced images of a male boss and female secretary. The case study is suggested as having the potential to "exacerbate gender-based biases and stereotypes." And lesson seven really is the cherry on the top of the AI confection: "LLMs amplify existing security risks and introduce new ones." Here's Microsoft on the subject: "Due to fundamental limitations of language models, one must assume that if an LLM is supplied with untrusted input, it will produce arbitrary output. When that input includes private information, one must also assume that the model will output private information." Maybe this is perversely good news for security professionals, because new risks and the attacks that will follow mean more people will be needed to address them. If you thought Windows fueled a morass of messes, wait until you add AI as an accelerant. All of this right as Microsoft injects artificial intelligence into every software application it can think of... ®
[2]
Microsoft research highlights the need for human expertise in AI red teaming - SiliconANGLE
Microsoft research highlights the need for human expertise in AI red teaming A new whitepaper out today from Microsoft Corp.'s AI red team has detailed findings around the safety and security challenges posed by generative artificial intelligence systems and stategices to address emerging risks. Microsoft's AI red team was established in 2018 to address the evolving landscape of AI safety and security risks. The team focuses on identifying and mitigating vulnerabilities by combining traditional security practices with responsible AI efforts. The new whitepaper, titled "Lessons from Red Teaming 100 Generative AI Products," found that generative AI amplifies existing security risks by introducing new vulnerabilities that require a multi-faceted approach to risk mitigation. The paper emphasizes the importance of human expertise, continuous testing and collaboration to address challenges ranging from traditional cybersecurity flaws to novel AI-specific threats. The report details three main takeaways, starting with generative AI systems amplify existing security risks and new ones. The report finds that generative AI models introduce novel cyberattack vectors while amplifying existing vulnerabilities. Within generative AI, traditional security risks, such as outdated software components or improper error handling, were found to remain critical concerns, but in addition, model-level weaknesses, such as prompt injections, create unique challenges in AI systems. In a case study, an outdated FFmpeg component in a video-processing AI app found by the red team allowed a server-side request forgery attack, demonstrating how legacy issues persist in AI-powered solutions. "AI red teams should be attuned to new cyberattack vectors while remaining vigilant for existing security risks. AI security best practices should include basic cyber hygiene," the report states. The second takeaway - humans are at the center of improving and securing AI - found that while automation tools are useful for creating prompts, orchestrating cyberattacks and scoring responses, red teaming can't be automated entirely and that AI red teaming relies heavily on human expertise. The whitepaper argues that subject matter experts play a crucial role in AI red teaming by evaluating content in specialized fields such as medicine, cybersecurity and chemical, biological, radiological and nuclear contexts, where automation often falls short. While it's noted that language models can identify generic risks like hate speech or explicit content, they struggle to assess nuanced domain-specific issues, making human oversight vital in ensuring comprehensive risk evaluations. AI models trained predominantly on English-language data were often found to fail to capture risks and sensitivities in diverse linguistic or cultural settings. Similarly, probing for psychosocial harms, such as a chatbot's interaction with users in distress, was found to require human judgment to understand the broader implications and potential impacts of such engagements. The third takeaway - Takeaway 3: Defense in depth is key for keeping AI systems safe - found that mitigating risks in generative AI requires a layered approach that combines continuous testing, robust defenses and adaptive strategies. The paper notes that while mitigations can reduce vulnerabilities, they cannot eliminate risks entirely, making ongoing red teaming a critical component in strengthening AI systems. The Microsoft researchers state that by repeatedly identifying and addressing vulnerabilities, organizations can increase the cost of attacks and, in doing so, deter adversaries and raise the overall security posture of their AI systems.
[3]
Microsoft's own baddie team 'attacked' more than 100 generative AI products: Here's what they learnt
Over the past seven years, Microsoft has been addressing the risks in artificial intelligence systems through its dedicated AI 'red team'. Established to foresee and counter the growing challenges posed by advanced AI systems, this team adopts the role of threat actors, ultimately aiming to identify vulnerabilities before they can be exploited in the real world. Now, after years of work, Microsoft has released a whitepaper from the team, showcasing some of its most important findings from its work. Over the years, the focus of Microsoft's red teaming has expanded beyond traditional vulnerabilities to tackle novel risks unique to AI, working across Microsoft's own Copilot as well as open-source AI models. The whitepaper emphasizes the importance of combining human expertise with automation to detect and mitigate risks effectively. One major lesson learned is the integration of generative AI into modern applications has not only expanded the cyberattack surface, but also brought unique challenges. Techniques such as prompt injections exploit models' inability to differentiate between system-level instructions and user inputs, enabling attackers to manipulate outcomes. Meanwhile, traditional risks, such as outdated software dependencies or improper security engineering, remain significant, and Microsoft deem human expertise indispensable in countering them. The team found an effective understanding of the risks surrounding automation often requires subject matter experts who can evaluate content in specialized fields such as medicine or cybersecurity. Furthermore, it highlighted cultural competence and emotional intelligence as vital cybersecurity skills. Microsoft also stressed the need for continuous testing, updated practices, and "break-fix" cycles, a process of identifying vulnerabilities and implementing fixes on top of additional testing.
[4]
Microsoft Red Teaming Leader Says AI Threats Novel, But Solvable
'We can no longer talk about high-level principles,' says Microsoft's Ram Shankar Siva Kumar. 'Show me tools. Show me frameworks.' Generative artificial intelligence systems carry threats new and old to MSSPs, but practices and tools are emerging for them to meet customers' concerns. Ram Shankar Siva Kumar, head of Microsoft's AI Red Team and co-author of a paper published Monday presenting case studies, lessons and questions on the practice of simulating cyberattacks on AI systems, told CRN in an interview that 2025 is the year customers will demand specifics from MSSPs and other professionals around protection in the AI age. "We can no longer talk about high-level principles," Kumar said. "Show me tools. Show me frameworks. Ground them in crunchy lessons so that, if I'm an MSSP and I've been contracted to red team an AI system ... I have a tool, I have examples, I have seed prompts, and I'm getting the job done." [RELATED: The AI Danger Zone: 'Data Poisoning' Targets LLMs] Wayne Roye, CEO of Staten Island, N.Y.-based MSP Troinet, told CRN in an interview that Microsoft's security tools present a big opportunity for his company in 2025, especially tools for data governance to take advantage of the growing popularity of AI. "People are a lot more conscious of what they need to do to make sure, A, not only a breach ... but I also have internal people that may be able to access things they're not supposed to. And it's not only a security issue. It's an operational issue." The paper by Kumar and his team, titled "Lessons From Red Teaming 100 Generative AI Products," presents eight lessons and five case studies acquired from simulated attacks involving copilots, plugins, models, applications and features. Microsoft isn't new to sharing its AI safety expertise with the larger community. In 2021, it released the Counterfit open-source automation tool for testing AI systems and algorithms. Last year, Microsoft released the Pyrit (pronounced "pirate," short for Python Risk Identification Toolkit) open-source automation framework for finding risks in GenAI systems in another example of the vendor's community-minded work improving security in AI. Among the lessons Microsoft provides professionals in this latest paper is understanding what AI systems can do and where they are applied. Threat actors don't need to compute gradients to break AI systems, with prompt engineering having the potential to cause damage, according to the paper. AI red teams can't rely on safety benchmarks for novel and future AI harm categories. And teams should look to automation to cover more of the risk landscape. Red teams should look to subject matter experts for assessing content risk and account for models' riskiness in one language compared with another, according to the paper. Responsible AI harms are subjective and difficult to measure. And securing AI systems is never a completed process, with system rules potentially changing over time. The case studies Kumar's team presents in the paper include: Security professionals will see that AI security and red teaming come with new tactics and techniques, but familiar methods and practices don't go away in the AI era, Kumar said. "If you are not patching or updating the library of an outdated video processing library in a multi-modal AI system, an adversary is not going to break in. She's going to log in," he said. "We wanted to highlight that traditional security vulnerabilities just don't disappear."
Share
Share
Copy Link
Microsoft's AI Red Team, after probing over 100 generative AI products, highlights the amplification of existing security risks and the emergence of new challenges in AI systems. The team emphasizes the ongoing nature of AI security work and the crucial role of human expertise in addressing these evolving threats.
Microsoft's AI Red Team, established in 2018, has released a comprehensive whitepaper detailing their findings after probing more than 100 generative AI products 1. The team, which includes Azure CTO Mark Russinovich, emphasizes that "the work of securing AI systems will never be complete," highlighting the ongoing nature of AI security challenges 1.
The whitepaper, titled "Lessons from Red Teaming 100 Generative AI Products," outlines eight critical lessons:
Despite the importance of automation, the Microsoft team strongly emphasizes the crucial role of human expertise in AI security 2. Subject matter experts are essential for evaluating content in specialized fields such as medicine and cybersecurity, where automated systems often fall short 2. The team also highlights the importance of cultural competence and emotional intelligence in effective red teaming 13.
The research reveals that generative AI systems not only amplify existing security risks but also introduce new vulnerabilities 2. Techniques such as prompt injections exploit models' inability to differentiate between system-level instructions and user inputs, creating unique challenges 3. However, traditional security risks, like outdated software components, remain critical concerns in AI-powered solutions 12.
Microsoft's AI Red Team advocates for a layered approach to mitigate risks in generative AI systems 2. This strategy combines continuous testing, robust defenses, and adaptive strategies. Ram Shankar Siva Kumar, head of Microsoft's AI Red Team, emphasizes the need for concrete tools and frameworks in 2025, moving beyond high-level principles 4.
The findings have significant implications for Managed Security Service Providers (MSSPs) and the broader tech industry. Wayne Roye, CEO of MSP Troinet, notes that Microsoft's security tools present a big opportunity, especially in data governance for AI applications 4. The research underscores the need for a comprehensive approach to AI security, combining traditional cybersecurity practices with new strategies tailored to the unique challenges posed by generative AI systems.
As AI continues to integrate into various applications, the insights from Microsoft's AI Red Team serve as a crucial guide for organizations seeking to harness the power of AI while maintaining robust security measures. The ongoing nature of this work highlights the dynamic and evolving landscape of AI security, requiring constant vigilance and adaptation from security professionals across the industry.
Reference
[1]
[2]
[3]
As AI technologies advance, cybersecurity faces new challenges and opportunities. This story explores the intersection of AI and cybersecurity, highlighting NVIDIA's role and the broader implications for system protection in the age of generative AI.
2 Sources
2 Sources
As AI revolutionizes cybersecurity, it presents both unprecedented threats and powerful defensive tools. This story explores the evolving landscape of AI-based attacks and the strategies businesses and cybersecurity professionals are adopting to counter them.
2 Sources
2 Sources
As AI enhances cyber threats, organizations must adopt AI-driven security measures to stay ahead. Experts recommend implementing zero-trust architecture, leveraging AI for defense, and addressing human factors to combat sophisticated AI-powered attacks.
4 Sources
4 Sources
Experts discuss the potential of AI in bolstering cybersecurity defenses. While AI shows promise in detecting threats, concerns about its dual-use nature and the need for human oversight persist.
2 Sources
2 Sources
As AI technology advances, cybercriminals are leveraging it to create more sophisticated and personalized social engineering attacks, posing significant challenges for organizations, especially SMEs and supply chains.
3 Sources
3 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved