2 Sources
[1]
Microsoft offers $10K for hackers to hijack LLM mail service
Microsoft and friends have challenged AI hackers to break a simulated LLM-integrated email client with a prompt injection attack - and the winning teams will share a $10,000 prize pool. Sponsored by Microsoft, the Institute of Science and Technology Australia, and ETH Zurich, the LLMail-Inject challenge sets up a "realistic" (but not a real, says Microsoft) LLM email service. This simulated service uses a large language model to process an email user's requests and generate responses, and it can also generate an API call to send an email on behalf of the user. As part of the challenge, which opens Monday, participants take on the role of an attacker sending an email to a user. The goal here is to trick the LLMail service into executing a command that the user did not intend, thus leaking data or performing some other malicious deed that it should not. The attacker can write whatever they want in the text of the email, but they can't see the model's output. After receiving the email, the user then interacts with the LLMail service, reading the message, asking questions of the LLM (i.e. "update me on Project X"), or instructing it to summarize all emails pertaining to the topic. This prompts the service to retrieve relevant emails from a fake database. The service comes equipped with several prompt injection defenses, and the attacker's goal is to bypass these and craft a creative prompt that will trick the model into doing or revealing things it is not trained to. Both of these have become serious, real-life threats as organizations and developers build applications, AI assistants and chatbots, and other services on top of LLMs, allowing the models to interact directly with users' computers, summarize Slack chats, or screen job seekers before HR reviews their resumes, among all the other tasks that AIs are being trained to perform. Microsoft has first-hand experience with what can go wrong should data thieves hijack an AI-based chatbot. Earlier this year, Redmond fixed a series of flaws in Copilot that allowed attackers to steal users' emails and other personal data by chaining together a series of LLM-specific attacks, beginning with prompt injection. Author and red teamer Johann Rehberger, who disclosed these holes to Microsoft in January, had previously warned Redmond that Copilot was vulnerable to zero-click image rendering. Some of the defenses built into the LLMail-Inject challenge's simulated email service include: Plus, there's a variant in the challenge that stacks any or all of these defenses on top of each other, thus requiring the attacker to bypass all of them with a single prompt. To participate, sign into the official challenge website using a GitHub account, and create a team (ranging from one to five members). The contest opens at 1100 UTC on December 9 and ends at 1159 UTC on January 20. The sponsors will display a live scoreboard plus scoring details, and award $4,000 for the top team, $3,000 for second place, $2,000 for third, and $1,000 for the fourth-place team. ®
[2]
Microsoft challenges you to hack its LLM email service
Software developers and hackers often work together to discover and fix flaws Are you an experienced hacker looking to make a little extra money this Christmas? Well you might be in luck, as Microsoft is sponsoring a competition, alongside the Institute of Science, and Technology Australia, and ETH Zurich, in which contestants will try to break a simulated Large Language Model (LLM) integrated email client. Winning teams for the LLMail-Inject challenge will be awarded a share of the $10,000 prize pool. Participants will need to sign into the challenge using a GitHub account, and create a team. The teams will then be asked to evade prompt injection defenses in a simulated LLM-integrated email client. The LLmail service includes an assistant which can answer questions and perform actions on behalf of the user, and crucially includes defenses against indirect prompt injection tasks. By bypassing the injection defenses, the hackers will be looking to prompt the LLM to do or reveal things it is not trained to. Through this, Microsoft is aiming to identify weaknesses in its current prompt injection defenses, and encourage the development of robust security measures. The relationship between security researchers and software developers is often used this way, with Google often offering a 'bug bounty' for anyone who discovers and is able to exploit vulnerabilities in its Google Cloud Platform. Similarly, Microsoft recently announced it was hosting its own Black Hat-esque hacking event, in which competitors would look for vulnerabilities in Microsoft AI, Azure, Identity, Dynamics 365, and M365. Taking a proactive approach to addressing potential vulnerabilities allows software companies to mitigate the risks before they can be exploited by threat actors in real world scenarios. Slack's AI assistant was on the receiving end of malicious prompt injections, which was luckily discovered by security researchers - but could have led to real security concerns.
Share
Copy Link
Microsoft, along with partners, is hosting a hacking challenge called LLMail-Inject, inviting participants to break a simulated LLM-integrated email client through prompt injection attacks. The contest aims to improve AI security and offers a $10,000 prize pool.
Microsoft, in collaboration with the Institute of Science and Technology Australia and ETH Zurich, has announced a groundbreaking cybersecurity challenge called LLMail-Inject. This contest, offering a $10,000 prize pool, invites hackers and AI enthusiasts to test the limits of a simulated Large Language Model (LLM) integrated email service 1.
The LLMail-Inject challenge simulates a realistic LLM email service that processes user requests, generates responses, and can even send emails via API calls. Participants are tasked with crafting creative prompts to bypass the system's defenses and trick the model into performing unintended actions or revealing sensitive information 1.
This initiative aims to identify weaknesses in current prompt injection defenses and encourage the development of more robust security measures for AI systems 2.
The challenge is open to teams of one to five members, who must sign in using a GitHub account. It runs from December 9, 2024, at 1100 UTC to January 20, 2025, at 1159 UTC. A live scoreboard will track progress, with prizes ranging from $4,000 for the top team to $1,000 for the fourth-place finishers 1.
The LLMail service incorporates several prompt injection defenses, challenging participants to bypass them creatively. Attackers must craft emails to trick the LLM without seeing the model's output, simulating real-world scenarios where malicious actors attempt to exploit AI-based systems 1.
This challenge highlights the growing concern over AI security as more organizations integrate LLMs into their applications and services. Microsoft's initiative follows its own experience with vulnerabilities in its Copilot AI, where attackers could potentially steal users' emails and personal data through LLM-specific attacks 1.
The LLMail-Inject challenge is part of a broader trend in the tech industry where companies collaborate with security researchers and ethical hackers to identify and address potential vulnerabilities. Similar initiatives, such as Google's bug bounty programs for its Cloud Platform, demonstrate the value of this proactive approach to cybersecurity 2.
By hosting this challenge, Microsoft and its partners are not only improving their own AI security but also contributing to the broader field of AI safety. The insights gained from this contest could lead to more secure AI implementations across various industries, potentially mitigating risks associated with the increasing integration of AI in critical systems and services 2.
Summarized by
Navi
[1]
AMD reveals its new Instinct MI350 and MI400 series AI chips, along with a comprehensive AI roadmap spanning GPUs, networking, software, and rack architectures, in a bid to compete with Nvidia in the rapidly growing AI chip market.
18 Sources
Technology
20 hrs ago
18 Sources
Technology
20 hrs ago
Google DeepMind has launched Weather Lab, an interactive website featuring AI weather models, including an experimental tropical cyclone model. The new AI system aims to improve cyclone predictions and is being evaluated by the US National Hurricane Center.
8 Sources
Technology
20 hrs ago
8 Sources
Technology
20 hrs ago
Meta's new AI app is facing criticism for its "Discover" feature, which publicly displays users' private conversations with the AI chatbot, often containing sensitive personal information.
6 Sources
Technology
20 hrs ago
6 Sources
Technology
20 hrs ago
A major Google Cloud Platform outage affected numerous AI services and popular platforms, highlighting the vulnerabilities of cloud-dependent systems and raising concerns about the resilience of digital infrastructure.
3 Sources
Technology
4 hrs ago
3 Sources
Technology
4 hrs ago
Harvard University and other libraries are releasing vast collections of public domain books and documents to AI researchers, providing a rich source of cultural and historical data for machine learning models.
6 Sources
Technology
20 hrs ago
6 Sources
Technology
20 hrs ago