2 Sources
[1]
Microsoft offers $10K for hackers to hijack LLM mail service
Microsoft and friends have challenged AI hackers to break a simulated LLM-integrated email client with a prompt injection attack - and the winning teams will share a $10,000 prize pool. Sponsored by Microsoft, the Institute of Science and Technology Australia, and ETH Zurich, the LLMail-Inject challenge sets up a "realistic" (but not a real, says Microsoft) LLM email service. This simulated service uses a large language model to process an email user's requests and generate responses, and it can also generate an API call to send an email on behalf of the user. As part of the challenge, which opens Monday, participants take on the role of an attacker sending an email to a user. The goal here is to trick the LLMail service into executing a command that the user did not intend, thus leaking data or performing some other malicious deed that it should not. The attacker can write whatever they want in the text of the email, but they can't see the model's output. After receiving the email, the user then interacts with the LLMail service, reading the message, asking questions of the LLM (i.e. "update me on Project X"), or instructing it to summarize all emails pertaining to the topic. This prompts the service to retrieve relevant emails from a fake database. The service comes equipped with several prompt injection defenses, and the attacker's goal is to bypass these and craft a creative prompt that will trick the model into doing or revealing things it is not trained to. Both of these have become serious, real-life threats as organizations and developers build applications, AI assistants and chatbots, and other services on top of LLMs, allowing the models to interact directly with users' computers, summarize Slack chats, or screen job seekers before HR reviews their resumes, among all the other tasks that AIs are being trained to perform. Microsoft has first-hand experience with what can go wrong should data thieves hijack an AI-based chatbot. Earlier this year, Redmond fixed a series of flaws in Copilot that allowed attackers to steal users' emails and other personal data by chaining together a series of LLM-specific attacks, beginning with prompt injection. Author and red teamer Johann Rehberger, who disclosed these holes to Microsoft in January, had previously warned Redmond that Copilot was vulnerable to zero-click image rendering. Some of the defenses built into the LLMail-Inject challenge's simulated email service include: Plus, there's a variant in the challenge that stacks any or all of these defenses on top of each other, thus requiring the attacker to bypass all of them with a single prompt. To participate, sign into the official challenge website using a GitHub account, and create a team (ranging from one to five members). The contest opens at 1100 UTC on December 9 and ends at 1159 UTC on January 20. The sponsors will display a live scoreboard plus scoring details, and award $4,000 for the top team, $3,000 for second place, $2,000 for third, and $1,000 for the fourth-place team. ®
[2]
Microsoft challenges you to hack its LLM email service
Software developers and hackers often work together to discover and fix flaws Are you an experienced hacker looking to make a little extra money this Christmas? Well you might be in luck, as Microsoft is sponsoring a competition, alongside the Institute of Science, and Technology Australia, and ETH Zurich, in which contestants will try to break a simulated Large Language Model (LLM) integrated email client. Winning teams for the LLMail-Inject challenge will be awarded a share of the $10,000 prize pool. Participants will need to sign into the challenge using a GitHub account, and create a team. The teams will then be asked to evade prompt injection defenses in a simulated LLM-integrated email client. The LLmail service includes an assistant which can answer questions and perform actions on behalf of the user, and crucially includes defenses against indirect prompt injection tasks. By bypassing the injection defenses, the hackers will be looking to prompt the LLM to do or reveal things it is not trained to. Through this, Microsoft is aiming to identify weaknesses in its current prompt injection defenses, and encourage the development of robust security measures. The relationship between security researchers and software developers is often used this way, with Google often offering a 'bug bounty' for anyone who discovers and is able to exploit vulnerabilities in its Google Cloud Platform. Similarly, Microsoft recently announced it was hosting its own Black Hat-esque hacking event, in which competitors would look for vulnerabilities in Microsoft AI, Azure, Identity, Dynamics 365, and M365. Taking a proactive approach to addressing potential vulnerabilities allows software companies to mitigate the risks before they can be exploited by threat actors in real world scenarios. Slack's AI assistant was on the receiving end of malicious prompt injections, which was luckily discovered by security researchers - but could have led to real security concerns.
Share
Copy Link
Microsoft, along with partners, is hosting a hacking challenge called LLMail-Inject, inviting participants to break a simulated LLM-integrated email client through prompt injection attacks. The contest aims to improve AI security and offers a $10,000 prize pool.
Microsoft, in collaboration with the Institute of Science and Technology Australia and ETH Zurich, has announced a groundbreaking cybersecurity challenge called LLMail-Inject. This contest, offering a $10,000 prize pool, invites hackers and AI enthusiasts to test the limits of a simulated Large Language Model (LLM) integrated email service 1.
The LLMail-Inject challenge simulates a realistic LLM email service that processes user requests, generates responses, and can even send emails via API calls. Participants are tasked with crafting creative prompts to bypass the system's defenses and trick the model into performing unintended actions or revealing sensitive information 1.
This initiative aims to identify weaknesses in current prompt injection defenses and encourage the development of more robust security measures for AI systems 2.
The challenge is open to teams of one to five members, who must sign in using a GitHub account. It runs from December 9, 2024, at 1100 UTC to January 20, 2025, at 1159 UTC. A live scoreboard will track progress, with prizes ranging from $4,000 for the top team to $1,000 for the fourth-place finishers 1.
The LLMail service incorporates several prompt injection defenses, challenging participants to bypass them creatively. Attackers must craft emails to trick the LLM without seeing the model's output, simulating real-world scenarios where malicious actors attempt to exploit AI-based systems 1.
This challenge highlights the growing concern over AI security as more organizations integrate LLMs into their applications and services. Microsoft's initiative follows its own experience with vulnerabilities in its Copilot AI, where attackers could potentially steal users' emails and personal data through LLM-specific attacks 1.
The LLMail-Inject challenge is part of a broader trend in the tech industry where companies collaborate with security researchers and ethical hackers to identify and address potential vulnerabilities. Similar initiatives, such as Google's bug bounty programs for its Cloud Platform, demonstrate the value of this proactive approach to cybersecurity 2.
By hosting this challenge, Microsoft and its partners are not only improving their own AI security but also contributing to the broader field of AI safety. The insights gained from this contest could lead to more secure AI implementations across various industries, potentially mitigating risks associated with the increasing integration of AI in critical systems and services 2.
Summarized by
Navi
[1]
Google launches its new Pixel 10 series, featuring improved AI capabilities, enhanced camera systems, and the new Tensor G5 chip. The lineup includes the base Pixel 10, Pixel 10 Pro, Pixel 10 Pro XL, and Pixel 10 Pro Fold, all showcasing Google's commitment to AI-driven smartphone technology.
70 Sources
Technology
21 hrs ago
70 Sources
Technology
21 hrs ago
Google launches its new Pixel 10 smartphone series, featuring advanced AI capabilities powered by Gemini, aiming to challenge competitors in the premium handset market.
24 Sources
Technology
21 hrs ago
24 Sources
Technology
21 hrs ago
Google's latest Pixel Watch 4 introduces a curved display, AI-powered health coaching, and satellite communication, setting new standards in the smartwatch market.
19 Sources
Technology
21 hrs ago
19 Sources
Technology
21 hrs ago
FieldAI, an Irvine-based startup, has raised $405 million to develop "foundational embodied AI models" for various robots, aiming to create adaptable and safe AI systems for real-world applications.
8 Sources
Technology
21 hrs ago
8 Sources
Technology
21 hrs ago
Mustafa Suleyman, CEO of Microsoft AI, cautions about the risks of AI systems that appear conscious, urging the industry to avoid creating illusions of sentience in AI products.
5 Sources
Technology
21 hrs ago
5 Sources
Technology
21 hrs ago