3 Sources
3 Sources
[1]
Telling an AI model that it's an expert makes it worse
Researchers say persona-based prompting can improve works for safety but not for facts Many people start their work with AI by prompting the machine to imagine it is an expert at the task they want it to perform, a technique that boffins have found may be futile. Persona-based prompting - which involves using directives such as "You're an expert machine learning programmer" in a model prompt - dates back to 2023, when researchers began to explore how role-playing instructions influenced AI models' output. It's now common to find online prompting guides that include passages like, "You are an expert full-stack developer tasked with building a complete, production-ready full-stack web application from scratch." But academics who have researched this approach report it does not always produce superior results. In a pre-print paper titled "Expert Personas Improve LLM Alignment but Damage Accuracy: Bootstrapping Intent-Based Persona Routing with PRISM," researchers affiliated with the University of Southern California (USC) find that persona-based prompting is task-dependent - which they say explains the mixed results. For alignment-dependent tasks, like writing, role-playing, and safety, personas do improve model performance. For pretraining-dependent tasks like math and coding, using the technique produces worse results. The reason appears to be that telling a model it's an expert in a field does not actually impart any expertise - no facts are added to the training data. In fact, telling a model that it's an expert in a particular field hinders the model's ability to fetch facts from pretraining data. The researchers used the Measuring Massive Multitask Language Understanding (MMLU) benchmark, a means of evaluating LLM performance, to test persona-based prompting and found "when the LLM is asked to decide between multiple-choice answers, the expert persona underperforms the base model consistently across all four subject categories (overall accuracy: 68.0 percent vs. 71.6 percent base model). A possible explanation is that persona prefixes activate the model's instruction-following mode that would otherwise be devoted to factual recall." But persona-based guidance does help steer the model toward responses that satisfy the LLM-based judge assessing alignment. As an example, the authors note, "A dedicated 'Safety Monitor' persona boosts attack refusal rates across all three safety benchmarks, with the largest gain on JailbreakBench (+17.7 percentage points from 53.2 percent to 70.9 percent)." Zizhao Hu, a PhD student at USC and one of the study's co-authors, told The Register in an email that based on the study's findings, asking AI to adopt the persona of an expert programmer will not help code quality or utility. But pointing to the prompt guidance we linked to above, Hu said "many other aspects, such as UI-preference, project architecture, and tool-preference, are more towards the alignment direction, which do benefit from a detailed persona." "In the examples provided, we believe that the general expert persona is not necessary, such as 'You are an expert full-stack developer,' while the granular personalized project requirement might help the model to generate code that satisfies the user's requirements." Given that prompts about expertise do have an effect, the researchers - Hu and colleagues Mohammad Rostami and Jesse Thomason - proposed a technique they call PRISM (Persona Routing via Intent-based Self-Modeling) which attempts to harness the benefits of expert personas without the harm. "We use the gated LoRA [low-rank adaptation] mechanism, where the base model is entirely kept and used for generations that depend on pretrained knowledge," he explained, adding "This decision process is learned by the gate." The LoRA adapter is activated where persona-based behaviors improve output, and otherwise falls back on the unmodified model. The researchers designed PRISM to avoid the tradeoffs of other approaches - prompt-based routing, which applies expert personas at inference time, and supervised fine tuning, which bakes behavior into model weights. Asked whether there's a way to generalize about effective prompting methods, Hu said: "We cannot say for sure for general prompting, but from our discovery on expert persona prompt, a potential point is, 'When you care more about alignment (safety, rules, structure-following, etc), be specific about your requirement; if you care more about accuracy and facts, do not add anything, just send the query.'" ®
[2]
Stop telling AI it's an expert programmer, you're making it worse at its job -- new research shows the best results need specific prompts
* Telling AI it's an expert in something is causing it to go down a totally different route * By introducing a persona, AI might not be able to think for itself, reducing output quality * The best prompts explain the task to AI and give it all the context and tools it needs New research has claimed asking AI to 'act as an expert' doesn't actually improve result reliability despite being a widely used prompt enhancer. More specifically, it might help with alignment-style tasks such as writing, tone and structure guidance, but it likely hurts knowledge tasks like maths and coding. Per the data these so-called expert personas underperformance base models on benchmarks likely because they're triggering the AI to shift into instruction-following mode rather than fact recall. Stop over-engineering your AI prompts "We specifically discourage crafting (system) prompt for maximum performance by exploiting biases, as this may have unexpected side effects, reinforce societal biases and poison training data obtained with such prompts," the paper, written by researchers affiliated with the University of Southern California (USC) reads. Separate research along the same lines found that while persona prompting can help shape tone and style, it does nothing to add factual capability to a model. Instead, prompt length and accuracy matters. A comprehensively designed prompt will ultimately give AI as much context as it needs to act autonomously and generate higher-quality output. The paper introduces a new PRISM (Persona Routing via Intent-based Self-Modeling) solution, whereby AI generates answers with and without a persona and compares which answer is best. The AI then learned when to apply personas in the future, falling back on the base model's functionality when personas hurt output quality. Adding to the complexity of prompt engineering, the researchers also uncover differences in model types, noting that reasoning models benefit more from context length while instruction-tuned models can be most sensitive to personas. In short, it seems that model developers are doing all the work needed to ensure generative AI gives us the best output, and that we should only aim to give chatbots tasks and share relevant context without dictating how they should go about creating a response. Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button! And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.
[3]
Turns out, if you ask an AI to play an expert, it gets less reliable
Asking AI to pretend it's an expert can backfire, but researchers may have found a fix. you've probably seen the tip floating around: tell AI to act like an expert in a field, and you'll get better answers. It's popular advice, and it does work, sometimes. However, a new study suggests that using AI personas may not be as effective as we thought it would be. Researchers from the University of California tested 12 different personas across six language models. The personas ranged from math and coding experts to creative writers and safety monitors. The goal was to find out how well AI performs when it is instructed to act as an expert. Recommended Videos The results were mixed. Adopting a persona made the AI sound more professional and follow the rules better. But it also made the AI worse at recalling facts. According to the study, using an AI persona shifts it into an instruction-following mode rather than a knowledge-retrieval mode, and that tradeoff costs you accuracy. What's the solution? To fix this problem, the researchers developed PRISM, which stands for Persona Routing via Intent-based Self-Modeling. Instead of always using a persona or never using one, PRISM teaches AI to decide what's best for itself. When you ask a question, PRISM generates two answers: one from its default mode and one from its persona. It then compares the two and delivers the answer that performs better for a specific query. The expert answer isn't discarded even when the default answer wins. Instead, the reasoning style is saved in a lightweight component called a LoRA adapter, which the AI can draw from later when needed. The solution sounds simple, and yet, it's effective. How did PRISM perform? PRISM raised AI's overall score by one to two points on the MT-Bench, a test that measures how well an AI follows instructions and stays helpful. For writing and safety tasks, personas helped. For raw knowledge questions, skipping the persona proved to be the better option. The researchers plan to test PRISM with more personas and refine its ability to provide better answers. It's early days, but this could change how we prompt AI for good.
Share
Share
Copy Link
University of Southern California researchers found that popular expert persona prompts harm AI model accuracy on factual tasks. While instructing AI to act as an expert helps with writing and safety, it degrades performance on math and coding by shifting models into instruction-following over factual recall mode. A new technique called PRISM aims to solve this tradeoff.
A widely adopted prompting technique may be undermining AI model performance on factual tasks. Researchers from the University of Southern California discovered that persona-based prompting—telling an AI model it's an expert in a specific field—consistently degrades AI performance on knowledge-based tasks like math and coding
1
. The study tested expert persona prompts across 12 different personas and six language models, revealing a troubling pattern: accuracy dropped from 71.6 percent with base models to 68.0 percent when expert personas were applied on the MMLU benchmark1
.
Source: The Register
The research challenges conventional wisdom found in countless online prompting guides that recommend starting with phrases like "You are an expert full-stack developer" or "You're an expert machine learning programmer." While this approach has become standard practice since researchers first explored role-playing instructions in 2023, the USC team found that instructing AI to act as an expert doesn't actually impart any expertise—no facts are added to the training data
1
. Instead, expert personas activate the model's instruction-following mode at the expense of factual recall, creating a fundamental tradeoff between alignment and accuracy.The picture isn't entirely negative. The USC researchers, including PhD student Zizhao Hu along with Mohammad Rostami and Jesse Thomason, found that persona-based prompting delivers measurable benefits for alignment-dependent tasks such as writing, role-playing, and safety
1
. A dedicated "Safety Monitor" persona boosted attack refusal rates across three safety benchmarks, with JailbreakBench showing the largest gain—jumping 17.7 percentage points from 53.2 percent to 70.9 percent1
.This split reveals why effective prompting techniques require understanding task type. Hu explained that while asking an AI model to adopt the persona of an expert programmer won't improve code quality, detailed requirements about UI preferences, project architecture, and tool preferences—which lean toward alignment—do benefit from persona guidance
1
. The key distinction: personas help shape tone, structure, and rule-following behavior but add nothing to factual capability.Recognizing this fundamental tension, the researchers developed PRISM (Persona Routing via Intent-based Self-Modeling), a technique designed to harness the benefits of expert personas without sacrificing factual accuracy
1
. Rather than forcing users to choose between personas or base models, PRISM teaches the AI model to dynamically apply personas based on query type3
.The system works by generating two answers for each query—one from the default mode and one using a persona—then comparing which performs better
3
. PRISM uses a gated LoRA (low-rank adaptation) mechanism where the base model remains intact for generations requiring pretrained knowledge, while the LoRA adapter activates when persona-based behaviors improve output1
. This approach avoids the limitations of prompt-based routing applied at inference time and supervised fine-tuning that bakes behavior into model weights1
.Early results show promise. PRISM raised overall scores by one to two points on MT-Bench, a test measuring how well AI follows instructions and maintains helpfulness
3
. For writing and safety tasks, personas helped; for raw knowledge questions, skipping the persona proved superior3
.Related Stories
The research carries immediate implications for anyone working with large language models. Hu's guidance is direct: "When you care more about alignment (safety, rules, structure-following, etc), be specific about your requirement; if you care more about accuracy and facts, do not add anything, just send the query"
1
. This runs counter to the over-engineered prompts that have proliferated across the AI community.
Source: TechRadar
The study also highlights differences in model types, noting that reasoning models benefit more from context length while instruction-tuned models show greater sensitivity to personas. As the researchers plan to test PRISM with additional personas and refine its decision-making capabilities, this work could fundamentally reshape how we interact with AI systems
3
. The message is clear: model developers have optimized these systems extensively, and users should focus on providing clear tasks and relevant context rather than dictating how AI should think.Summarized by
Navi
[1]
[3]
1
Policy and Regulation

2
Policy and Regulation

3
Technology
