Telling AI Models They're Experts Actually Makes Them Worse at Knowledge Tasks

Reviewed byNidhi Govil

3 Sources

Share

University of Southern California researchers found that popular expert persona prompts harm AI model accuracy on factual tasks. While instructing AI to act as an expert helps with writing and safety, it degrades performance on math and coding by shifting models into instruction-following over factual recall mode. A new technique called PRISM aims to solve this tradeoff.

Expert Persona Prompts Backfire on Knowledge-Based Tasks

A widely adopted prompting technique may be undermining AI model performance on factual tasks. Researchers from the University of Southern California discovered that persona-based prompting—telling an AI model it's an expert in a specific field—consistently degrades AI performance on knowledge-based tasks like math and coding

1

. The study tested expert persona prompts across 12 different personas and six language models, revealing a troubling pattern: accuracy dropped from 71.6 percent with base models to 68.0 percent when expert personas were applied on the MMLU benchmark

1

.

Source: The Register

Source: The Register

The research challenges conventional wisdom found in countless online prompting guides that recommend starting with phrases like "You are an expert full-stack developer" or "You're an expert machine learning programmer." While this approach has become standard practice since researchers first explored role-playing instructions in 2023, the USC team found that instructing AI to act as an expert doesn't actually impart any expertise—no facts are added to the training data

1

. Instead, expert personas activate the model's instruction-following mode at the expense of factual recall, creating a fundamental tradeoff between alignment and accuracy.

When Expert Personas Actually Help

The picture isn't entirely negative. The USC researchers, including PhD student Zizhao Hu along with Mohammad Rostami and Jesse Thomason, found that persona-based prompting delivers measurable benefits for alignment-dependent tasks such as writing, role-playing, and safety

1

. A dedicated "Safety Monitor" persona boosted attack refusal rates across three safety benchmarks, with JailbreakBench showing the largest gain—jumping 17.7 percentage points from 53.2 percent to 70.9 percent

1

.

This split reveals why effective prompting techniques require understanding task type. Hu explained that while asking an AI model to adopt the persona of an expert programmer won't improve code quality, detailed requirements about UI preferences, project architecture, and tool preferences—which lean toward alignment—do benefit from persona guidance

1

. The key distinction: personas help shape tone, structure, and rule-following behavior but add nothing to factual capability.

PRISM: A Solution to the Persona Tradeoff

Recognizing this fundamental tension, the researchers developed PRISM (Persona Routing via Intent-based Self-Modeling), a technique designed to harness the benefits of expert personas without sacrificing factual accuracy

1

. Rather than forcing users to choose between personas or base models, PRISM teaches the AI model to dynamically apply personas based on query type

3

.

The system works by generating two answers for each query—one from the default mode and one using a persona—then comparing which performs better

3

. PRISM uses a gated LoRA (low-rank adaptation) mechanism where the base model remains intact for generations requiring pretrained knowledge, while the LoRA adapter activates when persona-based behaviors improve output

1

. This approach avoids the limitations of prompt-based routing applied at inference time and supervised fine-tuning that bakes behavior into model weights

1

.

Early results show promise. PRISM raised overall scores by one to two points on MT-Bench, a test measuring how well AI follows instructions and maintains helpfulness

3

. For writing and safety tasks, personas helped; for raw knowledge questions, skipping the persona proved superior

3

.

What This Means for AI Users

The research carries immediate implications for anyone working with large language models. Hu's guidance is direct: "When you care more about alignment (safety, rules, structure-following, etc), be specific about your requirement; if you care more about accuracy and facts, do not add anything, just send the query"

1

. This runs counter to the over-engineered prompts that have proliferated across the AI community.

Source: TechRadar

Source: TechRadar

The study also highlights differences in model types, noting that reasoning models benefit more from context length while instruction-tuned models show greater sensitivity to personas. As the researchers plan to test PRISM with additional personas and refine its decision-making capabilities, this work could fundamentally reshape how we interact with AI systems

3

. The message is clear: model developers have optimized these systems extensively, and users should focus on providing clear tasks and relevant context rather than dictating how AI should think.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo