2 Sources
2 Sources
[1]
Brits believe the bots even though they spout nonsense
Consumer group Which? warns AI assistants can dish out unclear, risky, or downright daft advice AI assistants can sometimes provide misleading or incorrect answers. However, almost half of British consumers using the services put more faith in them than they maybe should. Consumer stalwart Which? put the tools through their paces and found that the consumer advice dispensed could be unclear, risky, or downright dangerous if followed. It's something the IT world is all too familiar with. AI-powered assistants have their place, but it is also important to understand their limitations and spot arrant emissions. Which? surveyed more than 4,000 UK adults about their use of AI and also put 40 questions around consumer issues such as health, finance, and travel to six bots - ChatGPT, Google Gemini, Gemini AI Overview, Copilot, Meta AI, and Perplexity. Things did not go well. Meta's AI answered correctly just over 50 percent of the time in the tests, while the most widely used AI tool, ChatGPT, came second from bottom at 64 percent. Perplexity came top at 71 percent. While different questions might yield different results, the conclusion is clear: AI tools don't always come up with the correct answer. The problem is that consumers trust the output. According to Which?, just over half (51 percent) of the respondents use AI to search the web. Of these, almost half (47 percent) said "they trusted the information they received to a 'great' or 'reasonable' extent." Which? said the figure rose to 65 percent for frequent users. Then there were the sources used by the AI services. Where references were clear, some used old forum posts, while others relied on sources such as Reddit threads. Although these can sometimes be valid sources of information, they might not be as authoritative as the confident tone of an AI chatbot indicates. Which? found the chatbots generated wrong information all too frequently, noting: "As many as one in six (17 percent) people surveyed said they rely on AI for financial advice, yet responses to many money queries were worrying." These included tax code checks or advice on ISA allowances that could easily leave a user in hot water. Andrew Laughlin, tech expert at Which? said: "Everyday use of AI is soaring, but we've found that when it comes to getting the answers you need, the devil is in the details. "Our research uncovered far too many inaccuracies and misleading statements for comfort, especially when leaning on AI for important issues like financial or legal queries." As the use of AI assistants continues to rise, so too do the risks. While the IT industry is aware of the dangers involved - a recent analysis showed that AI-assisted developers can produce three or four times the code of their unassisted peers but also generate ten times more security issues - consumers might be forgiven for being less tech-savvy, particularly considering the hype surrounding the technology. Laughlin ended with a warning: "When using AI, always make sure to define your question clearly, and check the sources the AI is drawing answers from. For particularly complex issues, always seek professional advice - particularly for medical queries - before making major financial decisions or embarking on legal action." ®
[2]
UK consumers warned over AI chatbots giving inaccurate financial advice
Which? study of ChatGPT, Copilot and others uncovers incorrect and misleading tips on investments, tax and insurance Artificial intelligence chatbots are giving inaccurate money tips, offering British consumers misleading tax advice and suggesting they buy unnecessary travel insurance, research has revealed. Tests on the most popular chatbots found Microsoft's Copilot and ChatGPT advised breaking HMRC investment limits on Isas; ChatGPT wrongly said it was mandatory to have travel insurance to visit most EU countries; and Meta's AI gave incorrect information about how to claim compensation for delayed flights. Google's Gemini advised withholding money from a builder if a job went wrong, a move that the consumer organisation Which? said risked exposing the consumer to a claim of breach of contract. Which? said its research, conducted by putting 40 questions to the rival AI tools, "uncovered far too many inaccuracies and misleading statements for comfort, especially when leaning on AI for important issues like financial or legal queries". Meta's AI received the worst score, followed by ChatGPT; Copilot and Gemini scored slightly higher. The highest score was given to Perplexity, an AI known for specialising in search. Estimates on the number of people in the UK using AI for financial advice range from one in six to as many as half. When asked about their experiences, Guardian readers said they had recently used AI to find the best credit cards to use abroad, for advice on how to reduce investment fees, and to secure good deals on household appliances - including an artist who used it to get a good price on a ceramic kiln. Several said they were pleased with the results, but Kathryn Boyd, 65, who runs a fashion business in Wexford, Ireland, said she turned to ChatGPT for advice on her self-employed tax and it used an out-of-date code. "It just gave me all the wrong information," she said, adding that she had to correct it at least three times. "My concern is that I am very well-informed but ... other people asking the same question may easily have relied on the assumptions used by ChatGPT which were just plain wrong - wrong tax credits, wrong tax and insurance rates etc." When the Which? researchers asked the AI tools how to claim a tax refund from HMRC, ChatGPT and Perplexity presented links to premium tax-refund companies alongside the free government service, which was "worrying" as "these companies are notorious for charging high fees and adding on spurious charges". After they placed a deliberate mistake in a question about the ISA allowance, asking: "How should I invest my £25k annual ISA allowance?", ChatGPT and Copilot failed to notice the correct allowance was £20,000 and gave advice that could have led a consumer to oversubscribe, breaching HMRC rules. NEW The Financial Conduct Authority regulator said: "Unlike regulated advice provided by authorised firms, any advice provided by these general-purpose AI tools are not covered by the Financial Ombudsman Service and the Financial Services Compensation Scheme." NEWENDS In response, Google said it was transparent about the limitations of generative AI and that Gemini reminded users to double check information and consult professionals on legal, medical and financial matters. A spokesperson for Microsoft said: "With any AI system, we encourage people to verify the accuracy of content, and we remain committed to listening to feedback to improve our AI technologies." Open AI said: "Improving accuracy is something the whole industry's working on. We're making good progress and our latest default model, GPT-5, is the smartest and most accurate we've built."
Share
Share
Copy Link
Consumer group Which? finds major AI chatbots providing incorrect financial, tax, and legal advice while nearly half of British users place significant trust in these tools despite frequent inaccuracies.

A comprehensive study by UK consumer group Which? has revealed alarming discrepancies between consumer trust in AI chatbots and their actual accuracy, particularly when dispensing financial and legal advice. The research, which tested six major AI platforms including ChatGPT, Google Gemini, Microsoft Copilot, Meta AI, and Perplexity, found that nearly half of British users place significant trust in AI-generated information despite frequent inaccuracies
1
.The study surveyed over 4,000 UK adults and subjected the AI tools to 40 consumer-focused questions covering health, finance, and travel topics. Results showed troubling accuracy rates, with Meta's AI answering correctly just 51% of the time, while ChatGPT—the most widely used AI tool—managed only 64% accuracy. Perplexity performed best at 71% correct responses
1
.The research uncovered particularly concerning issues with financial advice, an area where 17% of survey respondents reported relying on AI assistance. Multiple instances of potentially harmful guidance were documented, including ChatGPT and Microsoft Copilot advising users to exceed HMRC investment limits on Individual Savings Accounts (ISAs)
2
.When researchers deliberately included an incorrect ISA allowance figure of £25,000 instead of the actual £20,000 limit, both ChatGPT and Copilot failed to identify the error and provided advice that could lead consumers to breach HMRC regulations. Additionally, when asked about tax refunds, these platforms presented links to premium tax-refund companies alongside free government services, potentially exposing users to unnecessary fees
2
.The study highlighted significant concerns about the sources AI chatbots use to generate responses. Where references were identifiable, many tools relied on outdated forum posts and Reddit threads rather than authoritative sources. This reliance on potentially unreliable information becomes particularly problematic given the confident tone AI chatbots typically employ when delivering responses
1
.Meta's AI provided incorrect information about flight delay compensation claims, while Google's Gemini advised withholding payment from builders for unsatisfactory work—advice that Which? warned could expose consumers to breach of contract claims. ChatGPT incorrectly stated that travel insurance was mandatory for visiting most EU countries
2
.Related Stories
The Financial Conduct Authority emphasized a critical distinction between AI-generated advice and regulated financial guidance. Unlike advice from authorized firms, AI chatbot recommendations lack protection from the Financial Ombudsman Service and Financial Services Compensation Scheme, leaving consumers without recourse when receiving incorrect information
2
.Tech companies acknowledged the limitations while emphasizing ongoing improvements. Google highlighted its transparency about generative AI limitations and built-in reminders for users to verify information. Microsoft encouraged content verification and committed to incorporating feedback for improvements. OpenAI pointed to industry-wide accuracy improvements and claimed their latest GPT-5 model represents their most accurate system to date
2
.Real-world user experiences reflect these concerns. Kathryn Boyd, who runs a fashion business in Ireland, reported receiving outdated tax information from ChatGPT that required multiple corrections, expressing worry that less-informed users might rely on such incorrect guidance
2
.Summarized by
Navi
[1]
1
Technology

2
Business and Economy

3
Business and Economy
