Asking AI for advice? Be VERY careful! Asking AI for advice? Be VERY careful!

AI-powered assistants are truly ubiquitous nowadays, whether you like them or not. Every major browser now has a native connection to a large language model (LLM, which is how professionals call what goes for artificial intelligence today) that users can summon and query with a single click or a hotkey combo. Those in the creative industries use AI all the time: while prone to hallucinations – it is a term here, not a subjective opinion stemming from frustration – such tools are good at collecting, listing, digging the info field and fetching the respective results. Everything important should be double-checked, though, because of the mentioned proneness; knowing that, a group of researchers from the University of Erlangen-Nuremberg in Germany and an institute in Belgium staged an experiment seeking to see how accurate Microsoft's Copilot is when prompted to give medical advice. The results were… alarming, to say the least.

Scientific investigation of AI’s accuracy in replying to medicine-related queries

To test Microsoft’s Copilot, the researchers tasked it with answering 10 commonly asked medical questions about each of the 50 most prescribed drugs in the United States, which yielded an array of 500 queries. The topics covered were typical use scenarios, mechanisms, usage instructions, common side effects, etc.

After initial appraisal of the responses for completeness – 23% were considered lacking essential information – they were submitted to a board of seven drug safety experts, who evaluated the actual accuracy of the information and advice given by AI. Without further ado, here’s the roundup:

  • only 54% of Copilot's responses actually aligned with what’s universally agreed upon in the medical circles;
  • 36% of the replies were ultimately deemed harmless;
  • 42% of the pieces of advice given, if followed, could lead to moderate or severe harm;
  • 22% of the responses could actually cause death.

How to prompt AI to get accurate replies?

There are several proven techniques that help get good, trustworthy answers from AI assistants; we’ll cover them in a separate piece, so stay tuned. For the purpose of the subject given in this post, here are some AI prompting best practices you can adopt to feel safer when asking an LLM for some information.

Be specific. Frame questions with clear and specific details, avoiding even a hint of ambiguity.

Limit Scope. Narrow your questions down; don’t force AI to dive into a broad inquiry if you want an actual answer and not ruminations watered with hallucinations.

Request Citations. Ask for sources or references to ensure that the information provided can be verified against established knowledge.

Follow-up questions. Make sure AI really understands what you’re after by asking follow-up questions to clarify or delve deeper into specific areas of interest.

Cross-verify. Google is still there; double-check everything essential and important from AI replies, especially when talking about health.

Author's other posts

macOS 15.4: what’s new?
Article
macOS 15.4: what’s new?
Unveiling macOS 15.4: AI-powered Photo, enhanced Mail, seamlessly integrated Password app, faster Safari, ChatGPT integration, Quick Start & more.
BSOD is now black; old practices to avoid it are still valid
Article
BSOD is now black; old practices to avoid it are still valid
Microsoft makes BSOD black, not blue; here's some history of the trademark screen, and tips about how to avoid seeing it.
iOS 18.4: what’s new?
Article
iOS 18.4: what’s new?
iOS 18.4 is out, and so is iPadOS 18.4. Let's see what's inside that makes it worth upgrading to this version.
How to install Windows 11 without a Microsoft account
Article
How to install Windows 11 without a Microsoft account
Microsoft's Windows 11 update will ban installation without a Microsoft account. Discover workarounds like the "OOBE\BYPASSNRO" command and more tricks.