Asking AI for advice? Be VERY careful! Asking AI for advice? Be VERY careful!

AI-powered assistants are truly ubiquitous nowadays, whether you like them or not. Every major browser now has a native connection to a large language model (LLM, which is how professionals call what goes for artificial intelligence today) that users can summon and query with a single click or a hotkey combo. Those in the creative industries use AI all the time: while prone to hallucinations – it is a term here, not a subjective opinion stemming from frustration – such tools are good at collecting, listing, digging the info field and fetching the respective results. Everything important should be double-checked, though, because of the mentioned proneness; knowing that, a group of researchers from the University of Erlangen-Nuremberg in Germany and an institute in Belgium staged an experiment seeking to see how accurate Microsoft's Copilot is when prompted to give medical advice. The results were… alarming, to say the least.

Scientific investigation of AI’s accuracy in replying to medicine-related queries

To test Microsoft’s Copilot, the researchers tasked it with answering 10 commonly asked medical questions about each of the 50 most prescribed drugs in the United States, which yielded an array of 500 queries. The topics covered were typical use scenarios, mechanisms, usage instructions, common side effects, etc.

After initial appraisal of the responses for completeness – 23% were considered lacking essential information – they were submitted to a board of seven drug safety experts, who evaluated the actual accuracy of the information and advice given by AI. Without further ado, here’s the roundup:

  • only 54% of Copilot's responses actually aligned with what’s universally agreed upon in the medical circles;
  • 36% of the replies were ultimately deemed harmless;
  • 42% of the pieces of advice given, if followed, could lead to moderate or severe harm;
  • 22% of the responses could actually cause death.

How to prompt AI to get accurate replies?

There are several proven techniques that help get good, trustworthy answers from AI assistants; we’ll cover them in a separate piece, so stay tuned. For the purpose of the subject given in this post, here are some AI prompting best practices you can adopt to feel safer when asking an LLM for some information.

Be specific. Frame questions with clear and specific details, avoiding even a hint of ambiguity.

Limit Scope. Narrow your questions down; don’t force AI to dive into a broad inquiry if you want an actual answer and not ruminations watered with hallucinations.

Request Citations. Ask for sources or references to ensure that the information provided can be verified against established knowledge.

Follow-up questions. Make sure AI really understands what you’re after by asking follow-up questions to clarify or delve deeper into specific areas of interest.

Cross-verify. Google is still there; double-check everything essential and important from AI replies, especially when talking about health.

Author's other posts

Cybersecurity for the masses: best practices and tools
Article
Cybersecurity for the masses: best practices and tools
Here are some tips and tricks related to cybersecurity, from antivirus software and VPNs to password managers and secure browsing practices.
Is it on? How top AI makers responded DeepSeek’s success
Article
Is it on? How top AI makers responded DeepSeek’s success
DeepSeek shakes up AI, prompting OpenAI, Microsoft, and NVIDIA to react swiftly while Alibaba and ByteDance push forward with new LLMs in a competitive AI landscape.
Two weeks with DeepSeek: the scoop
Article
Two weeks with DeepSeek: the scoop
DeepSeek, a Chinese AI, soared past ChatGPT in the App Store. The startup claims to have spent only $5.6M on training the model. Let's see what's what after 2 weeks on the air.
Seven valid reasons to set up a VPN
Article
Seven valid reasons to set up a VPN
In case you were thinking a VPN setup or subscription isn't something you need, here are 7 real-world reasons that may have you reconsider.