Why AI chatbots often make mistakes

 They deliberately designed the questions to encourage the chatbot to provide misleading answers. This standar stress-testing technique in AI security research is known as red kubuing.


The studi also tested the free versions of each chatbot bentuk available in February 2025. Although paid versions and newer releases may perform better.

Why AI chatbots often make mistakes

In fact, most people use the free version. Most of the health questions they ask are not carefully worded. The conditions in this studi reflect how people actually use AI chatbots.


This article's findings are not isolated. They emerge amidst a growing bodi of evidence that suggests a consistent picture.



For example, a February 2026 studi in the journal Nature Medicine revealed quite surprising findings.


Chatbots are actually capable of providing correct medical answers almost 95% of the time. However, when used by real-world humans, the correct answer rate drops drastically to below 35%. This sosok is no better than that of people who don't use chatbots at all.


Simply put, the question is not just whether the chatbot provides the right answer, but rather: "Can a lay pemakai understand and use the answer correctly?"


Confirmed by other studies

Additionally, a recent studi published in the journal JAMA Network Open tested 21 leading AI models. The researchers asked the models to generate a kisaran of possible medical diagnoses.


When these models were given only basic rincis (such as the patient's age, gender, and symptoms) they struggled and failed to suggest the correct set of conditions in more than 80% of trials.

monetary precarity gases

However, after researchers included physical examination findings and laboratory results, the akiracy rate jumped to above 90%.


Postingan populer dari blog ini

On Friday, pointing out worries over unapproved settings up and also prohibited tasks,

The president’s good fortunes were buoyed

Protect your eyes from mechanical and UV damage