Stay up to date with notifications from The Independent

Notifications can be managed in browser preferences.

New study exposes critical flaw of ChatGPT

OpenAI launches GPT-5, its first flagship update to ChatGPT model in years
  • A new study by Stanford University revealed that artificial intelligence tools, including major chatbots like ChatGPT, struggle to differentiate between belief and fact.
  • The research found that all 24 large language models tested failed to consistently recognise false beliefs and statements, making them prone to hallucination and spreading misinformation.
  • This critical flaw has significant implications for the deployment of AI in high-stakes sectors such as law, medicine, and journalism, where distinguishing truth from falsehood is paramount.
  • Researchers evaluated the LLMs using 13,000 questions to assess their ability to discern beliefs, knowledge, and facts, exposing a structural weakness in their understanding.
  • The study, published in Nature Machine Intelligence, urges tech companies to urgently improve these models before their widespread use in critical domains.
In full

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in