See every side of every news story
Published loading...Updated

Experiments show adding CoT windows to chatbots teaches them to lie less obviously

Summary by TechXplore
Over the past year, AI researchers have found that when AI chatbots such as ChatGPT find themselves unable to answer questions that satisfy users' requests, they tend to offer false answers. In a new study, as part of a program aimed at stopping chatbots from lying or making up answers, a research team added Chain of Thought (CoT) windows. These force the chatbot to explain its reasoning as it carries out each step on its path to finding a final…

4 Articles

All
Left
Center
2
Right
Think freely.Subscribe and get full access to Ground NewsSubscriptions start at $9.99/yearSubscribe

Bias Distribution

  • 100% of the sources are Center
100% Center
Factuality

To view factuality data please Upgrade to Premium

Ownership

To view ownership data please Upgrade to Vantage

the-decoder.de broke the news in Germany on Monday, March 31, 2025.
Sources are mostly out of (0)

You have read out of your 5 free daily articles.

Join us as a member to unlock exclusive access to diverse content.