back
Get SIGNAL/NOISE in your inbox daily
Recent work from Anthropic and others claims that LLMs’ chains of thoughts can be “unfaithful”. These papers make an important point: you can’t take everything in the CoT at face value. As a result, people often use these results to conclude the CoT is useless for analyzing and monitoring AIs. Here, instead of asking whether the CoT always contains all information relevant to a model’s decision-making in all problems, we ask if it contains enough information to allow developers to monitor models in practice. Our experiments suggest that it might.
Recent Stories
Jan 19, 2026
Frontiers | Artificial Intelligence vs Human Evaluation of Anesthesia Education Videos: A Comparative Analysis Using Validated Quality Scales
Background: YouTube has become an increasingly popular platform for medical education, yet the accuracy and educational quality of anesthesia-related videos ...
Jan 19, 2026The Race to Build the DeepSeek of Europe Is On
As Europe’s longstanding alliance with the US falters, its push to become a self-sufficient AI superpower has become more urgent.
Jan 18, 2026Ed Zitron on big tech, backlash, boom and bust: ‘AI has taught us that people are excited to replace human beings’
His blunt, brash scepticism has made the podcaster and writer something of a cult figure. But as concern over large language models builds, he’s no longer the outsider he once was