Can We Trust AI Explanations? Evidence of Systematic Underreporting in Chain-of-Thought Reasoning

By: Deep Pankajbhai Mehta

Published: 2026-01-15

View on arXiv →
#cs.AI

Abstract

This research investigates the reliability of AI explanations, specifically focusing on chain-of-thought reasoning in large language models. The study provides evidence of systematic underreporting, where AI models fail to fully disclose all contributing factors to their conclusions. This highlights critical challenges for building trustworthy AI systems in real-world applications where transparency and accountability are paramount.

FEEDBACK

Projects

No projects yet

Can We Trust AI Explanations? Evidence of Systematic Underreporting in Chain-of-Thought Reasoning | ArXiv Intelligence