LLMs are known for, often criticised, and feared for their lack of explainability in their outputs and poor interpretability. Even when we peek into those black boxes and observe the steps taken to generate the final responses, we don’t understand how they ‘think’. This flaw had been a major showstopper for many real-world applications. For example, suppose a bank uses LLMs to decide whether a loan should be approved, but cannot explain the exact reasons behind the decision to the customer or the regulators. It will be very awkward.

Fast-forward to June 2025, three years after LLMs made the headlines, leading LLM providers started to expose LLMs’ internal reasoning process, hoping to ease the concern about using LLMs in critical decision-making processes. Google announced Thought Summaries, and Anthropic introduced an open-source tool to trace the thoughts of LLMs with attribution graphs in June 2025.

Referecnes