Anthropic scientists expose how AI actually ‘thinks’
Anthropic has developed a new method for peering inside large language models (LLMs) like Claude, revealing for the first time how these AI systems process information and make decisions.
The research, published today in two papers (available here and here), shows these models are more sophisticated than previously understood — they plan ahead when writing poetry, use the same internal blueprint to interpret ideas regardless of language, and sometimes even work backward from a desired outcome instead of simply building up from the facts.