RSS VentureBeat

Meta researchers open the LLM black box to repair flawed AI reasoning

Researchers have developed Circuit-based Reasoning Verification (CRV), a novel technique to inspect and correct large language model (LLM) reasoning. CRV monitors internal "reasoning circuits" within an LLM by building a computational graph from its internal activations. This method accurately detects reasoning errors by observing these computational traces. A key breakthrough is the ability to use this insight for real-time interventions to fix faulty reasoning. This advancement could significantly improve the trustworthiness and reliability of AI applications, especially in enterprise settings. Current methods for verifying LLM reasoning, such as black-box and gray-box approaches, lack the ability to explain the root cause of computational failures. CRV, as a white-box approach, treats LLMs as executing latent algorithms within specialized neuron circuits. By making LLMs interpretable with transcoders, CRV can observe how information flows through these circuits. It then extracts "structural fingerprints" from an attribution graph of this flow to predict reasoning correctness. CRV demonstrated superior performance over existing methods in detecting errors across various datasets. Crucially, the technique can pinpoint specific computational flaws, enabling targeted interventions that correct mistakes on the fly, as shown in a case study involving an order-of-operations error. This work represents a significant step towards a more rigorous science of AI interpretability and control.
favicon
bsky.app
AI and ML News on Bluesky @ai-news.at.thenote.app
favicon
venturebeat.com
venturebeat.com
Create attached notes ...