Corti, a company focused on healthcare AI infrastructure, announced a significant advancement in artificial intelligence transparency by securing the top position on the Hugging Face Mechanistic Interpretability Benchmark. This benchmark, maintained by researchers from institutions including MIT, Stanford, Cambridge, Boston, and ETH Zurich, serves as the industry standard for evaluating interpretability in AI. Corti outperformed well-known approaches from Meta, DeepMind-affiliated researchers, and Harvard with its innovative Gradient Interaction Modification technique, demonstrating that specialized labs can drive fundamental research that shapes how AI systems are built, understood, and applied.
Health Technology Insights: Slone Partners Places Jeffrey Kasten as Chief Business Officer at Vertero Therapeutics
The company describes this advancement as a “surgical” improvement, reflecting a shift in the AI field from simply increasing model size to understanding the internal mechanisms that govern model behavior. Lars Maaløe, Co-Founder and Chief Technology Officer at Corti, emphasized the importance of this change, explaining that most AI development has relied on trial-and-error. Maaløe stated that the GIM method exposes the underlying logic of models in ways conventional tools cannot, allowing improvements to be made with precision rather than guesswork.
Health Technology Insights: Norgine Expands Liver and Specialty Portfolio With Vir Biotechnology In-Licensing Agreement
Traditional interpretability techniques often examine individual neurons in isolation, which can be misleading. A neuron may appear irrelevant if it only contributes to an outcome when combined with others. GIM addresses this by analyzing how components interact simultaneously, revealing the actual circuits that drive model outputs. This approach allows researchers to identify the true causes behind a model’s decisions, trace errors such as hallucinations to their origins, and perform deep analysis on production models quickly, which is particularly important for compliance with regulations like the EU AI Act.
High-stakes fields such as healthcare require a higher level of explainability compared to consumer applications. Andreas Cleve, Co-Founder and Chief Executive Officer at Corti, highlighted that models affecting clinical decisions must provide precise reasoning that can be traced and validated. Cleve explained that the GIM tool was initially developed to meet the needs of Corti’s own clinical infrastructure, but its capabilities now benefit the broader AI community by enabling safer and more reliable model deployment.
This development addresses a growing challenge across regulated industries, including finance, autonomous systems, and government services. As transparency requirements become stricter, tools like GIM, initially designed for healthcare, are increasingly necessary to ensure AI systems operate safely and effectively. Corti’s achievement demonstrates that understanding AI at a mechanistic level is not only possible but essential for the responsible and scalable application of artificial intelligence in critical sectors.
Health Technology Insights: Ferronova Raises $6 Million to Boost Image-Guided Cancer Surgery
To participate in our interviews, please write to our HealthTech Media Room at info@intentamplify.com


