Making News Recommendations Explainable with Large Language Models

Successful AI Ethics & Governance at Scale: Bridging The Organizational and Implementation Gaps

Discover What Every Neuron in the Llama Model Does

Successful AI Ethics & Governance at Scale: Bridging The Interpretation Gap

FormulaFeatures: A Tool to Generate Highly Predictive Features for Interpretable Models

Sparse Autoencoders, Additive Decision Trees, and Other Emerging Topics in AI Interpretability

Beyond the Blind Zone

Feature Selection with Hierarchical Clustering for Interpretable Models

Get more out of XAI: 10 Tips

TE2Rules: Explaining “Why did my model say that?”