Programme

Model Readability and Explanatory Alignment

  • Talk detail
  • 11:45

Session

Panel 1: AI for Modeling and Forecasting

Time

11:45

Session window

11:00 - 12:00

Abstract

Feature attribution is the dominant paradigm for explaining the predictions of complex machine learning models like neural networks. However, most existing methods offer little guarantee of reflecting the model's prediction-making process. We define the notion of explanatory alignment and argue that it is central to trustworthy predictive modeling: in short, it requires that explanations directly underlie predictions rather than serve as rationalizations. We present model readability as a design principle enabling alignment, and Pointwise-interpretable Networks (PiNets) as a modeling framework to pursue it in a deep learning context. PiNets combine statistical intelligence with a pseudo-linear structure that yields instance-wise linear predictions in an arbitrary feature space. We illustrate their use on image classification and segmentation tasks, demonstrating that PiNets produce explanations that are not only aligned by design but also faithful across other dimensions: meaningfulness, robustness, and sufficiency.

Speakers

Corentin Lobet

Corentin Lobet

Sant’Anna School of Advanced Studies

Corentin Lobet is a Ph.D. Student at the Scuola Superiore Sant’Anna. His research spans from statistical modelling and explainable AI to the role of model errors in shaping inequalities and instability in society.