Table 1 Overview of questions that can be answered by SemanticLens
From: Mechanistic understanding and validation of large AI models with SemanticLens
Type | Question to the model \({\mathcal{M}}\) | Results |
---|---|---|
Search | ‘Has my model learned to encode a specific concept?’ through convenient ‘search-engine’-like text or image descriptions | |
‘Which components have encoded a concept, how is it used and which data are related?’ | Fig. 2d | |
Describe | ‘What concepts has my model learned?’ in a structured, condensed and understandable manner through textual descriptions | |
‘What concepts are contributing to a decision and how?’ by visualizing concept interactions throughout the model | ||
‘What do I not yet understand about my model?’, by offering an understanding of unexpected concepts and their role for the model and origin in data | ||
Compare | ‘What concepts are shared between two models, and which are unique to each one?’ by comparing learned concepts qualitatively and quantitatively | |
‘How do my concepts in my model change when the architecture or training changes?’ by comparing and tracking semantics of components | ||
Audit | ‘Does my model rely on valid information only?’ by separating learned concepts into valid, spurious and unexpected knowledge | |
Evaluate | ‘How interpretable is my model?’ with easy-to-compute measures | Fig. 5b |
‘How can I improve the interpretability of my model?’ by evaluating interpretability measures when the model architecture or training procedure change |