Uncovering and Inducing Interpretable Causal Structure in Deep Learning Models | Atticus Geiger
Автор: Valence Labs
Загружено: 2024-04-18
Просмотров: 1063
Описание:
Portal is the home of the TechBio community. Join for more details on this talk and to connect with the speakers: https://portal.valencelabs.com/care
Summary: A faithful and interpretable explanation of an AI model’s behavior and internal structure is a high-level explanation that is human-intelligible but also consistent with the known, but often opaque low-level causal details of the model. We argue that the theory of causal abstraction provides the mathematical foundations for the desired kinds of model explanations. In the analysis mode, we uncover causal structure using interventions on model-internal states to assess whether an interpretable high-level causal model is a faithful description of a deep learning model. In the training mode, we induce interpretable causal structure using interventions during model training to simulate counterfactuals in the deep learning model's activation space. We show how to uncover and induce causal structures in a variety of case studies on deep learning models that reason over language and/or images.
Speaker: Atticus Geiger
Twitter Chandler: / chandlersquires
Twitter Dhanya: / dhanya_sridhar
Twitter Jason: / jasonhartford
~
Chapters
00:00 - Discussant Slide
04:48 - Introduction
08:16 - Causal Abstraction
11:47 - Interchange Interventions
32:15 - Distributed Alignment Search
Повторяем попытку...
Доступные форматы для скачивания:
Скачать видео
-
Информация по загрузке: