Interpretability Latent Space Method: Exploiting Shapley Representation to Explain Latent Space

Zitu Liu, Jiawang Li, Yue Liu, Qun Liu, Guoyin Wang, Yike Guo

Research output: Chapter in Book/Conference Proceeding/ReportConference Paper published in a bookpeer-review

Abstract

Shapley values have become one of the most popular interpretation methods for feature attribution. These methods attribute the prediction of the input by the machine learning model to its basic feature. However, most of the previous work is based on Shapley's interpretation after the training is completed because of Shapley's calculation requirements (exponential time complexity), so that they can focus on post-hoc Shapley explanations. Therefore, we suggest trying to use Shapley value itself as a latent representation in the deep model to guide model training. First, we extract the latent space of the generation model as input of Shapley network. Then, we utilize Shapley network, which provides layer-wise transformation in the same forward pass and produce Shapley representation. Further, we apply Shapley representation to the calculation of the new model to explain the training process of the model. Finally, the experimental results show on the two datasets of MNIST and Fashion MNIST that our proposed method has certain interpretability.

Original languageEnglish
Title of host publicationProceedings - 2021 7th International Conference on Big Data and Information Analytics, BigDIA 2021
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages87-92
Number of pages6
ISBN (Electronic)9781665424660
DOIs
Publication statusPublished - 2021
Externally publishedYes
Event7th International Conference on Big Data and Information Analytics, BigDIA 2021 - Chongqing, China
Duration: 29 Oct 202131 Oct 2021

Publication series

NameProceedings - 2021 7th International Conference on Big Data and Information Analytics, BigDIA 2021

Conference

Conference7th International Conference on Big Data and Information Analytics, BigDIA 2021
Country/TerritoryChina
CityChongqing
Period29/10/2131/10/21

Bibliographical note

Publisher Copyright:
© 2021 IEEE.

Keywords

  • Interpretability
  • Latent space
  • Shapley value

Fingerprint

Dive into the research topics of 'Interpretability Latent Space Method: Exploiting Shapley Representation to Explain Latent Space'. Together they form a unique fingerprint.

Cite this