2021年最新-可解释机器学习相关研究最新论⽂、书籍、博
客、资源整理分享
理解(interpret)表⽰⽤可被认知(understandable)的说法去解释(explain)或呈现(present)。在机器学习的场景中,可解释性(interpretability)就表⽰模型能够使⽤⼈类可认知的说法进⾏解释和呈现。[Finale Doshi-Velez]
机器学习模型被许多⼈称为“⿊盒”。这意味着虽然我们可以从中获得准确的预测,但我们⽆法清楚地解释或识别这些预测背后的逻辑。但是我们如何从模型中提取重要的见解呢?要记住哪些事项以及我们需要实现哪些功能或⼯具?这些是在提出模型可解释性问题时会想到的重要问题。
本⽂整理了可解释机器学习相关领域最新的论⽂,书籍、资源、博客等,分享给需要朋友。
所有资源下载地址,见源地址。
本资源含了近年来热门的可解释⼈⼯智能(XAI)的前沿研究。从下图我们可以看到可解释/可解释AI的趋势。关于这个主题的出版物正在蓬
勃发展。
下图展⽰了XAI的⼏个⽤例。在这⾥,根据这个数字将出版物分成⼏个类别。
研究性论⽂
The elephant in the interpretability room: Why use attention as explanation when we have saliency methods, EMNLP Workshop 2020
Explainable Machine Learning in Deployment, FAT 2020
A brief survey of visualization methods for deep learning models from the perspective of Explainable AI, Information Visualization 2020
Explaining Explanations in AI, ACM FAT 2019
Machine learning interpretability: A survey on methods and metrics, Electronics, 2019
A Survey on Explainable Artificial Intelligence (XAI): Towards Medical XAI, IEEE TNNLS 2020
Interpretable machine learning: definitions, methods, and applications, Arxiv preprint 2019
Visual Analytics in Deep Learning: An Interrogative Survey for the Next Frontiers, IEEE Transactions on Visualization and Computer Graphics, 2019
Explainable Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI, Information Fusion, 2019
Evaluating Explanation Without Ground Truth in Interpretable Machine Learning, Arxiv preprint 2019
A survey of methods for explaining black box models, ACM Computing Surveys, 2018
Explaining Explanations: An Overview of Interpretability of Machine Learning, IEEE DSAA, 2018
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI), IEEE Access, 2018
Explainable artificial intelligence: A survey, MIPRO, 2018
How Convolutional Neural Networks See the World — A Survey of Convolutional Neural Network Visualization Methods, Mathematical Foundations of Computing 2018
Explainable Artificial Intelligence: Understanding, Visualizing and Interpreting Deep Learning Models, Arxiv 2017
Towards A Rigorous Science of Interpretable Machine Learning, Arxiv preprint 2017
Explaining Explanation, Part 1: Theoretical Foundations, IEEE Intelligent System 2017
Explaining Explanation, Part 2: Empirical Foundations, IEEE Intelligent System 2017
Explaining Explanation, Part 3: The Causal Landscape, IEEE Intelligent System 2017
Explaining Explanation, Part 4: A Deep Dive on Deep Nets, IEEE Intelligent System 2017
An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data, Ecological Modelling 2004
Review and comparison of methods to study the contribution of variables in artificial neural network models, Ecological Modelling 2003
书籍
Explainable Artificial Intelligence (xAI) Approaches and Deep Meta-Learning Models, Advances in Deep Learning Chapter 2020
Explainable AI: Interpreting, Explaining and Visualizing Deep Learning, Springer 2019
Explanation in Artificial Intelligence: Insights from the Social Sciences, 2017 arxiv preprint
Visualizations of Deep Neural Networks in Computer Vision: A Survey, Springer Transparent Data Mining for Big and Small Data 2017
Explanatory Model Analysis Explore, Explain and Examine Predictive Models
Interpretable Machine Learning A Guide for Making Black Box Models Explainable
An Introduction to Machine Learning Interpretability An Applied Perspective on Fairness, Accountability,
Transparency,and Explainable AI
关于python的书开源课程
Interpretability and Explainability in Machine Learning, Harvard University
⽂章
We mainly follow the taxonomy in the survey paper and divide the XAI/XML papers into the several branches.
1. Transparent Model Design
2. Post-Explanation
2.1 Model Explanation(Model-level)
2.2 Model Inspection
2.3 Outcome Explanation
2.3.1 Feature Attribution/Importance(Saliency Map)
2.4 Neuron Importance
2.5 Example-based Explanations
2.5.1 Counterfactual Explanations(Recourse)
2.5.2 Influential Instances
2.5.3 Prototypes&Criticisms
Uncategorized Papers on Model/Instance Explanation
Does Explainable Artificial Intelligence Improve Human Decision-Making?, AAAI 2021
Incorporating Interpretable Output Constraints in Bayesian Neural Networks, NeuIPS 2020
Towards Interpretable Natural Language Understanding with Explanations as Latent Variables, NeurIPS 2020
Learning identifiable and interpretable latent models of high-dimensional neural activity using pi-VAE, NeurIPS 2020
Generative causal explanations of black-box classifiers, NeurIPS 2020
Learning outside the Black-Box: The pursuit of interpretable models, NeurIPS 2020
Explaining Groups of Points in Low-Dimensional Representations, ICML 2020
Explaining Knowledge Distillation by Quantifying the Knowledge, CVPR 2020
Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems, IJCAI 2020
Machine Learning Explainability for External Stakeholders, IJCAI 2020
Py-CIU: A Python Library for Explaining Machine Learning Predictions Using Contextual Importance and Utility, IJCAI 2020
Machine Learning Explainability for External Stakeholders, IJCAI 2020
Interpretable Models for Understanding Immersive Simulations, IJCAI 2020
Towards Automatic Concept-based Explanations, NIPS 2019
Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead, Nature Machine Intelligence 2019
Interpretml: A unified framework for machine learning interpretability, arxiv preprint 2019
All Models are Wrong, but Many are Useful: Learning a Variable’s Importance by Studying an Entire Class of Prediction Models Simultaneously, JMLR 2019
On the Robustness of Interpretability Methods, ICML 2018 workshop
Towards A Rigorous Science of Interpretable Machine Learning, Arxiv preprint 2017
Object Region Mining With Adversarial Erasing: A Simple Classification to Semantic Segmentation Approach, CVPR 2017
LOCO, Distribution-Free Predictive Inference For Regression, Arxiv preprint 2016
Explaining data-driven document classifications, MIS Quarterly 2014
评测⽅法
Evaluations and Methods for Explanation through Robustness Analysis, arxiv preprint 2020
Evaluating and Aggregating Feature-based Model Explanations, IJCAI 2020
Sanity Checks for Saliency Metrics, AAAI 2020
A benchmark for interpretability methods in deep neural networks, NIPS 2019
Methods for interpreting and understanding deep neural networks, Digital Signal Processing 2017
Evaluating the visualization of what a Deep Neural Network has learned, IEEE Transactions on Neural Networks and Learning Systems 2015
Python开源库
Related Repositories
版权声明:本站内容均来自互联网,仅供演示用,请勿用于商业和其他非法用途。如果侵犯了您的权益请与我们联系QQ:729038198,我们将在24小时内删除。
发表评论