Resource list on adversarial explainable AI (AdvXAI)
Top 85.7% on sourcepulse
This repository serves as a curated list of academic papers focusing on adversarial attacks and defenses within Explainable Artificial Intelligence (XAI). It aims to consolidate research on the vulnerabilities of XAI methods, providing a resource for researchers and practitioners interested in the security and trustworthiness of AI explanations.
How It Works
The project compiles a comprehensive survey of literature that explores how adversarial manipulations can fool or compromise the integrity of XAI techniques. It categorizes various attack strategies and defense mechanisms, offering a unified notation and taxonomy to structure the field of Adversarial Explainable AI (AdvXAI). The goal is to highlight existing insecurities and propose future research directions for developing more robust interpretation methods.
Quick Start & Requirements
This repository is a collection of research papers and does not involve direct code execution or installation. All requirements are related to accessing and reading academic publications.
Highlighted Details
Maintenance & Community
The primary contributor is Hubert Baniecki, with a published survey paper in Information Fusion. The repository is a static collection of research links.
Licensing & Compatibility
The repository itself does not have a license as it is a collection of links to academic papers. The licensing of the individual papers would be governed by their respective publishers.
Limitations & Caveats
This repository is a curated list of papers and does not provide any code or tools for implementing or testing adversarial XAI techniques. Its utility is limited to academic research and literature review.
8 months ago
Inactive