Audio-XAI

Audio XAI

PyPI version

Python Boilerplate contains all the boilerplate you need to create a Python package.

Features

Documentation

Documentation is built with Zensical and deployed to GitHub Pages.

API documentation is auto-generated from docstrings using mkdocstrings.

Docs deploy automatically on push to master via GitHub Actions. To enable this, go to your repo’s Settings > Pages and set the source to GitHub Actions.

Development

To set up for local development:

# Clone your fork
git clone git@github.com:your_username/Audio-XAI.git
cd Audio-XAI

# Install in editable mode with live updates
uv tool install --editable .

This installs the CLI globally but with live updates - any changes you make to the source code are immediately available when you run audio_xai.

Run tests:

uv run pytest

Run quality checks (format, lint, type check, test):

just qa

Author

Audio XAI was created in 2026 by Piotr Kitłowski.

Built with Cookiecutter and the audreyfeldroy/cookiecutter-pypackage project template.

1. General Information and Project Objective

The main objective of the project is to investigate the perceptual fragility of explanations (XAI methods) for deep learning models in the audio domain while keeping predictions unchanged.

2. Planned scope of experiments

3. Planned Program Features

4. Planned Technology Stack

The project will implement a robust base structure, automatically generated by tool cookiecutter.

5. Project schedule

| Deadline dates in 2026 | Planned scope of work and progress | Status | | :————————: | :— | :—: | | 30.03 - 05.04 | Repository configuration (Cookiecutter, Ruff, Uv). Defining the directory structure and ensuring that audio files remain immutable. | ✔ | | 06.04 - 12.04 | Connecting W&B/TensorBoard. Training base classifiers using the PyTorch Lightning framework. (Estimated resource requirements: 15 hours of GPU computation) | | | 13.04 - 19.04 | Implementation of explanation-generating (XAI) modules in clean code, after first exporting experiments from notebooks. Writing the first tests. | | | 20.04 - 26.04 | Separating configuration from executable code. Preparing baseline attacks on attribution maps using standard distance metrics. | | | 27.04 - 03.05 | Implementation of PESQ/STOI/PEAQ metric approximations directly into the attack optimization loop (generation of perceptual perturbations). | | | 04.05 - 10.05 | Launch of the main research experiments on a dedicated cluster. (Estimated resource requirements: 25–30 hours of GPU computing for iterative processes). | | | 11.05 - 17.05 | Scripting the execution of the entire experiment using the just tool and CLI libraries (e.g., typer). Aggregating tables containing the results. | | | 18.05 - 21.05 | Finalization of the work: creating documentation and clear instructions for using the finished system. Organizing the code in accordance with PEP8. | | | 21.05 - 31.05 | Preparation of the paper(?) | |