Science Score: 26.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Committers with academic emails
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (14.0%) to scientific vocabulary
Keywords
Repository
XAI-Tris
Basic Info
Statistics
- Stars: 3
- Watchers: 2
- Forks: 2
- Open Issues: 0
- Releases: 0
Topics
Metadata Files
README.md
XAI-Tris
This repository contains all the experiments presented in the paper: "XAI-TRIS: Non-linear benchmarks to quantify ML explanation performance".
We use Pipfiles to create Python environments. Our main experiments use Captum, however inclusion of some methods (PatternNet, PatternAttribution, DTD) from the innvestigate library requires particular dependencies, such as Python=3.7, tensorflow==1.13.1/1.14, and keras==2.2.4. If you do not wish to include these methods, you are able to remove tensorflow and keras from the Pipfile and change any specific package requirements to * along with the Python version of choice (we have also tested this with Python 3.10).
To setup pipenv, run:
shell
pipenv install
and then
shell
pipenv shell
In four steps we can reproduce the results: (i) generate tetromino data, (ii) train models (iii) apply XAI methods, (iv) run the evaluation steps and generate plots.
Generate data
Download ImageNet data
If one wishes to use the data from the ImageNet-1k, we sourced it from HuggingFace by making an account and agreeing to the license terms. Please respect these license terms when using this data. Download and extract images into a folder imagenet_images/ in the top level of this repository. We used the N=50,000 validation set due to the appropriate volume size for our datasets and benchmarks.
64x64 Data
Once ImageNet data is downloaded and extracted above (if using it), configuration is set by the data/data_config.json file, with fields explained in data/README.md. Generate data via:
shell
python -m data.generate_data
8x8 Data
One can also generate the original 8x8 data (without ImageNet backgrounds), which is quicker and less computationally demanding to run, and also produces interesting results. Configuration here is set by the data/data_config_8by8.json file. Generate data via:
shell
python -m data.generate_data data_config_8by8
Train models
Update the data_path parameter of the training/training_config.json with the path to the folder of freshly generated pickle files containing data, of the form artifacts/tetris/data/YYYY-MM-DD-HH-mm.
To train models for a particular scenario, background, and experiment number:
Torch
shell
python -m training.train_models SCENARIO BACKGROUND EXP_NUM OUT_DIR
where SCENARIO=[linear|multiplicative|translations_rotations|xor], BACKGROUND=[uncorrelated|correlated|imagenet], EXP_NUM=[0,...,N], and OUT_DIR=[str]. For example python -m training.train_models linear uncorrelated 0 neurips_training will train models for the linear scenario with uncorrelated (WHITE) background for the 0th experiment and output models and logs to the artifacts/tetris/training/neurips_training directory.
Keras
shell
python -m training.train_models_keras SCENARIO BACKGROUND EXP_NUM OUT_DIR
where input args are the same as specified above.
Train all models
As you may have noticed, you will have to run this individually for all scenario-background pairs. You can also input regex into the SCENARIO and BACKGROUND input args, so running python -m training.train_models * * EXP_NUM OUT_DIR will train all scenario and background types for PyTorch.
One can use a somewhat approach like ./train_all.sh which simply runs all parameterizations one after another. We can also provide a script combining all parameterizations for each frameworks if needed.
XAI Methods
Update the data_path, training_output, and num_experiments parameters of the xai/xai_config.json file with the same data_path as used in the training step, and the corresponding training output path of the form artifacts/tetris/training/OUT_DIR.
To calculate explanations for a particular scenario and background and across all experiments:
Torch
shell
python -m xai.calculate_explanations SCENARIO BACKGROUND OUT_DIR
where SCENARIO=[linear|multiplicative|translations_rotations|xor], BACKGROUND=[uncorrelated|correlated|imagenet], and OUT_DIR=[str] as in the training step. Note: as of 14/06/23, you can't use regex for SCENARIO and BACKGROUND as in the training step, as it is incompatible with the input to the next step. If/when changed, this README will be updated.
Keras
shell
python -m xai.calculate_explanations_keras SCENARIO BACKGROUND OUT_DIR
where input args are the same as specified above.
Run evaluation metrics
Update the parameters data_path, training_path, xai_path, and num_experiments of eval/eval_config.json, and specify the desired output directory here too, in the form ./artifacts/tetris/eval/OUT_DIR.
Torch
shell
python -m eval.quantitative_metrics SCENARIO BACKGROUND
where SCENARIO=[linear|multiplicative|translations_rotations|xor] and BACKGROUND=[uncorrelated|correlated|imagenet]. Note: as of 14/06/23, you can't use regex for SCENARIO and BACKGROUND as in the training step, as it is incompatible with the input to the next step. If/when changed, this README will be updated.
Keras
shell
python -m eval.quantitative_metrics_keras SCENARIO BACKGROUND OUT_DIR
where input args are the same as specified above.
Generate plots
With the same parameters of eval/eval_config.json as before, simply run:
Qualitative plots (local and global)
shell
python -m eval.plot_qualitative
Quantitative plots
shell
python -m eval.plot_quantitative
Questions and Issues
If you have any questions and/or issues with the above, please do not hesitate to contact the authors!
Owner
- Name: QAI Labs
- Login: braindatalab
- Kind: organization
- Location: Berlin, Germany
- Website: tu.berlin/uniml
- Twitter: QAILabs
- Repositories: 13
- Profile: https://github.com/braindatalab
Quality in Artificial Intelligence Labs Berlin
GitHub Events
Total
- Push event: 2
- Create event: 1
Last Year
- Push event: 2
- Create event: 1
Committers
Last synced: about 2 years ago
Top Committers
| Name | Commits | |
|---|---|---|
| Benny Clark | b****2@g****m | 10 |
| Benny Clark | b****2@g****m | 1 |
Issues and Pull Requests
Last synced: 6 months ago
All Time
- Total issues: 0
- Total pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Total issue authors: 0
- Total pull request authors: 0
- Average comments per issue: 0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Issue authors: 0
- Pull request authors: 0
- Average comments per issue: 0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
Pull Request Authors
Top Labels
Issue Labels
Pull Request Labels
Dependencies
- captum *
- h5py ==2.10.0
- innvestigate *
- ipykernel *
- jupyter *
- keras ==2.2.4
- matplotlib *
- mlxtend *
- numpy *
- openpyxl *
- pip *
- pot *
- protobuf ==3.20.2
- scikit-learn *
- seaborn *
- tensorflow ==1.13.1
- torch ==1.8.0
- torchmetrics *
- torchvision *
- tqdm *
- 121 dependencies