Software repository for more sustainable and trustworthy reporting in machine learning and artificial intelligence, as proposed in my PhD thesis and original research paper. With the publicly available Exploration tool, you can investigate all results - no code needs to run on your machine!
Note that this software is under active development - it relfects work in progress and is subject to change, so you might encounter delays, off-times, and slight differences to earlier publications. Check out the paper branch and dissertation branch for frozen repository states at the respective time of publication.
- 22/09/2025 - Published my PhD thesis based on STREP
- 17/04/2025 - Some changes and lots of new figures, soon to be found in my PhD thesis
- 13/01/2025 - Many fixes, updated Papers with Code and EdgeAccUSB databases
- 02/10/2024 - Improved scaling methodology (x15 speed), updated MetaQuRe and AutoXPCR databases
- 11/09/2024 - Improved functionality and presented the work at ECML-PKDD '24
- 30/04/2024 - paper published in Data Mining and Knowledge Discovery, alongside initial repository version
Instead of exploring the pre-assembled databases, you can also investigate your own custom results by following these steps:
- Prepare your database as a
pandasDataFrame (each row lists one model performance result on some data set, with different measures as columns). - Store it in a directory, optionally add some
JSONmeta information (check our databases folder for examples and follow these naming conventions). - Clone the repo and install necessary libraries via
pip install -r requirements.txt(tested on Python 3.10). - Either run
python main.py --custom path/to/database.pkl, or use the following code snippet:
from strep.index_scale import load_database, scale_and_rate
from strep.elex.app import Visualization
fname = 'path/to/your/database.pkl'
# load database and meta information (if available)
database, meta = load_database(fname)
# index-scale and rate database
rated_database = scale_and_rate(database, meta)
# start the interactive exploration tool
app = Visualization(rated_database)
app.run_server()I firmaly believe that sustainable and trustworthy reporting is a community effort. If you perform large-scale benchmark experiments, stress-test models, or have any other important evaluations to report - get in touch! I would love to showcase other resource-aware evaluation databases and highlight your work.
- ImageNetEff22 (Fischer et al. 2022): Efficiency information of popular ImageNet models
- EdgeAccUSB (Staay et al. 2024): Efficiency results of stress-tested USB accelerators for edge inference with computer vision models
- XPCR / Forecasting (Fischer et al. 2024): Efficiency information of DNNs for time series forecasting tasks
- MetaQuRe (Fischer et al. 2024): Resource and quality information of ML algorithm performance on tabular data
- RobustBench (Croce et al. 2020): Robustness and quality information of image classification models
- Papers With Code: The most popular benchmarks from this public database (code for re-assembling can be found here)
If you appreciate our work and code, please cite my PhD thesis and original research paper:
Fischer, R. Advancing the Sustainability of Machine Learning and Artificial Intelligence via Labeling and Meta-Learning.
Ph.D. Dissertation, TU Dortmund University (2025). https://doi.org/10.17877/DE290R-25716
Fischer, R., Liebig, T. & Morik, K. Towards More Sustainable and Trustworthy Reporting in Machine Learning. Data Mining and Knowledge Discovery 38, 1909–1928 (2024). https://doi.org/10.1007/s10618-024-01020-3
You can also use the these bibtext entries:
@phdthesis{fischer_diss,
title={Advancing the Sustainability of Machine Learning and Artificial Intelligence via Labeling and Meta-Learning},
author={Fischer, Raphael},
school={TU Dortmund University},
url={http://doi.org/10.17877/DE290R-25716},
doi={10.17877/DE290R-25716},
year={2025}
}@article{fischer_dami,
title = {Towards More Sustainable and Trustworthy Reporting in Machine Learning},
volume = {38},
issn = {1573-756X},
url = {https://doi.org/10.1007/s10618-024-01020-3},
doi = {10.1007/s10618-024-01020-3},
number = {4},
journal = {Data Mining and Knowledge Discovery},
author = {Fischer, Raphael and Liebig, Thomas and Morik, Katharina},
year = {2024},
pages = {1909--1928},
}databasescontains different gathered evaluation databases of ML reports, including scripts to assemble some of them.strepcontains software that processes the databases, calculates index values and compound scores, and visualizes them.materialscontains some additional data, scripts, and figures used in papers and my thesis.- The top level scripts are used to deploy the exploration tool on render, and a main script for running it locally.
Copyright (c) 2025 Raphael Fischer
