Skip to content

Initiating a paradigm shift in reporting and helping with making ML advances more considerate of sustainability and trustworthiness.

Notifications You must be signed in to change notification settings

raphischer/strep

Repository files navigation

STREP - Sustainable and Trustworthy Reporting for ML and AI

Software repository for more sustainable and trustworthy reporting in machine learning and artificial intelligence, as proposed in my PhD thesis and original research paper. With the publicly available Exploration tool, you can investigate all results - no code needs to run on your machine!

Framework Overview

Note that this software is under active development - it relfects work in progress and is subject to change, so you might encounter delays, off-times, and slight differences to earlier publications. Check out the paper branch and dissertation branch for frozen repository states at the respective time of publication.

News and Release History

  • 22/09/2025 - Published my PhD thesis based on STREP
  • 17/04/2025 - Some changes and lots of new figures, soon to be found in my PhD thesis
  • 13/01/2025 - Many fixes, updated Papers with Code and EdgeAccUSB databases
  • 02/10/2024 - Improved scaling methodology (x15 speed), updated MetaQuRe and AutoXPCR databases
  • 11/09/2024 - Improved functionality and presented the work at ECML-PKDD '24
  • 30/04/2024 - paper published in Data Mining and Knowledge Discovery, alongside initial repository version

Explore your own databases

Instead of exploring the pre-assembled databases, you can also investigate your own custom results by following these steps:

  1. Prepare your database as a pandas DataFrame (each row lists one model performance result on some data set, with different measures as columns).
  2. Store it in a directory, optionally add some JSON meta information (check our databases folder for examples and follow these naming conventions).
  3. Clone the repo and install necessary libraries via pip install -r requirements.txt (tested on Python 3.10).
  4. Either run python main.py --custom path/to/database.pkl, or use the following code snippet:
from strep.index_scale import load_database, scale_and_rate
from strep.elex.app import Visualization

fname = 'path/to/your/database.pkl'
# load database and meta information (if available)
database, meta = load_database(fname)
# index-scale and rate database
rated_database = scale_and_rate(database, meta)
# start the interactive exploration tool
app = Visualization(rated_database)
app.run_server()

Contributing

I firmaly believe that sustainable and trustworthy reporting is a community effort. If you perform large-scale benchmark experiments, stress-test models, or have any other important evaluations to report - get in touch! I would love to showcase other resource-aware evaluation databases and highlight your work.

Current available databases:

Citing

If you appreciate our work and code, please cite my PhD thesis and original research paper:

Fischer, R. Advancing the Sustainability of Machine Learning and Artificial Intelligence via Labeling and Meta-Learning.
Ph.D. Dissertation, TU Dortmund University (2025). https://doi.org/10.17877/DE290R-25716

Fischer, R., Liebig, T. & Morik, K. Towards More Sustainable and Trustworthy Reporting in Machine Learning. Data Mining and Knowledge Discovery 38, 1909–1928 (2024). https://doi.org/10.1007/s10618-024-01020-3

You can also use the these bibtext entries:

@phdthesis{fischer_diss,
  title={Advancing the Sustainability of Machine Learning and Artificial Intelligence via Labeling and Meta-Learning},
  author={Fischer, Raphael},
  school={TU Dortmund University},
  url={http://doi.org/10.17877/DE290R-25716},
  doi={10.17877/DE290R-25716},
  year={2025}
}
@article{fischer_dami,
	title = {Towards More Sustainable and Trustworthy Reporting in Machine Learning},
	volume = {38},
	issn = {1573-756X},
	url = {https://doi.org/10.1007/s10618-024-01020-3},
	doi = {10.1007/s10618-024-01020-3},
	number = {4},
	journal = {Data Mining and Knowledge Discovery},
	author = {Fischer, Raphael and Liebig, Thomas and Morik, Katharina},
	year = {2024},
	pages = {1909--1928},
}

Repository Structure

  • databases contains different gathered evaluation databases of ML reports, including scripts to assemble some of them.
  • strep contains software that processes the databases, calculates index values and compound scores, and visualizes them.
  • materials contains some additional data, scripts, and figures used in papers and my thesis.
  • The top level scripts are used to deploy the exploration tool on render, and a main script for running it locally.

Terms of Use

Copyright (c) 2025 Raphael Fischer

About

Initiating a paradigm shift in reporting and helping with making ML advances more considerate of sustainability and trustworthiness.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published