Conversation
Added a new Retired Division section to outline rules for submitting benchmarks that are no longer active, including submission guidelines and limitations.
Clarified that retired division submissions cannot be compared in any public communications to ensure fairness.
Updated the submission timeline for retired benchmarks from two years to three years.
|
MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅ |
| * Since reviewing retired benchmarks is an extra overhead, all mlperf logging checkers from the <round-of-retirement>* should be used to verify the retired benchmark results. If the <round-of-retirement> logging checkers fail for a retired division submission, the submission will be rejected. | ||
| * No new RCPs can be added for retired benchmarks, no changes to the reference code and no modifications will be allowed to the logging checkers. | ||
| * Retired division submissions CANNOT be compared to any other official MLPerf Training results from the current round or any previous rounds in any public communications (including but not limited to press statements, analyst discussions, website, blogs) to ensure fairness. | ||
| * The retired division submissions will not be displayed in the final results table to avoid confusion around result interpretability. It will get a submitter ID and it should be submitted under the division retired_<round-of-retirement>*. For example, if submitting Stable Diffusion v2 (v5.0 was the last round when the benchmark was in the suite) in round v6.0, the division for the submission should be "retired_v5.0". |
There was a problem hiding this comment.
Maybe we put the result with invalid results on the website? So folks can find it if needed in a table?
There was a problem hiding this comment.
Retired results shouldn't be invalid because they pass a submission checker and invalid is used for other means. I agree they should be visible on the website for easy reference, and if classified as a separate division this should be reasonable to keep separate from closed and open.
There was a problem hiding this comment.
If not visible on the website then they can only be found by running an offline submission check to print the result. Is this the correct assumption?
There was a problem hiding this comment.
Yes
the logs will be available so score can be generated
|
|
||
| The retired division is governed by all existing closed division rules, with the sole modification being the retirement of the benchmark. These additional rules also apply | ||
|
|
||
| * Since reviewing retired benchmarks is an extra overhead, all mlperf logging checkers from the <round-of-retirement>* should be used to verify the retired benchmark results. If the <round-of-retirement> logging checkers fail for a retired division submission, the submission will be rejected. |
Addresses mlcommons/training#847