That level of reproducibility (byte-to-byte) is extremely difficult with CUDA, even with the identical random number generators. Different generations of GPUs will give you different results due to the hardware float point implementation and the instruction sets. But the model performance should be close to the reference baseline even with different GPU models.
The whole machine learning academia appreciates open source including fully open training code for reproducing the original work.
Note, you are moving the subject to “non-AI” program. We are talking about OSAID, let’s focus on AI programs.
I don’t want to ask people to include training data to be included in the media for blender distribution. The dataset should be separately accessible. I don’t think I specified that the training dataset has to be distributed together with the software part. Does “separately” resolve your concern about the blender example?