Score: 0

Statistical Quality and Reproducibility of Pseudorandom Number Generators in Machine Learning technologies

Published: July 2, 2025 | arXiv ID: 2507.03007v1

By: Benjamin A. Antunes

Potential Business Impact:

Makes computer learning more fair and reliable.

Business Areas:
A/B Testing Data and Analytics

Machine learning (ML) frameworks rely heavily on pseudorandom number generators (PRNGs) for tasks such as data shuffling, weight initialization, dropout, and optimization. Yet, the statistical quality and reproducibility of these generators-particularly when integrated into frameworks like PyTorch, TensorFlow, and NumPy-are underexplored. In this paper, we compare the statistical quality of PRNGs used in ML frameworks (Mersenne Twister, PCG, and Philox) against their original C implementations. Using the rigorous TestU01 BigCrush test suite, we evaluate 896 independent random streams for each generator. Our findings challenge claims of statistical robustness, revealing that even generators labeled ''crush-resistant'' (e.g., PCG, Philox) may fail certain statistical tests. Surprisingly, we can observe some differences in failure profiles between the native and framework-integrated versions of the same algorithm, highlighting some implementation differences that may exist.

Page Count
12 pages

Category
Computer Science:
Other Computer Science