You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am using H2O version 3.46.0.6 in Python. According to the H2O reproducibility page (https://docs.h2o.ai/h2o/latest-stable/h2o-docs/data-science/gbm-faq/reproducibility.html) the parallelization level (number of cores, nthreads) is supposed to control how the dataset is partitioned in memory (into "chunks"). However, I noticed that regardless of the number of threads I specify when initializing the cluster, the number of chunks remains the same.
This behavior seems inconsistent with the documentation. I am aware that the number of chunks can affect reproducibility. Does this mean that even if I explicitly control nthreads(), reproducibility is not guaranteed, as different machines with varying numbers of cores may produce different results?
The text was updated successfully, but these errors were encountered:
I am using H2O version 3.46.0.6 in Python. According to the H2O reproducibility page (https://docs.h2o.ai/h2o/latest-stable/h2o-docs/data-science/gbm-faq/reproducibility.html) the parallelization level (number of cores, nthreads) is supposed to control how the dataset is partitioned in memory (into "chunks"). However, I noticed that regardless of the number of threads I specify when initializing the cluster, the number of chunks remains the same.
ParseSetup heuristic: cloudSize: 1, cores: 28, numCols: 10, maxLineLength: 42, totalSize: 8758450, localParseSize: 8758450, chunkSize: 78201, numChunks: 111, numChunks * cols: 1110
This behavior seems inconsistent with the documentation. I am aware that the number of chunks can affect reproducibility. Does this mean that even if I explicitly control nthreads(), reproducibility is not guaranteed, as different machines with varying numbers of cores may produce different results?
The text was updated successfully, but these errors were encountered: