-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ENH add batching benchmarks #3
base: master
Are you sure you want to change the base?
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Some comments.
Yes it would be nice to also see the results with no batching.
# defines a variety of task duration profiles. | ||
random_state = np.random.RandomState(42) | ||
high_variance = np.abs( | ||
random_state.normal(loc=0.000001, scale=0.01, size=5000) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For float
like this, I'd rather use 1e-6
as it is easier to read in my opinion
params = ([10000, 100000, 1000000], [1, 2, 4]) | ||
parallel_parameters = dict( | ||
verbose=10, backend="loky", pre_dispatch="2*n_jobs" | ||
) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe you could also run the benchmark with auto-memmap disabled max_nbytes=None
?
Actually, we also need to wait before |
This PR adds a few benchmarks to assess the performance of joblib's batching strategy.