-
-
Notifications
You must be signed in to change notification settings - Fork 25.8k
Changed number of features and max iteration limits in plot_tweedie_regression_insurance_claims.py #21622
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
@sveneschlbeck please remember to paste before and after outputs and times with your PRs on examples :) @rth @lorentzenchr you may have a good idea on the details of this example |
@adrinjalali Yes, time tests are currently running, will paste them in soon :) |
Here what would be necessary as well is to check the figures / scores in the rendered example before and after this PR (see link in "ci/circleci: doc artifact") and make sure there are no major changes, or at least that the comments in the example are still valid with this smaller number of samples. |
@adrinjalali @rth Agreed, I could need your expertise to judge the effect of the changes, I am quite disappointed since the sample number reduction did not really save time. However, the results look quite a bit different. Maybe there is another way to achieve faster execution in this example? |
@sveneschlbeck Could you check which part of the example code takes the most time? Then we can focus on that part. Maybe it's the |
@lorentzenchr Will do :) |
@lorentzenchr I checked it out and these are the results analysing the partial execution times: line 1 - line 207 --> 0.001 sec So, this part seems to be the slowest by far:
|
@lorentzenchr What about reducing the number of samples? |
If the bottleneck is
then it depends on the ratio between download time and parsing time. If it's download time, reducing |
@lorentzenchr |
@sveneschlbeck do we have any conclusions here? |
Since the bottleneck is |
This will be fixed by #21938. Closing. Thanks for the work @sveneschlbeck |
#21598 @sply88 @adrinjalali @cakiki
Adapted the number of features at the beginning and then reduced the number of max possible iterations