-
-
Notifications
You must be signed in to change notification settings - Fork 25.9k
Adapted the number of splits in shuffle split to increase speed in plot_learning_curve.py #21628
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, could you please just propagate the ylim
param to the last plot to make results comparable across models?
Please also pass which can be very confusing. I will open a dedicated issue. Edit: actually the |
we could exclude the very fast runs from the example maybe? |
Added ``random_state=0`` to call of ``learning_curve``
@ogrisel Thanks for the input :) I added the |
As I explained in my edited comment, the
I want the y axis (score values) of both models to be on the same scale on the last row as they are on the first row. |
Maybe. Or we can keep the dataset large enough for the fastest model (NB) and subsample it only for the slowest model (SVC). Or we can just live with it. |
Added ``shuffle=True`` for ``random_state`` to make an impact
Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The new plot doesn't look more odd than what we have. Merging this one, we can merge this one and leave the fix for a separate PR.
@adrinjalali Alright |
* Adapted the number of splits * Update plot_learning_curve.py * Update plot_learning_curve.py Added ``random_state=0`` to call of ``learning_curve`` * Update plot_learning_curve.py Added ``shuffle=True`` for ``random_state`` to make an impact * Update examples/model_selection/plot_learning_curve.py Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org> Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org>
* Adapted the number of splits * Update plot_learning_curve.py * Update plot_learning_curve.py Added ``random_state=0`` to call of ``learning_curve`` * Update plot_learning_curve.py Added ``shuffle=True`` for ``random_state`` to make an impact * Update examples/model_selection/plot_learning_curve.py Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org> Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org>
* Adapted the number of splits * Update plot_learning_curve.py * Update plot_learning_curve.py Added ``random_state=0`` to call of ``learning_curve`` * Update plot_learning_curve.py Added ``shuffle=True`` for ``random_state`` to make an impact * Update examples/model_selection/plot_learning_curve.py Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org> Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org>
* Adapted the number of splits * Update plot_learning_curve.py * Update plot_learning_curve.py Added ``random_state=0`` to call of ``learning_curve`` * Update plot_learning_curve.py Added ``shuffle=True`` for ``random_state`` to make an impact * Update examples/model_selection/plot_learning_curve.py Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org> Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org>
* Adapted the number of splits * Update plot_learning_curve.py * Update plot_learning_curve.py Added ``random_state=0`` to call of ``learning_curve`` * Update plot_learning_curve.py Added ``shuffle=True`` for ``random_state`` to make an impact * Update examples/model_selection/plot_learning_curve.py Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org> Co-authored-by: Olivier Grisel <olivier.grisel@ensta.org>
#21598 @adrinjalali @sply88
Cut the number of splits in half twice. Didn't really change the outcome but time dropped significantly.
