10000 [MRG] Model selection documentation by raghavrv · Pull Request #4 · raghavrv/scikit-learn · GitHub
[go: up one dir, main page]

Skip to content

[MRG] Model selection documentation #4

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
want 8000 s to merge 5 commits into from

Conversation

raghavrv
Copy link
Owner

Documentation for the scikit-learn#4294

The build docs are over here

TODO -

  • Move grid_search.rst to search.rst

@@ -143,43 +143,58 @@ Classes
covariance.oas
covariance.graph_lasso

.. _model_selection_ref:

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As I said, I think I would keep the modules in the references under a "old and will be removed" header. I'm not sure what @vene @GaelVaroquaux or @jnothman think about that, though.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is that model_selection_ref tag used anywhere?

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ok so all the "ref" tags were added in 6709ed5, the usage later got removed, and we kept adding the tags in good old cargo-cult fashion?

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm +0.5 about keeping the "old and will be removed". It might hurt googlability so we need good links to the new non-deprecated files.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

also ping @larsmans ;)

Copy link
Owner Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ok so all the "ref" tags were added in 6709ed5, the usage later got removed, and we kept adding the tags in good old cargo-cult fashion?

Ah :P

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm -1 for keeping the old modules listed here, but don't object to having their docs generated (by putting automodule somewhere hidden, say "deprecated.rst") with a clear deprecation-oriented docstring, if that's what you're more-or-less going for, @amueller. I'm not sure whether there's an easy way to set rel="canonical" to point to new equivalents, if that's your concern @vene.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm also convinced now that we should not list them here. Maybe just not list them. People can always look at the docs in ipython or the source if they like.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

+.. _model_selection_ref:

As I said, I think I would ke 8000 ep the modules in the references under a
"old and will be removed" header. I'm not sure what @vladn
@GaelVaroquaux or @jnothman think about that, though.

Sorry for the slow reply. I think that it is a good suggestion.

@amueller
Copy link

the developers/contributing.rst file still has references to the old modules, I found cross_validation there and grid_search.

@amueller
Copy link

Not sure what to do with all the mentions in whatsnew.....

@@ -618,24 +610,6 @@ From text
lda.LDA


.. _learning_curve_ref:

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm confused that these tags were not used anywhere....

@raghavrv
Copy link
Owner Author

Not sure what to do with all the mentions in whatsnew.....

I was also wondering about that :P but I thought maybe its better to leave it unchanged? since we'll be adding a whatsnew entry to note that they are grouped into model_selection...

@raghavrv
Copy link
Owner Author

the developers/contributing.rst file still has references to the old modules, I found cross_validation there and grid_search.

Oh are there? I'll fix them right away!

EDIT: fixed!

model_selection.cross_val_score
model_selection.cross_val_predict
model_selection.permutation_test_score
model_selection.check_cv

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maybe a stupid question, but why is check_cv public? Input from @vene @GaelVaroquaux @larsmans @jnothman welcome.

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Keep it public.

>>> slo = LabelShuffleSplit(labels, n_iter=4, test_size=0.5,
... random_state=0)
>>> for train, test in slo:
>>> slo = LabelShuffleSplit(n_iter=4, test_size=0.5, random_state=0)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

slo? maybe lss

@amueller
Copy link

LGTM apart from minor nitpicks. If we rename n_labels in LeavePLabelsOut that needs to be fixed in the doctests.

Traceback (most recent call last):
ValueError: 'wrong_choice' is not a valid scoring value. Valid options are ['accuracy', 'adjusted_rand_score', 'average_precision', 'f1', 'f1_macro', 'f1_micro', 'f1_samples', 'f1_weighted', 'log_loss', 'mean_absolute_error', 'mean_squared_error', 'median_absolute_error', 'precision', 'precision_macro', 'precision_micro', 'precision_samples', 'precision_weighted', 'r2', 'recall', 'recall_macro', 'recall_micro', 'recall_samples', 'recall_weighted', 'roc_auc']
>>> clf = svm.SVC(probability=True, random_state=0)
>>> cross_validation.cross_val_score(clf, X, y, scoring='log_loss') # doctest: +ELLIPSIS
>>> cross_val_score(clf, X, y, scoring='log_loss') # doctest: +ELLIPSIS
array([-0.07..., -0.16..., -0.06...])
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please revert the order of the 2 examples: start with the valid use case (scoring='log_loss') first before the failure case (scoring='wrong_choice').

@raghavrv raghavrv force-pushed the model_selection branch 2 times, most recently from 64eebe2 to 1c47e5d Compare October 22, 2015 16:01
@raghavrv raghavrv force-pushed the model_selection_documentation branch 2 times, most recently from 5a6ce7e to d147a85 Compare October 22, 2015 16:51
@raghavrv raghavrv force-pushed the model_selection_documentation branch from d147a85 to 4dcfa19 Compare October 23, 2015 12:47
@raghavrv raghavrv force-pushed the model_selection_documentation branch from 4dcfa19 to 602fc8e Compare October 23, 2015 12:58
@raghavrv raghavrv force-pushed the model_selection_documentation branch from 602fc8e to efff882 Compare October 23, 2015 13:02
dohmatob and others added 5 commits October 23, 2015 17:28
--------------------

* ENH Reogranize classes/fn from grid_search into search.py
* ENH Reogranize classes/fn from cross_validation into split.py
* ENH Reogranize cls/fn from cross_validation/learning_curve into validate.py

* MAINT Merge _check_cv into check_cv inside the model_selection module
* MAINT Update all the imports to point to the model_selection module
* FIX use iter_cv to iterate throught the new style/old style cv objs
* TST Add tests for the new model_selection members
* ENH Wrap the old-style cv obj/iterables instead of using iter_cv

* ENH Use scipy's binomial coefficient function comb for calucation of nCk
* ENH Few enhancements to the split module
* ENH Improve check_cv input validation and docstring
* MAINT _get_test_folds(X, y, labels) --> _get_test_folds(labels)
* TST if 1d arrays for X introduce any errors
* ENH use 1d X arrays for all tests;
* ENH X_10 --> X (global var)

Minor
-----

* ENH _PartitionIterator --> _BaseCrossValidator;
* ENH CVIterator --> CVIterableWrapper
* TST Import the old SKF locally
* FIX/TST Clean up the split module's tests.
* DOC Improve documentation of the cv parameter
* COSMIT consistently hyphenate cross-validation/cross-validator
* TST Calculate n_samples from X
* COSMIT Use separate lines for each import.
* COSMIT cross_validation_generator --> cross_validator

Commits merged manually
-----------------------

* FIX Document the random_state attribute in RandomSearchCV
* MAINT Use check_cv instead of _check_cv
* ENH refactor OVO decision function, use it in SVC for sklearn-like
  decision_function shape
* FIX avoid memory cost when sampling from large parameter grids

ENH Major to Minor incremental enhancements to the model_selection

Squashed commit messages - (For reference)

Major
-----

* ENH p --> n_labels
* FIX *ShuffleSplit: all float/invalid type errors at init and int error at split
* FIX make PredefinedSplit accept test_folds in constructor; Cleanup docstrings
* ENH+TST KFold: make rng to be generated at every split call for reproducibility
* FIX/MAINT KFold: make shuffle a public attr
* FIX Make CVIterableWrapper private.
* FIX reuse len_cv instead of recalculating it
* FIX Prevent adding *SearchCV estimators from the old grid_search module
* re-FIX In all_estimators: the sorting to use only the 1st item (name)
    To avoid collision between the old and the new GridSearch classes.
* FIX test_validate.py: Use 2D X (1D X is being detected as a single sample)
* MAINT validate.py --> validation.py
* MAINT make the submodules private
* MAINT Support old cv/gs/lc until 0.19
* FIX/MAINT n_splits --> get_n_splits
* FIX/TST test_logistic.py/test_ovr_multinomial_iris:
    pass predefined folds as an iterable
* MAINT expose BaseCrossValidator
* Update the model_selection module with changes from master
  - From scikit-learn#5161
  -  - MAINT remove redundant p variable
  -  - Add check for sparse prediction in cross_val_predict
  - From scikit-learn#5201 - DOC improve random_state param doc
  - From scikit-learn#5190 - LabelKFold and test
  - From scikit-learn#4583 - LabelShuffleSplit and tests
  - From scikit-learn#5300 - shuffle the `labels` not the `indxs` in LabelKFold + tests
  - From scikit-learn#5378 - Make the GridSearchCV docs more accurate.
  - From scikit-learn#5458 - Remove shuffle from LabelKFold
  - From scikit-learn#5466(scikit-learn#4270) - Gaussian Process by Jan Metzen
  - From scikit-learn#4826 - Move custom error / warnings into sklearn.exception

Minor
-----

* ENH Make the KFold shuffling test stronger
* FIX/DOC Use the higher level model_selection module as ref
* DOC in check_cv "y : array-like, optional"
* DOC a supervised learning problem --> supervised learning problems
* DOC cross-validators --> cross-validation strategies
* DOC Correct Olivier Grisel's name ;)
* MINOR/FIX cv_indices --> kfold
* FIX/DOC Align the 'See also' section of the new KFold, LeaveOneOut
* TST/FIX imports on separate lines
* FIX use __class__ instead of classmethod
* TST/FIX import directly from model_selection
* COSMIT Relocate the random_state documentation
* COSMIT remove pass
* MAINT Remove deprecation warnings from old tests
* FIX correct import at test_split
* FIX/MAINT Move P_sparse, X, y defns to top; rm unused W_sparse, X_sparse
* FIX random state to avoid doctest failure
* TST n_splits and split wrapping of _CVIterableWrapper
* FIX/MAINT Use multilabel indicator matrix directly
* TST/DOC clarify why we conflate classes 0 and 1
* DOC add comment that this was taken from BaseEstimator
* FIX use of labels is not needed in stratified k fold
* Fix cross_validation reference
* Fix the labels param doc

FIX/DOC/MAINT Addressing the review comments by Arnaud and Andy

COSMIT Sort the members alphabetically
COSMIT len_cv --> n_splits
COSMIT Merge 2 if; FIX Use kwargs
DOC Add my name to the authors :D
DOC make labels parameter consistent
FIX Remove hack for boolean indices; + COSMIT idx --> indices; DOC Add Returns
COSMIT preds --> predictions
DOC Add Returns and neatly arrange X, y, labels
FIX idx(s)/ind(s)--> indice(s)
COSMIT Merge if and else to elif
COSMIT n --> n_samples
COSMIT Use bincount only once
COSMIT cls --> class_i / class_i (ith class indices) -->
perm_indices_class_i

FIX/ENH/TST Addressing the final reviews

COSMIT c --> count
FIX/TST make check_cv raise ValueError for string cv value
TST nested cv (gs inside cross_val_score) works for diff cvs
FIX/ENH Raise ValueError when labels is None for label based cvs;
TST if labels is being passed correctly to the cv and that the
ValueError is being propagated to the cross_val_score/predict and grid
search
FIX pass labels to cross_val_score
FIX use make_classification
DOC Add Returns; COSMIT Remove scaffolding
TST add a test to check the _build_repr helper
REVERT the old GS/RS should also be tested by the common tests.
ENH Add a tuple of all/label based CVS
FIX raise VE even at get_n_splits if labels is None
FIX Fabian's comments
PEP8
@raghavrv raghavrv force-pushed the model_selection_documentation branch from efff882 to cff6258 Compare October 23, 2015 15:34
@raghavrv raghavrv closed this Oct 23, 2015
@raghavrv raghavrv deleted the model_selection_documentation branch February 11, 2016 13:12
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

8 participants
0