Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FIX Make cross-validation figures less inexact #765

Merged
merged 3 commits into from
May 17, 2024

Conversation

ArturoAmorQ
Copy link
Collaborator

@ArturoAmorQ ArturoAmorQ commented Feb 23, 2024

Some people has mentioned in RL that the figures on train-validation-test split and nested cross-validation may be misleading, as they can make the user think the testing samples are used for evaluation in each cv iteration.

Instead of remaking the figures, this PR proposes a more conservative (but still somewhat inexact) solution by changing the legend labels.

Edit: This PR now does change the figures to better show the re-train process done by default by the grid search.

cc @lesteve and @tuscland

@glemaitre
Copy link
Collaborator

Could we order the label into: 1. training, 2. validation (reserved for hyperparameters evaluation), 3. testing (reserved for final or ... evaluation).

@ArturoAmorQ
Copy link
Collaborator Author

Could we order the label into: 1. training, 2. validation (reserved for hyperparameters evaluation), 3. testing (reserved for final or ... evaluation).

Validation sets are not reserved, but actually used in the given cv iteration. In such case there is no big advantage on factorizing the legend labels as 1., 2. and 3.

@glemaitre glemaitre merged commit 9a9e74f into INRIA:main May 17, 2024
3 checks passed
github-actions bot pushed a commit that referenced this pull request May 17, 2024
@ArturoAmorQ ArturoAmorQ deleted the testing_samples branch May 17, 2024 09:17
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants