-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feat: add hyperparameters tuning #361
base: master
Are you sure you want to change the base?
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is great, thank you @LucasDedieu ! I've left a few comments throughout your code
hidden_dropout_prob: 0.1 | ||
attention_probs_dropout_prob: 0.1 | ||
classifier_dropout: 0.1 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Are these parameters passed to the underlying transformer object ? If so, can you add a comment to explain this, since these aren't documented in the eds.transformer
object
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, they are passed to the underlying transformer.
Their meanings is specified at beginning of section 2.3. Do you think I should add additional details ?
Coverage Report
Files without new missing coverage
273 files skipped due to complete coverage. Coverage failure: total of 97.93% is less than 98.06% ❌ |
… dummy trial to compute time, now using the first tuning trial
…ost. Now at the end of the study, we check if there is gpu time left and add trials if possible regarding EMA of past trials.
Quality Gate passedIssues Measures |
Description
The goal of this PR is to introduce a hyperparameter tuning script to EDS-NLP. This new feature enables users to optimize their model's hyperparameters by specifying either the available GPU hours or the desired number of trials. By doing so, users can efficiently find the optimal hyperparameters for training their models, leading to improved performance and efficiency.
Changes
edsnlp/tune.py
: Implemented the tuning functionality.tests/tuning/
: Added unit teststest_tuning.py
andtest_update_config.py
for the tuning functionality.docs/tutorials/tuning.md
: Created a new tutorial for hyperparameter tuning.docs/tutorials/index.md
: Added a link to the new tuning tutorial.mkdocs.yml
: Updated the navigation to include the new tuning tutorial.pyproject.toml
: Updated dependencies to include Optuna.Checklist