-
Notifications
You must be signed in to change notification settings - Fork 183
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem with result caching if evaluation is interrupted #653
Comments
Indeed, we should save the results of the benchmark immediately after computation to avoid this issue. I tried implementing this, but something else was breaking.I could not find what. If you have time, you can take over this PR :) #422 |
MAybe a way to fix this with not too much code would be to rely on |
Indeed, replacing the for loops in the evaluations with joblib would be great! We started to discuss it here #481 |
I think it saves the results or it marks the first processed subjects as such, but it does not continue processing the rest of the subjects after the interruption. It decides that the dataset is already entirely processed. Is there some flag for that? There should be a check that counts how many subjects were cached and how many needs to be further processed. There can even be a user message: "3 subjects already cached, proceeding with N-3" where N is the total number of subjects for this dataset. |
@toncho11 do you have a minimal example to reproduce this behavior? |
It is a subtle problem that I am trying to explain here.
Same goes for an error like calculation or memory allocation. Thank you! |
I am running MOABB benchmark or just using the normal evaluation and the process is interrupted:
For example, if the interruption was at 3/64 subject then running again with overwrite=false will not process the other 64-3 subjects but it will simply use the 3 processed subjects in the final result as if the dataset was of 3 subjects (and not 64).
Actually the same happens when you use less subjects manually than the total number of subjects for the dataset.
I think this can lead to false results if you are not aware of this behavior.
The text was updated successfully, but these errors were encountered: