-
Notifications
You must be signed in to change notification settings - Fork 51
Duplicated trials #207
Comments
How duplicate trials are handled depends on the search algorithm itself, and it looks like HEBO doesn't account for those. |
So, we can fix that by sending existing results to HEBO without additional training. |
Isn't it just due to Pandas trying to fit the dataframe on screen? |
Thank you, it works. Don't you think it would be good to skip all duplicates for all searchers by default? I think it's a big problem when you think you tune hp's but in fact you're training the same over and over again. |
It's not a straightforward thing, as those should be ideally handled by the search algorithm itself. For example, if we were to reject duplicates from a search algorithm that doesn't check for them, it is possible for the situation to become an infinite loop where the tuner rejects the duplicate suggestion only for the algorithm to suggest it again, as this is what it considers to be the best configuration. In any case, that should be done in Ray Tune itself, and not here. @krfricke what do you think? |
Hmm, so yeah this seems to be a common request. We've actually implemented something similar in Bayesopt (see ray/python/ray/tune/suggest/bayesopt.py). |
Besides HEBO counts duplicates as newly tested parameters and throws an error when it reaches all possible combinations count. This leads to the fact that not all parameters are tested. |
@richardliaw |
I run your HEBO custom example and see that it runs the same trials multiple times. Can I skip them and finish when there are no unique hp configurations left?
Setting
cv=5
I expected to see 5 test scores for each trial, but it shows only 3 of them:split0_test_score, split1_test_score, split2_test_score
. Can you clarify how it works?The text was updated successfully, but these errors were encountered: