-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add: support for ignore_pretraining_limits to PHE #12
Conversation
Closes #11 |
Thanks Lennart!
WDYT? There may be some interaction with PHE I'm not aware of. I'm pretty confident we should add |
Mhm, it depends on how complicated you want to do this. Another approach would be to provide code instead to pass the models you want to use. PHE already supports this https://github.com/PriorLabs/tabpfn-extensions/blob/main/src/tabpfn_extensions/post_hoc_ensembles/pfn_phe.py#L133 So we can dettach model HPs from PHE HPs. But this would be a bit much for me IMO. |
Makes sense. Do you think |
I think What do you think about this line of reasoning? |
I thought it's two separate optimizations and the outer optimization relies on the inner to be aligned. Assume inner would have the opposite optimization direction of the outer. Now it would be very hard to optimize the outer ensemble. But maybe that's not the point you meant I realize. We should rather add a parameter that is passed to TabPFN to optimize any metric also internally? |
Hmm yeah I think that's fair. We might think about adding some more metrics then, for instance |
It is two separate optimizations, but for ensembling, you do not want to add such a bias to the base model, as it is more likely problematic for aggregating the predictions of base models. For example, if you can, you want to avoid temperature scaling base models of an ensemble but instead scale the ensemble's predictions, as done in AutoGluon. You could add a metric to the base model and check whether balancing predictions helps, but the overhead would be too large. This would require validation data, which is exactly what the PHE already does. Adding it to the search space is a good idea for sure! (One more small ensembling intuition: it is very easy for a weighted average to approximate balanced probabilities given a few different predictions from base models) |
See #11