-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the use of deep-forest #1
Comments
Hi, DeepForest was used as part of my final solution! Its just that it takes a bit long to run (~10mins per fold), so I just commented it out when running some experiments to prepare for my presentation. For your information, out of the 7 base models (GLM, XGB, Catboost, LGBM, DeepForest, NN, Baynesian_NN), Deepforest is slightly worse than the others 6. Interestingly, it does not have strong correlation (0.94) with the other GBMs (e.g. XGB), while among the GBMs they will have 0.97-0.98 in correlation. So, the final stack model performance is improved a bit after including DeepForest. Thanks! |
Glad to hear that @davidlkl 😄. You could think of Deep forest as a random forest based ensemble, and it is reasonable that deep forest has more different predictions compared to the predictions between GBDTs. Once again, congrats! Another question is that are you willing to share your winning solution in our community, the link to this repo will be posted in a webpage in our documentation. More details are available in this issue. |
No problem! Glad to have my solution shared! |
Hi @davidlkl, congrats on winning the competition, and thanks for trying out the deep forest package ;-)
It looks like that the
CascadeForestRegressor
was not used in your final solution, right? (since thedf_regressor
was commented out in train.py). Therefore, we will appreciate it very much if you could tell us the reason, such as adding it simply makes the performance worse, or it takes forever to finish training. Thanks!The text was updated successfully, but these errors were encountered: