Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Decreased predictiveness when converting integers to floats #10683

Open
tsmith-perchgroup opened this issue Aug 7, 2024 · 1 comment
Open

Comments

@tsmith-perchgroup
Copy link

tsmith-perchgroup commented Aug 7, 2024

I am running some experimentation on a dataset with roughly 300 features and around 300k datapoints. There are roughly 50 integer variables, representing a range of one-hot encoded, label encoded and ordinal numerical data.

When I convert all integer columns from integer to float before fitting my model, I see a significant reduction in model predictiveness on the test set.

Can anyone shed some light on why this might be? I can't find anything when performing a web search. I'm running XGBoost 2.0.3 using the sklearn API.

@trivialfis
Copy link
Member

Hi, could you please share a reproducible example? It's not easy to make guess based on your description.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants