(I’d very much appreciate feedback or just plain thoughts on my weird case; thanks in advance!)
I wonder if I’m doing something wrong, not code-wise but let’s say machine-wise…
I am running an XGBClassifier on various combinations of a couple dozens different features. I get various reasonably good classification outcomes.
Then, as a sanity check, I run on one single feature, which is furthermore expected to not be crucially relevant to the task. Amazingly, I still get a classifier output that looks meaningful.
Is there any chance that the forest picks up something from previous runs?
I run on Jupyter and restart the kernel between runs. I’d like to hear from someone more knowledgable if this is adequate.
Is there any chance that the forest behaves erratically when it is given only one feature?
Other suggestions?