I’m trying to figure out whether the default behavior for
Booster.predict in the Python API is to utilize all available threads. I came across a number of posts on the default threading behavior specifically in the context of model training, but couldn’t find the same information for model prediction.
the model would by default use all the available threads (as if nthread=0), provided that
OMP_NUM_THREADSenvironment variable wasn’t set.
I tried digging into the C++ source code and even attempted to look into OpenMP but haven’t had much luck, as I have limited experience in either.
Thanks in advance!