Search code examples
pythonmachine-learningcatboost

Negative Feature Importance Value in CatBoost LossFunctionChange


I am using CatBoost for ranking task. I am using QueryRMSE as my loss function. I notice for some features, the feature importance values are negative and I don't know how to interpret them.

It says in the documentation, the i-th feature importance is calculated as the difference between loss(model with i-th feature excluded) - loss(model).

So a negative feature importance value means that feature makes my loss go up? What does that suggest then?


Solution

  • Negative feature importance value means that feature makes the loss go up. This means that your model is not getting good use of this feature. This might mean that your model is underfit (not enough iteration and it has not used the feature enough) or that the feature is not good and you can try removing it to improve final quality.