Random Forest - LIME
Last updated
Last updated
Random Forest is an ensemble learning method used for classification and regression
Random Forests are a modification of bagging technique that builds a large collection of non-correlated trees and then averages them
This technique is sometimes even called "feature bagging".
Dataset: Automotive marketing ; Target: Customer lifetime value
The data is passed through pre-processing and then trained using Scikit learn Library with RandomForestRegressor function.
R2 score: 0.70
For this model we will be using LIME for interpretation. In particular, we use the LimeTabularExplainer which is part of the open source lime package.
We feed some data as shown above and then request the explainer to generate an explanation for a particular row. In the code, we have generated an explanation for 4th row and kept a cap on the number of features as 6.
Row Prediction
Positive/Negative Effects
The above plot gives us an idea of the features that impact the prediction positively and negatively. By positive impact, we say that the feature brought the prediction closer to the predicted value. Negative effect implies the feature took the prediction away from the predicted value. The length of the bars also show the impact. Longer the bar, higher the impact.
Individual Feature Values
The above plot shows the feature values for that particular row.