You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @seralouk, in my experience, the more iterations of permutation importance, the more reliable the results. Permutation importance essentially returns the decrease in model accuracy due to randomly shuffling the values of a feature (i.e. column of your data matrix). So the more random seeds you average over, the more robust the resulting estimate of feature importance. The number of iterations you need depends on your application: if you know how precise you want your estimates of feature importance to be, you could try plotting the variance of these estimates versus the number of iterations, and use that to choose n_iter.
Good idea to plot the variance as a function of iterations.
I was hoping that there would be a rule of thumb that connects the number of iterations with the actual number of samples that are available in a study.
Hi all,
For the Permutation feature importance procedure, the default iteration value
n_iter
is 5 .See: https://eli5.readthedocs.io/en/latest/autodocs/sklearn.html#eli5.sklearn.permutation_importance.PermutationImportance
I am looking for a reference or publication that justifies the selection of any
n_iter
value.What is the gold standard or most commonly used
n_iter
value?The text was updated successfully, but these errors were encountered: