relative variable importance with unbalanced model set

While developing a set of models, we would like to use this approach to show some weight of evidence for a particular variable. However, when all variables do not have equal representation throughout the model set you might run into some level of concern that there is some bias in the resulting accumulated weight score, just due to frequency.
I know there is some philosophy behind the fact that this is ok, and it's really just a function of your predefined a priori hypotheses, and you just have to recognize that and present/infer your results within that universe.
That said, I think there is some value for obtaining a estimate of variable importance where this is corrected. Examples might include interaction terms which require both variables to also be present in the model, or polynomial terms. We (Julie Yee and I) decided on: log(W/(1-W)) - log(Nv/(N-Nv)); were W is sum of model weights that a given variable is in, Nv is the number of models that the given variable is in and N is total number of models.
Searching the forum prior to posting I noticed Darryl (Mackenzie) suggested something similar here:
[url]http://www.phidot.org/forum/viewtopic.php?f=34&t=1228&p=3423&hilit=relative+variable+importance#p3423
[/url]
Darryl's is basically an odds-ratio, and we opted to use a log-odds to provide a bit more symmetric range. But logic is the same.
Question/Discussion topic here is two-fold:
1. How do folks feel about the use of the log-odds or odds ratio as an approach to correct for this?
2. Is there anything out there that is citeable to use such an approach? We don't want to have to spend too much time each paper trying to defend this approach... and perhaps we can do that once and then cite that paper each time going forward, but would definitely prefer to have something more specific to cite if it's available.
Thanks,
Mark
I know there is some philosophy behind the fact that this is ok, and it's really just a function of your predefined a priori hypotheses, and you just have to recognize that and present/infer your results within that universe.
That said, I think there is some value for obtaining a estimate of variable importance where this is corrected. Examples might include interaction terms which require both variables to also be present in the model, or polynomial terms. We (Julie Yee and I) decided on: log(W/(1-W)) - log(Nv/(N-Nv)); were W is sum of model weights that a given variable is in, Nv is the number of models that the given variable is in and N is total number of models.
Searching the forum prior to posting I noticed Darryl (Mackenzie) suggested something similar here:
[url]http://www.phidot.org/forum/viewtopic.php?f=34&t=1228&p=3423&hilit=relative+variable+importance#p3423
[/url]
Darryl's is basically an odds-ratio, and we opted to use a log-odds to provide a bit more symmetric range. But logic is the same.
Question/Discussion topic here is two-fold:
1. How do folks feel about the use of the log-odds or odds ratio as an approach to correct for this?
2. Is there anything out there that is citeable to use such an approach? We don't want to have to spend too much time each paper trying to defend this approach... and perhaps we can do that once and then cite that paper each time going forward, but would definitely prefer to have something more specific to cite if it's available.
Thanks,
Mark