by Mateo » Thu Jul 15, 2010 4:59 pm
25 presences out of how many surveys from how many sites?
I have 537 sites, surveyed 3 times. Out of ~1600 searches, I had 45 detections spread across 25 sites. Needle in a haystack!
Back to your original question though, a beta estimate of -37 is reasonable in this case and as for the SE for the beta, it should probably be a really large number rather than 0. Did you get any warnings? Essentially though you have a boundary problem in that for those sites your estimate of occupancy is 0. In these situations you can get unstable results for the standard errors.
I agree that a large SE makes more sense. I originally did this in RMark and got B=-37 and SE=E-9. The only warning from RMark was that the number of parameters was one too low, obviously due to this parameter. I just redid it in PRESENCE and got B=-44 and SE=E+10. I got the "numerical convergence may not have been reached" warning. Which is obviously true.
What I suggest you do is rerun that model but fix the beta estimate to something like -37 and see if that changes the SE's for the other parameters in the model.
Although I know how to fix a parameter, like p or psi, in MARK or PRESENCE, I don't know how to fix a beta estimate. I asked a couple people in the office, checked Google, and checked the forum, but maybe I overlooked something. I tried to fix psi, just to see, and it just fixes psi at all sites and does not estimate any covariates.
So I couldn't follow your advice. However, I think the real question was, "are the estimates unstable?", so I tried an alternate tact. I changed the covariate for one occupied site to remove the data separation. Although it is faking data, this is a crude solution that I have seen used. Then I reran the model. For the intercept, other covariates, and p, estimates changed by <7% and SEs changed by <2%. So I think the other estimates are pretty stable.
How does this model rank compared to other models that you've considered?
If I allow the separation and use B=-37, it gets 71% of the weight. If I subset the data to remove the variable, it is down to 0.3% weight. If I use the bias-reduced logistic regression, it is 9%, and if I use the fake data "fix" above, then it is 4%. So the choice of analysis makes a big difference.
Sorry for the long response, and thanks for taking an interest in my problem!!