Median C-hat Output

questions concerning analysis/theory using program MARK

Median C-hat Output

Postby npseudacris » Tue Apr 11, 2006 3:48 pm

Greetings:

The median c-hat output that is displayed in notepad says that the analysis was preformed as a known fate model which I am assuming is the default/only way since that is how the help section explains that the logistic regression analysis (for median c-hat) is preformed in MARK and it doesn’t matter that my model is a Live Recapture model, am I correct?

The output also says that I have one covariate. I did not run my median c-hat with my individual covariates (they are in the data read into MARK, but not in the design matrix for the general model I used for the median c-hat). Is this some default setting or should I be concerned that there is a problem with my data?

Finally, in setting the upper bound for the estimation, the example on page 5-27 of the book uses 5.5 which is slightly higher than the observed deviance c-hat. Should you always set the upper bound slightly higher than the observed deviance c-hat? I also checked the MARK help on this and it said “….to find the approximate range in which to simulate c to focus the simulated data around the likely value of c that will result.” Not the clarity I was hoping for, any further guidelines for setting the upper bound would be appreciated.

Thanks in advance
npseudacris
 
Posts: 8
Joined: Tue Mar 28, 2006 1:27 pm
Location: Murray State University, KY

Median C-hat Output

Postby gwhite » Tue Apr 11, 2006 4:02 pm

Nicole:
You are correct that the known fate model output is just the logistic regression to compute the median c-hat value. The reason there is a covariate is that the independent variable for the logistic regression is the true c value used to generate the data that are analyzed in the logistic regression.

The upper bound should be set such that you bracket the value of c estimated from the logistic regression, i.e., data are generated with c values lower than the estimated value AND data are generated with c values greater than the estimated c-hat. If you don't bracket the estimate, then you will not get a correct result.

Gary
gwhite
 
Posts: 340
Joined: Fri May 16, 2003 9:05 am

Re: Median C-hat Output

Postby cooch » Wed Apr 12, 2006 11:50 am

Adding a few things to Gary's response:

npseudacris wrote:The median c-hat output that is displayed in notepad says that the analysis was preformed as a known fate model which I am assuming is the default/only way since that is how the help section explains that the logistic regression analysis (for median c-hat) is preformed in MARK and it doesn’t matter that my model is a Live Recapture model, am I correct?


Correct - as noted in the most current version of chapter 5, MARK uses the known-fate analysis to handle logistic regression, which is used to derive an estimate of median c-hat. See the -sidebar- on p. 27 of Chapter 5.

Finally, in setting the upper bound for the estimation, the example on page 5-27 of the book uses 5.5 which is slightly higher than the observed deviance c-hat. Should you always set the upper bound slightly higher than the observed deviance c-hat? I also checked the MARK help on this and it said “….to find the approximate range in which to simulate c to focus the simulated data around the likely value of c that will result.” Not the clarity I was hoping for, any further guidelines for setting the upper bound would be appreciated.

You seem to have an older version of the chapter. But, nonetheless, the strategy I generally recommend is (i) run the design points from 1 -> some point greater than the observed value (i.e., bracket it, as Gary describes). Use 5 or 6 design points, with say 3-4 replicates per design point - just enough to give you a 'quick and dirty' idea of where the median c-hat is. Say its ~1.65. Then (ii) re-run the analysis, using a bracketing around this first estimate (say, 1 -> 2), again with 5-6 design points, but many more replicates (say, 15-20). The motivation is based on the fact that running the median c-hat routine s compute intensive. Trying it in two stages - one quick and dirty, to find the general part of the curve you want to be in), and then a second, more intensive run in the range of this part of the curve, is often a time-saver in the end, especially if your data set is large...

In the book, I mention (in the particular example being discussed) actually using an upper bound lower than the observed 5.2 (I use 3.0). The reason (as noted) is that in that case, is that if your c-hat is >3, then you're probably going to have have greater problems anyway - the general recommendation is that c-hat<= 3.0 is a reasonable adjustment to make. Anything >3 pretty well guarantees your best models will typically be of very low complexity (often 'dot' models). So, I simply went 1 -> 3 (but, the dipper data set being used in that example is one we're all *rather* familiar with, so I knew I was pretty safe).
Last edited by cooch on Thu Apr 13, 2006 8:29 pm, edited 1 time in total.
cooch
 
Posts: 1654
Joined: Thu May 15, 2003 4:11 pm
Location: Cornell University

Postby npseudacris » Wed Apr 12, 2006 3:32 pm

Gary and Evan,

Thanks for the help that really cleared things up. I'm using the 4th edition of the book, I know another version was recently posted, so I'll switch to that. My median c-hat was 1.72, so I'm a happy camper for now. I'm starting my multi-state robust models next (different project), so I suspect I will be spending some quality time with the gentle guide, my notes from the workshop and searching phidot posts.

Nicole
npseudacris
 
Posts: 8
Joined: Tue Mar 28, 2006 1:27 pm
Location: Murray State University, KY


Return to analysis help

Who is online

Users browsing this forum: No registered users and 3 guests