thaljoha wrote:I am performing a multi-state model in MARK of which I have 18 years of round-up data from bison. Each year, if the individual was on the island, they can be one of three states (calf, pregnant, or not pregnant). Sometimes, however, there is missing data so I have a dot in that place. Also, years before and after the individual was in the herd, I have zeros as place holders. So, my file without covariates looks like this:
/* AED8979*/ 0000000000N0000000 1;
/* AED8990*/ 00C000000000000000 1;
/* AED8998*/ 0000N0000000000000 1;
/* AEF7403*/ 000000N00000000000 1;
/* AEF7410*/ 00000000N000000000 1;
/* AEF7420*/ 00000NN00000000000 1;
/* AEF7421*/ 00000C000000000000 1;
/* AEF7455*/ 00000P000000000000 1;
/* AEF7459*/ 00000N..P000000000 1;
/* AEF7483*/ 000000000N00000000 1;
/* AEF7493*/ 00000000000P000000 1;
/* AEF7500*/ 000000N00000000000 1;
/* AEG6907*/ 0000000000N0000000 1;
/* AEG6910*/ 0000CNPP.PPPPPPN00 1;
/* AEG6911*/ 00000000N000000000 1;
/* AEG6923*/ 0000C0000000000000 1;
/* AEG6925*/ 00000N000000000000 1;
/* AEG6952*/ 0000C.NPNPNNP00000 1;
I want to estimate c-hat on my basic model s(.)p(.)psi(g) Apparently, however, I am not able to estimate c-hat using bootstrapping or median c-hat because of my missing data points (the dots), and I have tried to use U-CARE, however it will not run on my computer, it just crashes as soon as I open the .INP file. Any help would be greatly appreciated. Also, do I need to run estimate c-hat when I have the individual covariates in the model?
Thanks!
First, I'm puzzled how you have missing data, since you have per-individual histories? - the whole notion of a 'dot' makes sense in (say) occupancy, where you don't visit a site on a particular occasion (the 'dot' notation in the history was initially implemented for occupancy models). Is it really missing data, or is it simply that on a particular occasion, when the particular animal is encountered, that you don't know the state of the animal? If the latter, then you need another approach entirely (handling uncertain states in a MS context is a 'growth industry' in this field -- you'll probably end up using a HMM, or twisting the RD with hidden states data type into working for you). Failing that, if you really have missing values, so long as you don't have a lot of them (say, <5% of your data), then a quick-and-dirty approach is to (i) drop those individuals with 'dots' from the file, (ii) fit your general model, and (iii) run the GOF on that. I've found that if there are relatively few 'problem histories' that this works well enough to pass the sniff test.
Second, U-CARE requires numeric states (or used to, I presume it still does), not 'letters'. So, try re-coding your input file -- instead of N,C, P, say, 1,2,. Global search and replace takes care of this. This is actually noted on p. 47 of Chapter 10:
.... The only thing you need to do is make sure that all of your ‘state coding’ is numeric (e.g., if you use ‘B’ and ‘N’ in your input file, for example, you’ll need to change them to numbers, say 1 for ‘N’, and 2 for ‘B’...U-CARE cannot currently handle letters for state coding in the input file).
However, having said that, I rather doubt U-CARE can handle 'dots' either.
And, as for individual covariates, as noted in Chapter 11,
...the recommended approach is to perform GOF testing on the most general model that does not include the individual covariates, and use the c-hat value for this general model on all of the other models, even those including individual covariates. If individual covariates will serve to reduce (or at least explain) some of the variation, then this would imply that the c-hat from the general model without the covariates is likely to be too high, and thus, the analysis using this c-hat will be ’somewhat
conservative’. So, keep this in mind...
To get there from here, you need to make a copy of your input file without the covariates, fit your general model to it, and then look at GOF for this model. Take the c-hat from that analysis, and use it for your analysis of the 'full data' (i.e., the .inp file containing the individual covariates).