ACE or ADE and results interpretation
I'm wondering that in which condition we should choose ACE model, or ADE? Is there any criteria to be referred to? For example, in one trait of my own data, the ACE model indicate the CE model is the best fitted model, while the ADE model indicate the DE model is the best fitted model, is this trait heritable? And in the ADE model, how to calculate the heritability? Only the part of A, or the part of A and D, or only the D?
About the result interpreting, is there any criteria help to decide which model is the best fitted model and if the trait is heritable? I found in some paper just using the minus2LL and AIC to choose the best fitted model and in others just comparing the AE model to the E model in which if the model was significantly worsen, then this trait is heritable. Actually I have not found a clear standard criteria to apply the model selection and result interpreting. Is there any recommended literature about it? Many thanks.
I recommend fitting all
Log in or register to post comments
In reply to I recommend fitting all by AdminRobK
Only the AIC
Refer to the ACE vs ADE, should I estimate the ACE, ADE and their submodels each in one study? Is that make sense? If the DE modle is the best fitted one, the heritability shold be the d2/(a2+d2+e2) or (a2+d2)/(a2+d2+e2), even the effect of additive gene is not significant? Many thanks.
Log in or register to post comments
In reply to Only the AIC by Liz
Model selection vs. statistical inference
Again, if you're selecting a model by AIC, forget about pairwise comparisons of nested models until after model selection is complete. The LRT only allows for pairwise comparison of a model and a nested submodel, with one of the two (somewhat) arbitrarily treated as the null model. In contrast, AIC allows you to compare ALL models under consideration to one another, and they need not be a sequence of nested submodels. To address your question more directly: if two models differ by a single parameter, and the LRT p-value is greater than 0.05 but less than about 0.16, then AIC will prefer the model that includes the parameter. The interpretation would be something like "the AIC-selected model included parameter [whatever], which was not significant at conventional significance levels." If instead the LRT p-value is greater than about 0.16, then AIC will prefer the model that drops the parameter.
It may well make sense, but that's partly a subject-matter question about the phenotype, based on existing literature and theory. Also, sometimes you can tell from the saturated-model correlations and covariances that certain models will obviously fit the data poorly. However, it's not advised to fit a model that includes D but excludes A. That's sort of like using a power function of a predictor in regression without also including the first-degree effect. At any rate, even if the effect of A is non-significant, I would use (a2+d2)/(a2+d2+e2) as the heritability estimate, as it provides a lower-bound estimate on the broad-sense heritability.
All of the foregoing assumes you are going to select only one model, and base your conclusions solely off of it. I advocate multimodel inference instead--see my relevant publication for more information.
Log in or register to post comments
In reply to Model selection vs. statistical inference by AdminRobK
Many thanks. About the model
You said "It's not advised to fit a model that includes D but excludes A". But if the nested model DE is better fitted than the full model ADE and the other nested model AE (from the AIC and statistical inference), how could I interpret it (about only one model selection)? Thanks.
Log in or register to post comments
In reply to Many thanks. About the model by Liz
easier with numbers: MZr DZr and Ns
I'll just add that this would be easier to discuss if you posted your MZ and DZ correlations and pair-numbers, along with some data on measurement reliability (if I didn't miss that?)
If you have 50 pairs of twins, power is extremely low to make any of these comparisons. If measurement error is 50%, even quite large data-sets will be all over the place. 1-MZ correlation is often a good indicator of measurement error.
The idea that it is unlikely a trait is affected by dominance but has no additive effects is built out of biological literature on selection suggesting higher response to selection for traits built out of additive alleles. Of course some traits like blue/brown eye colour follow a predominantly dominance-based inheritance.
If dominance is operating, we'd expect DZ correlations much smaller than the MZ correlations. The absolute correlations are interesting here: MZ of .1 and DZ of -.2 (which can happen with noisy measures and tiny datasets) is indicative of what…
If DE fits better than ADE, you have the problem referred to in regression as the principle of marginality: We never keep an interaction term and drop the main effects underlying it as this can artificially inflate the interaction term.
Hence one would be likely to interpret the ADE model, noting that in this case A could be dropped, and all the genetic effects interpreted as dominance. One might compare the D estimate in the two models: If it is larger in the DE model, this might suggest simple lack of power to detect A.
Inversely, in tiny data sets and/or noisy measures, one can see DZ correlations exceeding the MZ correlations: not because of a strange inheritance model, but because all correlations are tiny, and there's lots of bounce in the data. But this is why people collect thousands of pairs of twins and pay great attention to measurement.
Best, tim
PS: re references: I don't know of a paper proving the link between AIC and LRT p-values. Posting one would be useful, but it's also not necessary for following Rob's logic. One thing that I found helpful in building intuitions about twin models was just playing with example models: reducing the n and seeing how p-values and AIC change. It's a solid tonic for whenever others suggest using short-scales, clinical cut-offs, or skimping on subject numbers :-)
Log in or register to post comments
In reply to easier with numbers: MZr DZr and Ns by tbates
Thanks Tim. My data is brain
Log in or register to post comments
In reply to Thanks Tim. My data is brain by Liz
DZ > MZ
Log in or register to post comments
In reply to DZ > MZ by tbates
Many thanks~
My logic is that: I firstly use ACE model to estimate whether each trait is heritable. And then includ the heritable traits into the multivariate modle.
Log in or register to post comments
In reply to Thanks Tim. My data is brain by Liz
Be careful about outliers
Log in or register to post comments
In reply to Be careful about outliers by neale
Multiple comparison correction
Log in or register to post comments
In reply to Multiple comparison correction by Liz
Dealing with the multiple testing in neuroimaging genetics
The rationale is that the goal in drawing a map of heritability is very different from assessing whether a particular region is associated with a particular stimulus or response. By analogy, consider drawing a map of Earth in the 1400's. Cartographers could have declared Mount Everest and K2 as the only two places that were significantly elevated, so ships should definitely avoid steering into them. However, it turned out that it was much better to draw an approximate map of coastlines - heavily error-prone as they were, since this avoided shipwrecks much more often than just knowing roughly where big mountain ranges are. To my mind, it's the same deal with heritability maps of cortical thickness or surface area. It's not critically important to know that, e.g., despite a lot of multiple testing a bit of the prefrontal cortex is still significantly heritable. That most places are moderately heritable, and that some seem highly heritable, is valuable knowledge. It is unlike the situation with genome-wide association analysis, where deciding that one of the SNPs remains significant despite correction for multiple testing is important because expensive follow-up studies depend on that result being correct. By and large, following up a particularly heritable voxel at great expense is not the scientific goal. Similarly, if one were interested in stimulating or lesioning a region of the brain on the basis of association in an fMRI study, it would be sensible to be pretty sure that you had the right region and that the association was real.
Log in or register to post comments
In reply to Dealing with the multiple testing in neuroimaging genetics by AdminNeale
Heritability calculation within ROIs
Log in or register to post comments
In reply to Model selection vs. statistical inference by AdminRobK
statistical inference
And if the p value of AE vs. E is also not significant, however, the confidential interval of standardized a2 in the AE model do not contain zero. Should we get the same conclusion that the heritability of A is significant, or actually the power of confidential interval is weaker than that of p value from the model comparison?
Log in or register to post comments
In reply to statistical inference by Liz
If inferences are to be based
However, the confidence interval is a more difficult optimization problem. In the second scenario you describe, it's possible that the lower confidence limit is slightly off-target. What does the
summary()
output look like for the AE and E models, with argumentverbose=TRUE
(e.g.,summary(myFittedAEModel,verbose=TRUE)
)?Log in or register to post comments
In reply to If inferences are to be based by AdminRobK
statistical inference
About the second scenario, I think I explained it in a wrong way. I meant sometimes the CE model is the best fittted model based on the model selection, and the p-value of ACE vs. CE and ACE vs. AE were both not significant, whereas the p-value of AE vs. E and CE vs. E were both significant, and the CI of heritability in the AE model did not contain zero. Could I conclude that this phenotype is under the genetic influence?
For that there are thousands of variables in my data which limited sample size contained only dozens of twins pairs, there were various kinds of results and I'd like to set a clear common rule to define that if this variable/phenotype is affected by the genes.
Log in or register to post comments
In reply to statistical inference by Liz
Sometimes I also found that
Something is definitely wrong there. That shouldn't be possible. I'd be very curious to see the output from those models.
Assuming you want to base your inferences on a single "best" model, your conclusion must necessarily be that the phenotype is not under any genetic influence, because the estimate of A variance from your "best" model, CE, is zero. Does that mean you have to believe that the influence of genetics on the phenotype is literally zero? Of course not. The whole point of AIC-based model selection is to choose a model that optimally (in an information-theoretic sense) balances the bias and repeated-sampling variance of parameter estimates, for a given sample size. Models with more parameters tend to have less bias, at the cost of greater variance, whereas models with fewer parameters tend to have less variance, at the cost of greater bias. In the case presently under discussion, AIC is telling you that the CE model for the phenotype best balances bias and variance, given your (quite small) sample size.
Now...do you want to base your inferences on a single "best" model? The fact that you seem to be repeatedly getting yourself confused over pairwise likelihood-ratio tests makes me suspect that you actually don't. If that's the case, consider model averaging and multimodel inference. Or, just forget about model-selection, and fit the ACE model to every phenotype, since that's the model that will give you a freely calculated estimate of all three biometric variance components.
Log in or register to post comments
In reply to Sometimes I also found that by AdminRobK
I see, many thanks
About your doubt on that why I standardized my data before the model fitting (my another question), I have remenbered the purpose which is to avoid the "RED status 6", such like:
Warning message: In model 'ACE' Optimizer returned a non-zero status code 6. The model does not satisfy the first-order optimality conditions to the required accuracy, and no improved point for the merit function could be found during the final linesearch (Mx status RED) (Which error means the model was not identified, right?)
Sometimes in the model with more than two phenotypes, the mean may be different between them and the difference may be huge. So if I standardize the data first and set the start value as 0, the model may be more easier to get "Optimization".
Log in or register to post comments
In reply to I see, many thanks by Liz
reply to #18
It's not the only way, but I'd say it's the best way.
It can, though not necessarily. But it makes sense to me that you'd standardize the phenotypes if they are on really different scales, as you describe.
Log in or register to post comments
About the model selection
I'll explain. A model's AIC = -2logL + 2k, where L is the likelihood at the MLE, and k is the number of free parameters. Now suppose I'm comparing two models that differ by the free/fixed status of a single parameter. AIC will only prefer the model that frees the parameter if freeing the parameter decreases -2logL by more than 2. The LRT statistic is the change in -2logL, and the p-value for a chi-square test statistic of 2, on 1 degree-of-freedom, is 0.16 when rounded to two decimal places.
Among the ADE, AE, ACE, and CE models, which has the smallest AIC? I have a feeling it's the CE model.
Log in or register to post comments
In reply to About the model selection by AdminRobK
Many thank Rob. Yes, it's CE
Log in or register to post comments
In reply to About the model selection by AdminRobK
ACDE model
I'm wondering that if the ACDE model is rationable? I noted that in the book written by Neale at 2004 the ACDE model could be tested in MX. Hence if it is possible to use OpenMx to set the ACDE model as the full model, and compare ACE, ADE, CE, AE, CE submodels to it respectively? However, is there a saying that the D and C could not be estimated in a same model and why? Thanks.
Log in or register to post comments
In reply to ACDE model by Liz
The ACDE model, when fit via
Log in or register to post comments
In reply to The ACDE model, when fit via by AdminRobK
That's great. Many thanks for
Log in or register to post comments