Difference between Cholesky decomposition and variance-based ACE model

Posted on
No user picture. Liz Joined: 04/20/2016
Dear OpenMx experts,

May I ask what's the difference between multi-variate ACE models based on paths (Cholesky decomposition) and models based on variance, such as advantages and disadvantages of the two methods? I was told that the models based on variance and covariance will not suffer from the type I error issues, by compared to the Cholesky decomposition, why? Will the variance-based multi-variate ACE models better than the Cholesky decomposition-based models? Could I model a multi-variate ACE model based on variance in OpenMx? Any demo codes would be appreciated.

Thanks and best wishes,

Zhi

Replied on Tue, 12/11/2018 - 12:43
Picture of user. AdminRobK Joined: 01/24/2014

May I ask what's the difference between multi-variate ACE models based on paths (Cholesky decomposition) and models based on variance, such as advantages and disadvantages of the two methods? I was told that the models based on variance and covariance will not suffer from the type I error issues, by compared to the Cholesky decomposition, why?

The two are equivalent parameterizations. Under the "Cholesky" parameterization, the free parameters that define the model-expected covariance matrix are the elements of the Cholesky factors of the _A_, _C_, and _E_ covariance matrices. Under the "direct-symmetric" parameterization, the free parameters that define the model-expected covariance matrix are the elements of the _A_, _C_, and _E_ covariance matrices themselves. Much of the time--probably, most of the time--results will not differ between the two parameterizations. But, sometimes results _will_ differ, because the Cholesky parameterization imposes an implicit constraint that the direct-symmetric parameterization does not, namely, that the _A_, _C_, and _E_ covariance matrices must be positive-definite. But, the model-expected covariance matrix remains within the parameter space of the multivariate-normal distribution as long as it itself is positive-definite, whether or not the _A_, _C_, and _E_ covariance matrices are positive-definite.

Thus, the Cholesky parameterization introduces a "boundary condition," which can interfere with the usual asymptotic maximum-likelihood theory. One of the first (perhaps first) researchers to recognize how the Cholesky parameterization can interfere with Type I error rates was Greg Carey, in a 2005 paper, "Cholesky Problems". This year's Boulder Workshop was the first year we taught the direct-symmetric parameterization.

It is true that the direct-symmetric parameterization can result in negative estimates of variance components, but that is not necessarily a bad thing. I'd say the only real advantage the Cholesky parameterization has is its interpretability.

Could I model a multi-variate ACE model based on variance in OpenMx? Any demo codes would be appreciated.

See here for a collection of twin-modeling scripts for one phenotype and one covariate, maintained by Hermine Maes. The scripts that are described as "estimating variance components" use the direct-symmetric parameterization, and those described as "estimating path coefficients" use the Cholesky parameterization.

Replied on Wed, 12/12/2018 - 09:59
No user picture. Liz Joined: 04/20/2016

In reply to by AdminRobK

Thank you very much for your always useful and detailed advices, Rob. May I ask another question which may be not related to this post? When I summarized a full trivariate Cholesky ACE model in OpenMx, the CIs of all the parameters were NA, however, I could get the CIs in the submodels such as AE or CE. In addition, if I add the flag "verbose = T" in the summarization of the full ACE model, I could also get the CIs of most parameters estimation. May I know why and if the CIs of the full model after adding "verbose = T" were trustable? (Actually I have found some previous related posts on this question, however, I cannot reach them not matter what the search engine or computer I use.)

Best wishes,

Zhi

Replied on Wed, 12/12/2018 - 11:24
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Liz

May I know why and if the CIs of the full model after adding "verbose = T" were trustable?

They are not trustworthy, which is why they are not presented in the non-verbose `summary()` output. However, they are not necessarily wrong, either, but there is reason to believe that they MAY be wrong. The reason for why each confidence limit may be wrong is given in the 'diagnostic' column of the CI details table. Which optimizer are you using?

(Actually I have found some previous related posts on this question, however, I cannot reach them not matter what the search engine or computer I use.)

The best way to search the OpenMx website is to enter your search terms into the search text box in the upper right of each page, click the "Search" button, and then when the search results page opens, click on "Content" above the text box near the top of the page.

Replied on Thu, 12/13/2018 - 00:51
No user picture. Liz Joined: 04/20/2016

In reply to by AdminRobK

Thank you very much, Rob. I have tried both the NPSOL and SLSQP, and their results were similar that the CIs were missing. Here is the version of R and OpenMx I use now:

OpenMx version: 2.11.5 [GIT v2.11.5]
R version: R version 3.5.1 (2018-07-02)
Platform: x86_64-w64-mingw32
Default optimiser: SLSQP
NPSOL-enabled?: Yes
OpenMP-enabled?: No

Please see attached for the part of CIs in my codes. I could run them smoothly and the results seemed reasonable, excepted for the missing CIs. Could you please kindly tell me if there is something wrong about my codes? Thank you very much.

sta <- mxAlgebra(iSD %% a, name="sta") #standarized path coefficients of A
stc <- mxAlgebra(iSD %% c, name="stc")
ste <- mxAlgebra(iSD %% e, name="ste")
StandardizedA <- mxAlgebra(A/V, name="StandardizedA")
StandardizedC <- mxAlgebra(C/V, name="StandardizedC")
StandardizedE <- mxAlgebra(E/V, name="StandardizedE")
CorA <- mxAlgebra(solve(sqrt(IA)) %&% A, name="CorA")
CorC <- mxAlgebra(solve(sqrt(IC)) %&% C, name="CorC")
CorE <- mxAlgebra(solve(sqrt(IE)) %&% E, name="CorE")
CorP <- mxAlgebra(solve(sqrt(I*V)) %&% V, name="CorP")
CI <- mxCI(c('sta', 'stc', 'ste','StandardizedA', 'StandardizedC', 'StandardizedE','CorA', 'CorC', 'CorE', 'CorP'))

dataMZ <- mxData( observed=mzData, type="raw" )
dataDZ <- mxData( observed=dzData, type="raw" )

objMZ <- mxExpectationNormal( covariance="expCovMZ", means="expMean",
dimnames=selVars )
objDZ <- mxExpectationNormal( covariance="expCovDZ", means="expMean",
dimnames=selVars )
funML <- mxFitFunctionML()

pars <- list( pathA, pathC, pathE, covA, covC, covE, covP,
matI, invSD, meanG, meanT, sta, stc, ste,
StandardizedA, StandardizedC, StandardizedE,
CorA, CorC, CorE, CorP)
modelMZ <- mxModel( pars, covMZ, dataMZ, objMZ, funML, name="MZ" )
modelDZ <- mxModel( pars, covDZ, dataDZ, objDZ, funML, name="DZ" )
fitML <- mxFitFunctionMultigroup(c("MZ.fitfunction","DZ.fitfunction") )
CholAceModel <- mxModel( "CholACE", pars, modelMZ, modelDZ, fitML, CI)

Replied on Thu, 12/13/2018 - 22:55
No user picture. Liz Joined: 04/20/2016

Thanks, Rob. Please see below for my script:

mzData <- subset(nl, zyg==1, selVars)
dzData <- subset(nl, zyg==2, selVars)
describe(mzData, skew=F)
describe(dzData, skew=F)
dim(mzData)
dim(dzData)

# Generate Descriptive Statistics
colMeans(mzData,na.rm=TRUE)
colMeans(dzData,na.rm=TRUE)
cov(mzData,use="complete")
cov(dzData,use="complete")
cor(mzData,use="complete")
cor(dzData,use="complete")

# Set Starting Values
meVals <- c(0.5,0.5,0.5) # start value for means
meanLabs <- paste(Vars,"mean",sep="") # create labels for means
paVal <- .6 # start value for path coefficient
paLBo <- .0001 # start value for lower bounds
paValD <- vech(diag(paVal,nv,nv)) # start values for diagonal of covariance matrix
paLBoD <- diag(paLBo,nv,nv) # lower bounds for diagonal of covariance matrix
paLBoD[lower.tri(paLBoD)] <- -10 # lower bounds for below diagonal elements
paLBoD[upper.tri(paLBoD)] <- NA # lower bounds for above diagonal elements

# Create Labels for Lower Triangular Matrices
aLabs <- paste("a",rev(nv+1-sequence(1:nv)),rep(1:nv,nv:1),sep="_")
cLabs <- paste("c",rev(nv+1-sequence(1:nv)),rep(1:nv,nv:1),sep="_")
eLabs <- paste("e",rev(nv+1-sequence(1:nv)),rep(1:nv,nv:1),sep="_")

# Matrices declared to store a, c, and e Path Coefficients
pathA <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=TRUE,
values=paValD, labels=aLabs, lbound=paLBoD, name="a" )
pathC <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=TRUE,
values=paValD, labels=cLabs, lbound=paLBoD, name="c" )
pathE <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=TRUE,
values=paValD, labels=eLabs, lbound=paLBoD, name="e" )

# Matrices generated to hold A, C, and E computed Variance Components
covA <- mxAlgebra( expression=a %*% t(a), name="A" )
covC <- mxAlgebra( expression=c %*% t(c), name="C" )
covE <- mxAlgebra( expression=e %*% t(e), name="E" )

# Algebra to compute total variances and standard deviations (diagonal only)
covP <- mxAlgebra( expression=A+C+E, name="V" )
matI <- mxMatrix( type="Iden", nrow=nv, ncol=nv, name="I")
invSD <- mxAlgebra( expression=solve(sqrt(I*V)), name="iSD")

# Algebra for expected Mean and Variance/Covariance Matrices in MZ & DZ twins
meanG <- mxMatrix( type="Full", nrow=1, ncol=nv, free=TRUE,
values=meVals, labels=meanLabs, name="Mean" )
meanT <- mxAlgebra( expression= cbind(Mean,Mean), name="expMean" )
covMZ <- mxAlgebra( expression=
rbind( cbind(A+C+E , A+C),
cbind(A+C , A+C+E)), name="expCovMZ" )
covDZ <- mxAlgebra( expression=
rbind( cbind(A+C+E , 0.5%x%A+C),
cbind(0.5%x%A+C , A+C+E)), name="expCovDZ" )

# Standeraized parameters and CI
sta <- mxAlgebra(iSD %*% a, name="sta")
stc <- mxAlgebra(iSD %*% c, name="stc")
ste <- mxAlgebra(iSD %*% e, name="ste")
StandardizedA <- mxAlgebra(A/V, name="StandardizedA")
StandardizedC <- mxAlgebra(C/V, name="StandardizedC")
StandardizedE <- mxAlgebra(E/V, name="StandardizedE")

CI <- mxCI(c('StandardizedA', 'StandardizedC', 'StandardizedE'))

# Data objects for Multiple Groups
dataMZ <- mxData( observed=mzData, type="raw" )
dataDZ <- mxData( observed=dzData, type="raw" )

# Objective objects for Multiple Groups
objMZ <- mxExpectationNormal( covariance="expCovMZ", means="expMean",
dimnames=selVars )
objDZ <- mxExpectationNormal( covariance="expCovDZ", means="expMean",
dimnames=selVars )
funML <- mxFitFunctionML()

# Combine Groups
pars <- list( pathA, pathC, pathE, covA, covC, covE, covP,
matI, invSD, meanG, meanT, sta, stc, ste,
StandardizedA, StandardizedC, StandardizedE)
modelMZ <- mxModel( pars, covMZ, dataMZ, objMZ, funML, name="MZ" )
modelDZ <- mxModel( pars, covDZ, dataDZ, objDZ, funML, name="DZ" )
fitML <- mxFitFunctionMultigroup(c("MZ.fitfunction","DZ.fitfunction") )
CholAceModel <- mxModel( "CholACE", pars, modelMZ, modelDZ, fitML, CI)

# Run Cholesky Decomposition ACE model
CholAceFit <- mxRun(CholAceModel, intervals = TRUE)
CholAceSumm <- summary(CholAceFit)
CholAceSumm

The results I got from this script were:
CholACE.StandardizedA[1,1] NA 0.34696886 NA !!!
CholACE.StandardizedA[2,2] NA 0.57933647 NA !!!
CholACE.StandardizedA[3,3] NA 0.27260377 0.5327416 !!!
CholACE.StandardizedC[1,1] NA 0.07631260 NA !!!
CholACE.StandardizedC[2,2] NA 0.02342186 NA !!!
CholACE.StandardizedC[3,3] NA 0.01657537 NA !!!
CholACE.StandardizedE[1,1] NA 0.57671854 NA !!!
CholACE.StandardizedE[2,2] NA 0.39724167 NA !!!
CholACE.StandardizedE[3,3] NA 0.71082086 NA !!!

After I added the verbose, I got the missing CIs:

parameter side value fit diagnostic statusCode
CholACE.StandardizedA[1,1] lower 0.00576769 1388.242545 active box constraint OK
CholACE.StandardizedA[1,1] upper 0.620436827 1388.2481 active box constraint OK
CholACE.StandardizedA[2,2] lower 0.220852389 1388.264244 active box constraint nonzero gradient/red
CholACE.StandardizedA[2,2] upper 0.74937414 1388.245092 active box constraint nonzero gradient/red
CholACE.StandardizedA[3,3] lower 0.015417542 1388.244836 active box constraint nonzero gradient/red
CholACE.StandardizedA[3,3] upper 0.532741586 1388.251373 success nonzero gradient/red
CholACE.StandardizedC[1,1] lower 8.69E-09 1388.23726 active box constraint nonzero gradient/red
CholACE.StandardizedC[1,1] upper 0.402550673 1388.252648 active box constraint nonzero gradient/red
CholACE.StandardizedC[2,2] lower 9.07E-09 1388.23726 active box constraint nonzero gradient/red
CholACE.StandardizedC[2,2] upper 0.290058001 1388.25808 active box constraint nonzero gradient/red
CholACE.StandardizedC[3,3] lower 9.24E-09 1388.23726 active box constraint nonzero gradient/red
CholACE.StandardizedC[3,3] upper 0.236906447 1388.253135 active box constraint nonzero gradient/red
CholACE.StandardizedE[1,1] lower 0.37856246 1388.247897 active box constraint nonzero gradient/red
CholACE.StandardizedE[1,1] upper 0.824304393 1388.252819 active box constraint nonzero gradient/red
CholACE.StandardizedE[2,2] lower 0.248666688 1388.244859 active box constraint nonzero gradient/red
CholACE.StandardizedE[2,2] upper 0.616918327 1388.253597 active box constraint nonzero gradient/red
CholACE.StandardizedE[3,3] lower 0.464477104 1388.251145 active box constraint nonzero gradient/red
CholACE.StandardizedE[3,3] upper 0.961257369 1388.246413 active box constraint nonzero gradient/red

Replied on Fri, 12/14/2018 - 10:36
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Liz

All the CI diagnostics say "active box constraint", which means that one or more free parameters are close to their lbounds or ubounds. Active box constraints represent a boundary condition (described upthread in the context of the Cholesky parameterization), and therefore mean that the confidence limit in question may be wrong. Try eliminating the `lbound`s from this code:

# Matrices declared to store a, c, and e Path Coefficients
pathA <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=TRUE,
values=paValD, labels=aLabs, lbound=paLBoD, name="a" )
pathC <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=TRUE,
values=paValD, labels=cLabs, lbound=paLBoD, name="c" )
pathE <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=TRUE,
values=paValD, labels=eLabs, lbound=paLBoD, name="e" )
Replied on Sun, 12/16/2018 - 23:34
No user picture. Liz Joined: 04/20/2016

Thank you very much, Rob. It works now, although the results are very close to the "verbose = T". May I ask how many kinds of diagnostic of CI results in openMx, and how to deal with them? Sorry I did not find it in the manual.

I have also tried the direct-symmetric parameterization. However, the results were hard to interpret:

Results of Cholesky decomposition:
CholACE.StandardizedA[1,1] 5.7629863e-03 3.4696840e-01 0.620436834
CholACE.StandardizedA[2,2] 2.2085240e-01 5.7933417e-01 0.749374149
CholACE.StandardizedA[3,3] 1.5417533e-02 2.7260201e-01 0.532741610

Results of direct-symmetric parameterization:
twoACEvc.StandardizedA[1,1] -0.129377109 0.632041104 1.390309000
twoACEvc.StandardizedA[2,2] 0.580567884 1.316408772 2.054897056
twoACEvc.StandardizedA[3,3] 0.236514247 1.057349468 1.794373138

The parameter estimation of direct-symmetric parameterization could be higher than 1. How could I know the heritability and its significance of phenotype in the variance-based model? Based on the above CIs, can I draw a conclusion that the first phenotype is heritable in the Cholesky decomposition, but not in direct-symmetric parameterization? Looking forward to your kind suggestions.

Replied on Mon, 12/17/2018 - 10:33
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Liz

May I ask how many kinds of diagnostic of CI results in openMx, and how to deal with them? Sorry I did not find it in the manual.

See the code in the source repository, in src/ComputeGD.cpp , starting at line 1541. Of those eight CI diagnostics, I have only ever seen 4 occur in practice: "success", "alpha level not reached", "bound infeasible", and "active box constraint".

The parameter estimation of direct-symmetric parameterization could be higher than 1. How could I know the heritability and its significance of phenotype in the variance-based model? Based on the above CIs, can I draw a conclusion that the first phenotype is heritable in the Cholesky decomposition, but not in direct-symmetric parameterization? Looking forward to your kind suggestions.

There doesn't look like there is any evidence for heritability of the first phenotype. Even under the Cholesky parameterization, the lower confidence limit is essentially zero (it can't actually reach zero because of the implicit constraint inherent to the Cholesky parameterization).

Replied on Mon, 12/17/2018 - 10:56
No user picture. Liz Joined: 04/20/2016

Thanks Rob. The parameter estimations were the results of the full ACE model. However, if the sub-models were introduced the AE model would survived. In this condition, which one would be more trustable, the CIs of the full ACE or the model comparison between the full genetic model and its sub-models?

How could we get the heritability from the direct-symmetric parameterization and explain the results? The parameters estimation were higher than 1 and the sum of variance of A, C and E seemed not equal to 1.

Replied on Tue, 12/18/2018 - 10:27
Picture of user. AdminRobK Joined: 01/24/2014

Thanks Rob. The parameter estimations were the results of the full ACE model. However, if the sub-models were introduced the AE model would survived. In this condition, which one would be more trustable, the CIs of the full ACE or the model comparison between the full genetic model and its sub-models?

Assuming that the full ACE model and the submodels converged, and that all of the CIs succeeded, then I would say both are informative. The model-comparison chi-square for ACE versus (say) AE is an omnibus test of the null hypothesis that there is no _C_, that is, that the 6(?) _C_-related parameters are all zero. The CIs instead represent inference about one quantity at a time.

How could we get the heritability from the direct-symmetric parameterization and explain the results? The parameters estimation were higher than 1 and the sum of variance of A, C and E seemed not equal to 1.

I'd like to see those results.

Replied on Wed, 12/19/2018 - 00:22
No user picture. Liz Joined: 04/20/2016

Thanks. So the results of CIs and model comparisons are independent to each other, could I understand it in this way? However, if the AE model survives and the CI of A in AE does not contain zero, could I conclude that this phenotype is heritable even the CI of A in the full ACE contains zero?

Please see below for the results of the direct-symmetric parameterization (the first phenotype):
twoACEvc.StandardizedA[1,1] -0.129377109 0.632041104 1.390309000
twoACEvc.StandardizedC[1,1] -0.851222975 -0.192160561 0.409548567
twoACEvc.StandardizedE[1,1] 0.369370659 0.560119458 0.825937713

How to explain the negative estimates and quantify the heritability based on this results? In addition, the correlation matrix of C were all NA:

twoACEvc.CorC[1,1] NA NaN NA !!!
twoACEvc.CorC[2,1] NA NaN NA !!!
twoACEvc.CorC[3,1] NA NaN NA !!!
twoACEvc.CorC[1,2] NA NaN NA !!!
twoACEvc.CorC[2,2] NA NaN NA !!!
twoACEvc.CorC[3,2] NA NaN NA !!!
twoACEvc.CorC[1,3] NA NaN NA !!!
twoACEvc.CorC[2,3] NA NaN NA !!!
twoACEvc.CorC[3,3] NA NaN NA !!!

I have tried "verbose = T", the diagnostic was "alpha level not reached". Could this be caused by negative estimates of factor C?

Replied on Wed, 12/19/2018 - 09:30
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Liz

If you're going to report and interpret results from a single "best" model, then my usual recommendation is to report point estimates and confidence intervals from the model with the smallest AIC.

if the AE model survives and the CI of A in AE does not contain zero, could I conclude that this phenotype is heritable even the CI of A in the full ACE contains zero?

That sounds like a reasonable interpretation to me.

How to explain the negative estimates and quantify the heritability based on this results? In addition, the correlation matrix of C were all NA:

The fact that the first phenotype's c² is estimated as negative in the ACE model is a signal that you should consider modeling the first phenotype as ADE. I would guess that yes, the _C_ correlation matrix is full of `NA`s due to the negative estimate of the first phenotype's _C_ variance component.

Replied on Wed, 12/19/2018 - 22:57
Picture of user. Nengzhi Joined: 04/30/2018

Hi! My similar bivariate ACE model results were as below:
confidence intervals:
CholCE.corC[2,1] NA 1.000000e+00 NA !!!
CholCE.corC[1,2] NA 1.000000e+00 NA !!!
CholCE.corC[2,2] NA 1.000000e+00 NA !!!
CholCE.corE[1,1] NA 1.000000e+00 NA !!!
CholCE.corE[2,1] NA -8.420869e-02 NA !!!
CholCE.corE[1,2] NA -8.420869e-02 NA !!!
CholCE.corE[2,2] NA 1.000000e+00 NA !!!
CI details:
58 CholCE.corC[1,1] 1.00000000 upper 1156.910 5.010251 0.10257746 -4.200825e-02
59 CholCE.corC[2,1] 0.96243518 lower 1153.141 5.238043 0.09977091 -2.814626e-02
60 CholCE.corC[2,1] 1.00000000 upper 1153.069 5.238043 0.09978666 -5.515351e-07
61 CholCE.corC[1,2] 1.00000000 lower 1153.069 5.238043 0.09978666 -5.515351e-07
62 CholCE.corC[1,2] 1.00000000 upper 1153.069 5.238043 0.09978666 -5.515351e-07
63 CholCE.corC[2,2] 1.00000000 lower 1153.069 5.238043 0.09978666 -5.515351e-07
64 CholCE.corC[2,2] 1.00000000 upper 1153.069 5.238043 0.09978666 -5.515351e-07
65 CholCE.corE[1,1] 1.00000000 lower 1153.069 5.238043 0.09978666 -5.515351e-07
66 CholCE.corE[1,1] 1.00000000 upper 1153.069 5.238043 0.09978666 -5.515351e-07

I am confused why there is no CholCE.corE[1,2] & CholCE.corE[2,2] results in CI details. How can I get it ?

Replied on Thu, 12/20/2018 - 02:41
Picture of user. Nengzhi Joined: 04/30/2018

Here is my scripts:

# Run Cholesky CE model
CholCeModel <- mxModel( CholAceFit, name="CholCE" )
pathA <- mxMatrix( type="Lower", nrow=nv, ncol=nv, free=FALSE, values=paZero,
labels=aLabs, lbound=paLBoD, name="a" )#pathA is 2X2 matrix
CholCeModel$MZ$a <-pathA
CholCeModel$DZ$a <-pathA
CholCeModel$a <-pathA
rA <- mxAlgebra(I*A, name="corA")
CholCeModel$corA <- rA
CholCeModel$MZ$corA <- rA
CholCeModel$DZ$corA <- rA

CholCeFit <- mxRun(CholCeModel, intervals = TRUE)
CholCeSumm <- summary(CholCeFit,verbose=T)
CholCeSumm
round(CholCeFit@output$estimate,4)
expectedMeansCovariances(CholCeFit)
tableFitStatistics(CholAceFit,CholCeFit)

*****Here is another part of the results:
58 success infeasible non-linear constraint
59 active box constraint infeasible non-linear constraint
60 alpha level not reached iteration limit/blue
61 alpha level not reached iteration limit/blue
62 alpha level not reached iteration limit/blue
63 alpha level not reached iteration limit/blue
64 alpha level not reached iteration limit/blue
65 alpha level not reached iteration limit/blue
66 alpha level not reached iteration limit/blue
[ reached getOption("max.print") -- omitted 14 rows ]

I don't know how to understand this part?
Wish for your help!
Thank you!

Replied on Thu, 12/20/2018 - 11:12
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Nengzhi

You must be running an out-of-date version of OpenMx. As of version 2.11 (see release announcement here), the columns of the "CI details" table have been reordered to make the table easier to read, and a bug that caused a lot of confidence limits to fail with "iteration limit/blue" was fixed. What is your `mxVersion()` output?

I am confused why there is no CholCE.corE[1,2] & CholCE.corE[2,2] results in CI details. How can I get it ?
[snip]
[ reached getOption("max.print") -- omitted 14 rows ]

I bet the missing rows are among the 14 that got cut off. Update to OpenMx v2.11.5. Then, find the CI-details table inside `CholCeSumm`, and print only its first 7 columns. If you restore a saved workspace, you'll have to at least re-create `CholCeSumm` under version 2.11.5.

Note that if 'corE' is a proper correlation matrix, 'corE[2,2]' is going to be fixed at 1. Also, the CI diagnostic 'active box constraint' is a signal that there's at least one parameter's `lbound` or `ubound` interfering with confidence-limit optimization.

Replied on Thu, 12/20/2018 - 13:26
No user picture. Liz Joined: 04/20/2016

In reply to by AdminRobK

Thank you Rob. I have tried the variance-based ADE model and found that if the parameter estimate is negative, then the genetic/environmental correlation would be NAs. I think you are right about that the negative estimates do influence the estimation of correlation matrix. My question is what we could get from the variance-based ACE/ADE model, except for that it may be not susceptible to the type I error? From the path-based ACE/ADE model, we could quantify the heritability that to which extent genes or environments influence the phenotype, because the estimates of A, D, C, E are positive in the path-based models. How could I we interpret the negative estimates in the variance-based models? In the estimates of variance-based ACE model below, I think we could not conclude that the heritability of this phenotype is 0.63, the ratio of factor A (heritability?) "included" by compared with the path-based ACE model. And how could we interpret the negative estimate of the common environmental factor C?

twoACEvc.StandardizedA[1,1] -0.129377109 0.632041104 1.390309000
twoACEvc.StandardizedC[1,1] -0.851222975 -0.192160561 0.409548567
twoACEvc.StandardizedE[1,1] 0.369370659 0.560119458 0.825937713

Replied on Fri, 12/21/2018 - 10:40
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Liz

Thank you Rob. I have tried the variance-based ADE model and found that if the parameter estimate is negative, then the genetic/environmental correlation would be NAs. I think you are right about that the negative estimates do influence the estimation of correlation matrix.

It could not be otherwise. Turning a covariance matrix into its corresponding correlation matrix involves taking the square roots of the covariance matrix's diagonal elements. If a diagonal element is negative, its `sqrt()` will be `NaN`, and the `NaN`s will propagate into at least part of the resulting correlation matrix via multiplication and addition.

How could I we interpret the negative estimates in the variance-based models?

The simple answer is that you can't really interpret them. They're a signal that the model being fitted is a poor choice for the data. Are you getting negative variance-component estimates in your "AIC-best" model?

Replied on Fri, 12/21/2018 - 18:53
No user picture. Liz Joined: 04/20/2016

In reply to by AdminRobK

I see now. All the variance-component estimates in my best-fitted model were positive. So the variance-based model may help us clarify the fit of certain model and the significance of parameter estimates. We still need the path-based model to quantify the heritability. Thank you very much for your kind help, Rob.
Replied on Fri, 12/21/2018 - 23:23
Picture of user. Nengzhi Joined: 04/30/2018

In reply to by AdminRobK

Thank you!Rob!
I have update OpenMx to latest release. However, the results didn't change.
I don't know how to use "[ reached getOption("max.print") -- omitted 14 rows ]" to see the omitted part. Could you please give me an example?
Replied on Mon, 12/24/2018 - 14:29
No user picture. Liz Joined: 04/20/2016

One more question: When I tried the Cholesky Decomposition back, I found that the shared environmental correlations were all 1, although their CIs contain zero, if this situation is interpretable?

CholACE.CorC[1,1] 1.00000000 1.00000000 1.000000000 !!!
CholACE.CorC[2,1] -1.00000000 1.00000000 1.000000000
CholACE.CorC[3,1] -1.00000000 1.00000000 1.000000000
CholACE.CorC[1,2] -1.00000000 1.00000000 1.000000000
CholACE.CorC[2,2] 1.00000000 1.00000000 1.000000000 !!!
CholACE.CorC[3,2] -1.00000000 1.00000000 1.000000000
CholACE.CorC[1,3] -1.00000000 1.00000000 1.000000000
CholACE.CorC[2,3] -1.00000000 1.00000000 1.000000000
CholACE.CorC[3,3] 1.00000000 1.00000000 1.000000000 !!!

Replied on Tue, 01/08/2019 - 14:02
Picture of user. AdminRobK Joined: 01/24/2014

In reply to by Liz

That is certainly a possible result if there is very little shared-environmental variance in the phenotypes, or if very little of their covariance is due to the shared environment. What this result is telling you is that your estimates of the shared-environmental correlations have essentially zero statistical precision.
Replied on Tue, 01/08/2019 - 13:52
Picture of user. AdminRobK Joined: 01/24/2014

Very relevant to the topic of this thread: Type I Error Rates and Parameter Bias in Multivariate Behavioral Genetic Models.

Abstract:

For many multivariate twin models, the numerical Type I error rates are lower than theoretically expected rates using a likelihood ratio test (LRT), which implies that the significance threshold for statistical hypothesis tests is more conservative than most twin researchers realize. This makes the numerical Type II error rates higher than theoretically expected. Furthermore, the discrepancy between the observed and expected error rates increases as more variables are included in the analysis and can have profound implications for hypothesis testing and statistical inference. In two simulation studies, we examine the Type I error rates for the Cholesky decomposition and Correlated Factors models. Both show markedly lower than nominal Type I error rates under the null hypothesis, a discrepancy that increases with the number of variables in the model. In addition, we observe slightly biased parameter estimates for the Cholesky decomposition and Correlated Factors models. By contrast, if the variance–covariance matrices for variance components are estimated directly (without constraints), the numerical Type I error rates are consistent with theoretical expectations and there is no bias in the parameter estimates regardless of the number of variables analyzed. We call this the direct symmetric approach. It appears that each model-implied boundary, whether explicit or implicit, increases the discrepancy between the numerical and theoretical Type I error rates by truncating the sampling distributions of the variance components and inducing bias in the parameters. The direct symmetric approach has several advantages over other multivariate twin models as it corrects the Type I error rate and parameter bias issues, is easy to implement in current software, and has fewer optimization problems. Implications for past and future research, and potential limitations associated with direct estimation of genetic and environmental covariance matrices are discussed.