Fit Confirmatory Factor Analysis Models
bcfa.Rd
Fit a Confirmatory Factor Analysis (CFA) model.
Arguments
- ...
Default lavaan arguments. See
lavaan
.- cp
Handling of prior distributions on covariance parameters: possible values are
"srs"
(default) or"fa"
. Option"fa"
is only available fortarget="jags"
.- dp
Default prior distributions on different types of parameters, typically the result of a call to
dpriors()
. See thedpriors()
help file for more information.- n.chains
Number of desired MCMC chains.
- burnin
Number of burnin/warmup iterations (not including the adaptive iterations, for target="jags"). Defaults to 4000 or target="jags" and 500 for Stan targets.
- sample
The total number of samples to take after burnin. Defaults to 10000 for target="jags" and 1000 for Stan targets.
- adapt
For target="jags", the number of adaptive iterations to use at the start of sampling. Defaults to 1000.
- mcmcfile
If
TRUE
, the JAGS/Stan model will be written to file (in the lavExport directory). Can also supply a character string, which serves as the name of the directory to which files will be written.- mcmcextra
A list with potential names
syntax
(unavailable for target="stan"
),monitor
,data
, andllnsamp
. Thesyntax
object is a text string containing extra code to insert in the JAGS/Stan model syntax. Thedata
object is a list of extra data to send to the JAGS/Stan model. Ifmoment_match_k_threshold
is specified withindata
the looic of the model will be calculated using moment matching. Themonitor
object is a character vector containing extra JAGS/Stan parameters to monitor. Thellnsamp
object is only relevant to models with ordinal variables, and specifies the number of samples that should be drawn to approximate the model log-likelihood (larger numbers imply higher accuracy and longer time). This log-likelihood is specifically used to compute information criteria.- inits
If it is a character string, the options are currently
"simple"
(default),"Mplus"
,"prior"
, or"jags"
. In the first two cases, parameter values are set as though they will be estimated via ML (seelavaan
). The starting parameter value for each chain is then perturbed from the original values through the addition of random uniform noise. If"prior"
is used, the starting parameter values are obtained based on the prior distributions (while also trying to ensure that the starting values will not crash the model estimation). If"jags"
, no starting values are specified and JAGS will choose values on its own (and this will probably crash Stan targets). You can also supply a list of starting values for each chain, where the list format can be obtained from, e.g.,blavInspect(fit, "inits")
. Finally, you can specify starting values in a similar way to lavaan, using the lavaanstart
argument (see the lavaan documentation for all the options there). In this case, you should also setinits="simple"
, and be aware that the same starting values will be used for each chain.- convergence
Useful only for
target="jags"
. If"auto"
, parameters are sampled until convergence is achieved (viaautorun.jags()
). In this case, the argumentsburnin
andsample
are passed toautorun.jags()
asstartburnin
andstartsample
, respectively. Otherwise, parameters are sampled as specified by the user (or by therun.jags
defaults).- target
Desired MCMC sampling, with
"stan"
(pre-compiled marginal approach) as default. Also available is"vb"
, which calls the rstan functionvb()
. Other options include"jags"
,"stancond"
, and"stanclassic"
, which sample latent variables and provide some greater functionality (because syntax is written "on the fly"). But they are slower and less efficient.- save.lvs
Should sampled latent variables (factor scores) be saved? Logical; defaults to FALSE
- wiggle
Labels of equality-constrained parameters that should be "approximately" equal. Can also be "intercepts", "loadings", "regressions", "means".
- wiggle.sd
The prior sd (of normal distribution) to be used in approximate equality constraints. Can be one value, or (for target="stan") a numeric vector of values that is the same length as wiggle.
- prisamp
Should samples be drawn from the prior, instead of the posterior (
target="stan"
only)? Logical; defaults to FALSE- jags.ic
Should DIC be computed the JAGS way, in addition to the BUGS way? Logical; defaults to FALSE
- seed
A vector of length
n.chains
(for target"jags"
) or an integer (for target"stan"
) containing random seeds for the MCMC run. IfNULL
, seeds will be chosen randomly.- bcontrol
A list containing additional parameters passed to
run.jags
(orautorun.jags
) orstan
. See the manpage of those functions for an overview of the additional parameters that can be set.
Details
The bcfa
function is a wrapper for the more general
blavaan
function, using the following default
lavaan
arguments:
int.ov.free = TRUE
, int.lv.free = FALSE
,
auto.fix.first = TRUE
(unless std.lv = TRUE
),
auto.fix.single = TRUE
, auto.var = TRUE
,
auto.cov.lv.x = TRUE
,
auto.th = TRUE
, auto.delta = TRUE
,
and auto.cov.y = TRUE
.
Value
An object that inherits from class lavaan, for which several methods
are available, including a summary
method.
References
Edgar C. Merkle, Ellen Fitzsimmons, James Uanhoro, & Ben Goodrich (2021). Efficient Bayesian Structural Equation Modeling in Stan. Journal of Statistical Software, 100(6), 1-22. URL http://www.jstatsoft.org/v100/i06/.
Edgar C. Merkle & Yves Rosseel (2018). blavaan: Bayesian Structural Equation Models via Parameter Expansion. Journal of Statistical Software, 85(4), 1-30. URL http://www.jstatsoft.org/v85/i04/.
Yves Rosseel (2012). lavaan: An R Package for Structural Equation Modeling. Journal of Statistical Software, 48(2), 1-36. URL http://www.jstatsoft.org/v48/i02/.
Examples
data(HolzingerSwineford1939, package = "lavaan")
# The Holzinger and Swineford (1939) example
HS.model <- ' visual =~ x1 + x2 + x3
textual =~ x4 + x5 + x6
speed =~ x7 + x8 + x9 '
if (FALSE) { # \dontrun{
fit <- bcfa(HS.model, data = HolzingerSwineford1939)
summary(fit)
} # }
# A short run for rough results
fit <- bcfa(HS.model, data = HolzingerSwineford1939, burnin = 100, sample = 100,
n.chains = 2)
#>
#> SAMPLING FOR MODEL 'stanmarg' NOW (CHAIN 1).
#> Chain 1:
#> Chain 1: Gradient evaluation took 0.000332 seconds
#> Chain 1: 1000 transitions using 10 leapfrog steps per transition would take 3.32 seconds.
#> Chain 1: Adjust your expectations accordingly!
#> Chain 1:
#> Chain 1:
#> Chain 1: WARNING: There aren't enough warmup iterations to fit the
#> Chain 1: three stages of adaptation as currently configured.
#> Chain 1: Reducing each adaptation stage to 15%/75%/10% of
#> Chain 1: the given number of warmup iterations:
#> Chain 1: init_buffer = 15
#> Chain 1: adapt_window = 75
#> Chain 1: term_buffer = 10
#> Chain 1:
#> Chain 1: Iteration: 1 / 200 [ 0%] (Warmup)
#> Chain 1: Iteration: 20 / 200 [ 10%] (Warmup)
#> Chain 1: Iteration: 40 / 200 [ 20%] (Warmup)
#> Chain 1: Iteration: 60 / 200 [ 30%] (Warmup)
#> Chain 1: Iteration: 80 / 200 [ 40%] (Warmup)
#> Chain 1: Iteration: 100 / 200 [ 50%] (Warmup)
#> Chain 1: Iteration: 101 / 200 [ 50%] (Sampling)
#> Chain 1: Iteration: 120 / 200 [ 60%] (Sampling)
#> Chain 1: Iteration: 140 / 200 [ 70%] (Sampling)
#> Chain 1: Iteration: 160 / 200 [ 80%] (Sampling)
#> Chain 1: Iteration: 180 / 200 [ 90%] (Sampling)
#> Chain 1: Iteration: 200 / 200 [100%] (Sampling)
#> Chain 1:
#> Chain 1: Elapsed Time: 0.44 seconds (Warm-up)
#> Chain 1: 0.45 seconds (Sampling)
#> Chain 1: 0.89 seconds (Total)
#> Chain 1:
#>
#> SAMPLING FOR MODEL 'stanmarg' NOW (CHAIN 2).
#> Chain 2:
#> Chain 2: Gradient evaluation took 0.000267 seconds
#> Chain 2: 1000 transitions using 10 leapfrog steps per transition would take 2.67 seconds.
#> Chain 2: Adjust your expectations accordingly!
#> Chain 2:
#> Chain 2:
#> Chain 2: WARNING: There aren't enough warmup iterations to fit the
#> Chain 2: three stages of adaptation as currently configured.
#> Chain 2: Reducing each adaptation stage to 15%/75%/10% of
#> Chain 2: the given number of warmup iterations:
#> Chain 2: init_buffer = 15
#> Chain 2: adapt_window = 75
#> Chain 2: term_buffer = 10
#> Chain 2:
#> Chain 2: Iteration: 1 / 200 [ 0%] (Warmup)
#> Chain 2: Iteration: 20 / 200 [ 10%] (Warmup)
#> Chain 2: Iteration: 40 / 200 [ 20%] (Warmup)
#> Chain 2: Iteration: 60 / 200 [ 30%] (Warmup)
#> Chain 2: Iteration: 80 / 200 [ 40%] (Warmup)
#> Chain 2: Iteration: 100 / 200 [ 50%] (Warmup)
#> Chain 2: Iteration: 101 / 200 [ 50%] (Sampling)
#> Chain 2: Iteration: 120 / 200 [ 60%] (Sampling)
#> Chain 2: Iteration: 140 / 200 [ 70%] (Sampling)
#> Chain 2: Iteration: 160 / 200 [ 80%] (Sampling)
#> Chain 2: Iteration: 180 / 200 [ 90%] (Sampling)
#> Chain 2: Iteration: 200 / 200 [100%] (Sampling)
#> Chain 2:
#> Chain 2: Elapsed Time: 0.753 seconds (Warm-up)
#> Chain 2: 0.426 seconds (Sampling)
#> Chain 2: 1.179 seconds (Total)
#> Chain 2:
#> Warning: The largest R-hat is NA, indicating chains have not mixed.
#> Running the chains for more iterations may help. See
#> https://mc-stan.org/misc/warnings.html#r-hat
#> Warning: Bulk Effective Samples Size (ESS) is too low, indicating posterior means and medians may be unreliable.
#> Running the chains for more iterations may help. See
#> https://mc-stan.org/misc/warnings.html#bulk-ess
#> Warning: Tail Effective Samples Size (ESS) is too low, indicating posterior variances and tail quantiles may be unreliable.
#> Running the chains for more iterations may help. See
#> https://mc-stan.org/misc/warnings.html#tail-ess
#> Computing post-estimation metrics (including lvs if requested)...
summary(fit)
#> blavaan 0.5.6.1314 ended normally after 100 iterations
#>
#> Estimator BAYES
#> Optimization method MCMC
#> Number of model parameters 21
#>
#> Number of observations 301
#>
#> Statistic MargLogLik PPP
#> Value -3806.252 0.000
#>
#> Parameter Estimates:
#>
#>
#> Latent Variables:
#> Estimate Post.SD pi.lower pi.upper Rhat Prior
#> visual =~
#> x1 1.000
#> x2 0.572 0.121 0.359 0.805 1.003 normal(0,10)
#> x3 0.739 0.131 0.498 1.045 1.014 normal(0,10)
#> textual =~
#> x4 1.000
#> x5 1.116 0.060 0.995 1.231 1.000 normal(0,10)
#> x6 0.933 0.055 0.831 1.027 0.992 normal(0,10)
#> speed =~
#> x7 1.000
#> x8 1.267 0.178 0.905 1.615 1.005 normal(0,10)
#> x9 1.219 0.289 0.769 2.041 1.015 normal(0,10)
#>
#> Covariances:
#> Estimate Post.SD pi.lower pi.upper Rhat Prior
#> visual ~~
#> textual 0.386 0.079 0.225 0.556 0.996 lkj_corr(1)
#> speed 0.240 0.052 0.158 0.350 1.005 lkj_corr(1)
#> textual ~~
#> speed 0.159 0.051 0.065 0.266 0.995 lkj_corr(1)
#>
#> Variances:
#> Estimate Post.SD pi.lower pi.upper Rhat Prior
#> .x1 0.569 0.143 0.242 0.820 1.020 gamma(1,.5)[sd]
#> .x2 1.149 0.093 0.957 1.327 0.998 gamma(1,.5)[sd]
#> .x3 0.855 0.105 0.656 1.078 1.009 gamma(1,.5)[sd]
#> .x4 0.380 0.047 0.294 0.475 0.994 gamma(1,.5)[sd]
#> .x5 0.450 0.053 0.355 0.566 0.993 gamma(1,.5)[sd]
#> .x6 0.361 0.049 0.276 0.469 1.007 gamma(1,.5)[sd]
#> .x7 0.839 0.092 0.670 1.022 1.004 gamma(1,.5)[sd]
#> .x8 0.505 0.098 0.327 0.726 1.006 gamma(1,.5)[sd]
#> .x9 0.549 0.101 0.328 0.736 1.012 gamma(1,.5)[sd]
#> visual 0.796 0.164 0.533 1.171 1.009 gamma(1,.5)[sd]
#> textual 0.981 0.104 0.789 1.182 1.000 gamma(1,.5)[sd]
#> speed 0.342 0.097 0.154 0.525 1.038 gamma(1,.5)[sd]
#>