首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 703 毫秒
1.
Parameter recovery was assessed within mixture confirmatory factor analysis across multiple estimator conditions under different simulated levels of mixture class separation. Mixture class separation was defined in the measurement model (through factor loadings) and the structural model (through factor variances). Maximum likelihood (ML) via the EM algorithm was compared to a Markov chain Monte Carlo (MCMC) estimator condition using weak priors and a condition using tight priors. Results indicated that the MCMC weak condition produced the highest bias, particularly with a weak Dirichlet prior for the mixture class proportions. Specifically, the weak Dirichlet prior affected parameter estimates under all mixture class separation conditions, even with moderate and large sample sizes. With little knowledge about parameters, ML/EM should be used over MCMC weak. However, MCMC tight produced the lowest bias under all mixture class separation conditions and should be used if tight and accurate priors can be placed on parameters.  相似文献   

2.
The recovery of weak factors has been extensively studied in the context of exploratory factor analysis. This article presents the results of a Monte Carlo simulation study of recovery of weak factor loadings in confirmatory factor analysis under conditions of estimation method (maximum likelihood vs. unweighted least squares), sample size, loading size, factor correlation, and model specification (correct vs. incorrect). The effects of these variables on goodness of fit and convergence are also examined. Results show that recovery of weak factor loadings, goodness of fit, and convergence are improved when factors are correlated and models are correctly specified. Additionally, unweighted least squares produces more convergent solutions and successfully recovers the weak factor loadings in some instances where maximum likelihood fails. The implications of these findings are discussed and compared to previous research.  相似文献   

3.
In testing the factorial invariance of a measure across groups, the groups are often of different sizes. Large imbalances in group size might affect the results of factorial invariance studies and lead to incorrect conclusions of invariance because the fit function in multiple-group factor analysis includes a weighting by group sample size. The implication is that violations of invariance might not be detected if the sample sizes of the 2 groups are severely unbalanced. In this study, we examined the effects of group size differences on results of factorial invariance tests, proposed a subsampling method to address unbalanced sample size issue in factorial invariance studies, and evaluated the proposed approach in various simulation conditions. Our findings confirm that violations of invariance might be masked in the case of severely unbalanced group size conditions and support the use of the proposed subsampling method to obtain accurate results for invariance studies.  相似文献   

4.
This simulation study assesses the statistical performance of two mathematically equivalent parameterizations for multitrait–multimethod data with interchangeable raters—a multilevel confirmatory factor analysis (CFA) and a classical CFA parameterization. The sample sizes of targets and raters, the factorial structure of the trait factors, and rater missingness are varied. The classical CFA approach yields a high proportion of improper solutions under conditions with small sample sizes and indicator-specific trait factors. In general, trait factor related parameters are more sensitive to bias than other types of parameters. For multilevel CFAs, there is a drastic bias in fit statistics under conditions with unidimensional trait factors on the between level, where root mean square error of approximation (RMSEA) and χ2 distributions reveal a downward bias, whereas the between standardized root mean square residual is biased upwards. In contrast, RMSEA and χ2 for classical CFA models are severely upwardly biased in conditions with a high number of raters and a small number of targets.  相似文献   

5.
The objective was to offer guidelines for applied researchers on how to weigh the consequences of errors made in evaluating measurement invariance (MI) on the assessment of factor mean differences. We conducted a simulation study to supplement the MI literature by focusing on choosing among analysis models with different number of between-group constraints imposed on loadings and intercepts of indicators. Data were generated with varying proportions, patterns, and magnitudes of differences in loadings and intercepts as well as factor mean differences and sample size. Based on the findings, we concluded that researchers who conduct MI analyses should recognize that relaxing as well as imposing constraints can affect Type I error rate, power, and bias of estimates in factor mean differences. In addition, fit indexes can be misleading in making decisions about constraints of loadings and intercepts. We offer suggestions for making MI decisions under uncertainty when assessing factor mean differences.  相似文献   

6.
This article examines the effects of clustering in latent class analysis. A comprehensive simulation study is conducted, which begins by specifying a true multilevel latent class model with varying within- and between-cluster sample sizes, varying latent class proportions, and varying intraclass correlations. These models are then estimated under the assumption of a single-level latent class model. The outcomes of interest are measures of bias in the Bayesian Information Criterion (BIC) and the entropy R 2 statistic relative to accounting for the multilevel structure of the data. The results indicate that the size of the intraclass correlation as well as between- and within-cluster sizes are the most prominent factors in determining the amount of bias in these outcome measures, with increasing intraclass correlations combined with small between-cluster sizes resulting in increased bias. Bias is particularly noticeable in the BIC. In addition, there is evidence that class separation interacts with the size of the intraclass correlations and cluster sizes in producing bias in these measures.  相似文献   

7.
Just as growth mixture models are useful with single-phase longitudinal data, multiphase growth mixture models can be used with multiple-phase longitudinal data. One of the practically important issues in single- and multiphase growth mixture models is the sample size requirements for accurate estimation. In a Monte Carlo simulation study, the sample sizes required for using these models are investigated under various theoretical and realistic conditions. In particular, the relationship between the sample size requirement and the number of indicator variables is examined, because the number of indicators can be relatively easily controlled by researchers in many multiphase data collection settings such as ecological momentary assessment. The findings not only provide tangible information about required sample sizes under various conditions to help researchers, but they also increase understanding of sample size requirements in single- and multiphase growth mixture models.  相似文献   

8.
基于联系数的直觉模糊数多指标评价研究   总被引:1,自引:0,他引:1  
集对分析的联系数理论是关于确定性与不确定性的理论,直觉模糊数多指标评价问题中具有不确定性,为了预测和分析不确定性的影响,可以运用集对分析理论,将直觉模糊数转化为联系数,从而建立基于联系数的直觉模糊数多指标评价模型.实例应用表明,方法有效、实用.  相似文献   

9.
This study presents a new approach to synthesizing differential item functioning (DIF) effect size: First, using correlation matrices from each study, we perform a multigroup confirmatory factor analysis (MGCFA) that examines measurement invariance of a test item between two subgroups (i.e., focal and reference groups). Then we synthesize, across the studies, the differences in the estimated factor loadings between the two subgroups, resulting in a meta-analytic summary of the MGCFA effect sizes (MGCFA-ES). The performance of this new approach was examined using a Monte Carlo simulation, where we created 108 conditions by four factors: (1) three levels of item difficulty, (2) four magnitudes of DIF, (3) three levels of sample size, and (4) three types of correlation matrix (tetrachoric, adjusted Pearson, and Pearson). Results indicate that when MGCFA is fitted to tetrachoric correlation matrices, the meta-analytic summary of the MGCFA-ES performed best in terms of bias and mean square error values, 95% confidence interval coverages, empirical standard errors, Type I error rates, and statistical power; and reasonably well with adjusted Pearson correlation matrices. In addition, when tetrachoric correlation matrices are used, a meta-analytic summary of the MGCFA-ES performed well, particularly, under the condition that a high difficulty item with a large DIF was administered to a large sample size. Our result offers an option for synthesizing the magnitude of DIF on a flagged item across studies in practice.  相似文献   

10.
This study examined the performance of the weighted root mean square residual (WRMR) through a simulation study using confirmatory factor analysis with ordinal data. Values and cut scores for the WRMR were examined, along with a comparison of its performance relative to commonly cited fit indexes. The findings showed that WRMR illustrated worse fit when sample size increased or model misspecification increased. Lower (i.e., better) values of WRMR were observed when nonnormal data were present, there were lower loadings, and when few categories were analyzed. WRMR generally illustrated expected patterns of relations to other well-known fit indexes. In general, a cutoff value of 1.0 appeared to work adequately under the tested conditions and the WRMR values of “good fit” were generally in agreement with other indexes. Users are cautioned that when the fitted model is misspeficifed, the index might provide misleading results under situations where extremely large sample sizes are used.  相似文献   

11.
Survival analysis is an advanced statistical method to investigate the occurrence and the timing of an important event such as school access, dropout, and graduation in a longitudinal framework. The aim of our study is to provide practical guidelines for empirical researchers in choosing an appropriate survival analysis model. For this goal, this study chose two major survival analytical models of a discrete-time hazard model and a Cox regression model and compared analytical outcomes considering time metrics, as well as sample sizes and censoring proportions. In the analytical model, the combined specifications of varying factors using two models were adopted to analyse the Educational Longitudinal Study of 2002. We chose the college access of Hispanic English Language Learners to understand the importance of adopting a proper survival model to examine the educational outcome in the educational context. Importantly, we considered the hazard probability for the target event in the model specification which is a fundamental yet often neglected component of survival analysis. We recommended discrete models for the cases with a smaller number of time points, larger time metrics, larger sample size, and smaller proportions of censored observations.  相似文献   

12.
Abstract

This paper and the accompanying tool are intended to complement existing supports for conducting power analysis tools by offering a tool based on the framework of Minimum Detectable Effect Sizes (MDES) formulae that can be used in determining sample size requirements and in estimating minimum detectable effect sizes for a range of individual- and group-random assignment design studies and for common quasi-experimental design studies. The paper and accompanying tool cover computation of minimum detectable effect sizes under the following study designs: individual random assignment designs, hierarchical random assignment designs (2-4 levels), block random assignment designs (2-4 levels), regression discontinuity designs (6 types), and short interrupted time-series designs. In each case, the discussion and accompanying tool consider the key factors associated with statistical power and minimum detectable effect sizes, including the level at which treatment occurs and the statistical models (e.g., fixed effect and random effect) used in the analysis. The tool also includes a module that estimates for one and two level random assignment design studies the minimum sample sizes required in order for studies to attain user-defined minimum detectable effect sizes.  相似文献   

13.
Recently, advancements in Bayesian structural equation modeling (SEM), particularly software developments, have allowed researchers to more easily employ it in data analysis. With the potential for greater use, come opportunities to apply Bayesian SEM in a wider array of situations, including for small sample size problems. Effective use of Bayseian estimation hinges on selection of appropriate prior distributions for model parameters. Researchers have suggested that informative priors may be useful with small samples, presuming that the mean of the prior is accurate with respect to the population mean. The purpose of this simulation study was to examine model parameter estimation for the Multiple Indicator Multiple Cause model when an informative prior distribution had an incorrect mean. Results demonstrated that the use of incorrect informative priors with somewhat larger variance than is typical, yields more accurate parameter estimates than do naïve priors, or maximum likelihood estimation. Implications for practice are discussed.  相似文献   

14.
Probabilistic analysis of linear elastic cracked structures   总被引:1,自引:0,他引:1  
This paper presents a probabilistic methodology for linear fracture mechanics analysis of cracked structures. The main focus is on probabilistic aspect related to the nature of crack in material. The methodology involves finite element analysis; sta- tistical models for uncertainty in material properties, crack size, fracture toughness and loads; and standard reliability methods for evaluating probabilistic characteristics of linear elastic fracture parameter. The uncertainty in the crack size can have a significant effect on the probability of failure, particularly when the crack size has a large coefficient of variation. Numerical example is presented to show that probabilistic methodology based on Monte Carlo simulation provides accurate estimates of failure prob- ability for use in linear elastic fracture mechanics.  相似文献   

15.
This study compared diagonal weighted least squares robust estimation techniques available in 2 popular statistical programs: diagonal weighted least squares (DWLS; LISREL version 8.80) and weighted least squares–mean (WLSM) and weighted least squares—mean and variance adjusted (WLSMV; Mplus version 6.11). A 20-item confirmatory factor analysis was estimated using item-level ordered categorical data. Three different nonnormality conditions were applied to 2- to 7-category data with sample sizes of 200, 400, and 800. Convergence problems were seen with nonnormal data when DWLS was used with few categories. Both DWLS and WLSMV produced accurate parameter estimates; however, bias in standard errors of parameter estimates was extreme for select conditions when nonnormal data were present. The robust estimators generally reported acceptable model–data fit, unless few categories were used with nonnormal data at smaller sample sizes; WLSMV yielded better fit than WLSM for most indices.  相似文献   

16.
Peer assessment (PA) is one of the central principles of formative assessment and assessment for learning (AfL) fields. There is ample empirical evidence as to the benefits for students’ learning when AfL principles are implemented. However, teachers play a critical role in mediating the implementation of intended policies. Hence, their experiences, beliefs, and attitudes towards PA are important factors in determining whether the policy is actually carried out. A survey of over 1500 primary, secondary, and higher education teachers in Spain elicited their beliefs and values around PA as well as other aspects of formative assessment; only 751 teachers provided complete responses to all PA items. Teachers reported occasional use of PA in their classrooms but with positive experience of it. The vast majority did not use anonymous forms of PA and half of the teachers considered the students were accurate when assessing peers. Confirmatory factor analysis and structural equation modeling were used to examine relationships of attitudes and beliefs to self-reported frequency of using of PA. The self-reported frequency of using PA was strongly predicted by teacher experience of PA which included positive reasons for using PA, rather than negative obstacles for avoiding, prior use, and beliefs that students should participate in assessment, and willingness to include PA in grading.  相似文献   

17.
In the presence of test speededness, the parameter estimates of item response theory models can be poorly estimated due to conditional dependencies among items, particularly for end‐of‐test items (i.e., speeded items). This article conducted a systematic comparison of five‐item calibration procedures—a two‐parameter logistic (2PL) model, a one‐dimensional mixture model, a two‐step strategy (a combination of the one‐dimensional mixture and the 2PL), a two‐dimensional mixture model, and a hybrid model‐–by examining how sample size, percentage of speeded examinees, percentage of missing responses, and way of scoring missing responses (incorrect vs. omitted) affect the item parameter estimation in speeded tests. For nonspeeded items, all five procedures showed similar results in recovering item parameters. For speeded items, the one‐dimensional mixture model, the two‐step strategy, and the two‐dimensional mixture model provided largely similar results and performed better than the 2PL model and the hybrid model in calibrating slope parameters. However, those three procedures performed similarly to the hybrid model in estimating intercept parameters. As expected, the 2PL model did not appear to be as accurate as the other models in recovering item parameters, especially when there were large numbers of examinees showing speededness and a high percentage of missing responses with incorrect scoring. Real data analysis further described the similarities and differences between the five procedures.  相似文献   

18.
The asymptotically distribution free (ADF) method is often used to estimate parameters or test models without a normal distribution assumption on variables, both in covariance structure analysis and in correlation structure analysis. However, little has been done to study the differences in behaviors of the ADF method in covariance versus correlation structure analysis. The behaviors of 3 test statistics frequently used to evaluate structural equation models with nonnormally distributed variables, χ2 test TAGLS and its small-sample variants TYB and TF(AGLS) were compared. Results showed that the ADF method in correlation structure analysis with test statistic TAGLS performs much better at small sample sizes than the corresponding test for covariance structures. In contrast, test statistics TYB and TF(AGLS) under the same conditions generally perform better with covariance structures than with correlation structures. It is proposed that excessively large and variable condition numbers of weight matrices are a cause of poor behavior of ADF test statistics in small samples, and results showed that these condition numbers are systematically increased with substantial increase in variance as sample size decreases. Implications for research and practice are discussed.  相似文献   

19.
To infer longitudinal relationships among latent factors, traditional analyses assume that the measurement model is invariant across measurement occasions. Alternative to placing cross-occasion equality constraints on parameters, approximate measurement invariance (MI) can be analyzed by specifying informative priors on parameter differences between occasions. This study evaluated the estimation of structural coefficients in multiple-indicator autoregressive cross-lagged models under various conditions of approximate MI using Bayesian structural equation modeling. Design factors included factor structures, conditions of non-invariance, sizes of structural coefficients, and sample sizes. Models were analyzed using two sets of small-variance priors on select model parameters. Results showed that autoregressive coefficient estimates were more accurate for the mixed pattern than the decreasing pattern of non-invariance. When a model included cross-loadings, an interaction was found between the cross-lagged estimates and the non-invariance conditions. Implications of findings and future research directions are discussed.  相似文献   

20.
A study was conducted to determine the performance effects of ability grouping and perceived probability of attaining a prize in a Paired-associate (PA) task. Sixty undergraduate students were pretested for PA ability and divided into high, medium, and low ability groups for random assignment to control (no prize probability), low prize probability, and high prize probability treatment groups. The dependent measure was the number of correct responses over four trials for intentional learning and number of correct responses on a single test for the incidental learning condition. Results of two-way ANOVA supported the hypothesis that performance in a PA task is affected by subject's ability and perceived probability of attaining a prize.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号