Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Aug;81(4):668-697.
doi: 10.1177/0013164420970614. Epub 2020 Nov 16.

A Comparison of Label Switching Algorithms in the Context of Growth Mixture Models

Affiliations

A Comparison of Label Switching Algorithms in the Context of Growth Mixture Models

Kristina R Cassiday et al. Educ Psychol Meas. 2021 Aug.

Abstract

Simulation studies involving mixture models inevitably aggregate parameter estimates and other output across numerous replications. A primary issue that arises in these methodological investigations is label switching. The current study compares several label switching corrections that are commonly used when dealing with mixture models. A growth mixture model is used in this simulation study, and the design crosses three manipulated variables-number of latent classes, latent class probabilities, and class separation, yielding a total of 18 conditions. Within each of these conditions, the accuracy of a priori identifiability constraints, a priori training of the algorithm, and four post hoc algorithms developed by Tueller et al.; Cho; Stephens; and Rodriguez and Walker are tested to determine their classification accuracy. Findings reveal that, of all a priori methods, training of the algorithm leads to the most accurate classification under all conditions. In a case where an a priori algorithm is not selected, Rodriguez and Walker's algorithm is an excellent choice if interested specifically in aggregating class output without consideration as to whether the classes are accurately ordered. Using any of the post hoc algorithms tested yields improvement over baseline accuracy and is most effective under two-class models when class separation is high. This study found that if the class constraint algorithm was used a priori, it should be combined with a post hoc algorithm for accurate classification.

Keywords: growth mixture model; label switching; mixture modeling; simulation; training data set.

PubMed Disclaimer

Conflict of interest statement

Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Figures

Figure 1.
Figure 1.
(a) Two-class model classification accuracy for single algorithms. (b) Two-class model classification accuracy for combinations of algorithms. (c) Three-class model classification accuracy for single algorithms. (d) Three-class model classification accuracy for combinations of algorithms. (e) Four-class model classification accuracy for single algorithms. (f) Four-class model classification accuracy for combinations of algorithms.
Figure 2.
Figure 2.
(a) Two-class model average classification accuracy for individuals by condition and algorithm. (b) Three-class model average classification accuracy for individuals by condition and algorithm. (c) Four-class model average classification accuracy for individuals by condition and algorithm.

References

    1. Casella G., George E. I. (1992). Explaining the Gibbs sampler. The American Statistician, 46(3), 167-174. 10.1080/00031305.1992.10475878 - DOI
    1. Cho Y. (2013). The mixture distribution polytomous Rasch model used to account for response styles on rating scales: A simulation study of parameter recovery and classification accuracy. [Unpublished doctoral dissertation]. University of Maryland.
    1. Dempster A. P., Laird N. M., Rubin D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B, 39(1), 1-22. 10.1111/j.2517-6161.1977.tb01600.x - DOI
    1. Depaoli S. (2013). Mixture class recovery in GMM under varying degrees of class separation: Frequentist versus Bayesian estimation. Psychological Methods, 18(2), 186-219. 10.1037/a0031609 - DOI - PubMed
    1. Diebolt J., Robert C. P. (1994). Estimation of finite mixture distributions through Bayesian sampling. Journal of the Royal Statistical Society, Series B, 56, 363-375. 10.1111/j.2517-6161.1994.tb01985.x - DOI

LinkOut - more resources