Skip to main content
Log in

Advanced Psychometric Methods for Developing and Evaluating Cut-Point-Based Indicators

  • Published:
Child Indicators Research Aims and scope Submit manuscript

Abstract

In this study, we offer some advice on the use of empirical techniques to create categorical indicators from responses to multiple survey questions that ostensibly compose scales. We discuss the use of advanced psychometric methods, including confirmatory factor analysis for ordered categorical measures (CFA-OCM) and item response theory (IRT), to expand 12 previously published criteria for cut-point development and evaluation offered by Moore et al. (Indicators of Child Well-Being: The Promise for Positive Youth Development. Annals of the American Academy of Political and Social Science. Special Issue: Positive Development: Realizing the Potential of Youth. 591:125–145, 2004). We present an application of this advice by developing categorical cut-points for the social skills and behavior problems subscales of the U.S. National Survey of Children’s Health (NSCH) Social Competence Scale. We used data from the 2007 NSCH, a large cross-sectional, random-digit-dial telephone survey of a representative sample of the noninstitutionalized population of U.S. children 0–17 years of age. Parents of children aged 6 to 17 years (n = 63,364) responded to 4 questions about their children’s behavior problems and 4 questions about their children’s positive social skills. IRT analyses indicated that IRT model-based scores of +1.8 and −0.7 provided suitable high and low cut-points, respectively, for the behavior problems subscale. These cut-points had marginal reliabilities of 0.70 and 0.67 respectively, and corresponded to raw subscale scores of 13 and 8 IRT analyses indicated that an IRT model-based score of −1.8 provided a suitable cut-point for identifying children with low levels of positive social skills. This cut-point had a marginal reliability of 0.63 and corresponded to a raw subscale score of 13. For both subscales and cut-point sets, frequency distributions demonstrated that the cut-points identified sufficient proportions of children generally as well as within demographic subgroups. Our findings indicate that analysts can use raw subscale scores and cut-points of 13 and 8 to identify children with high, moderate, and low levels of behavior problems. Our results demonstrate that, for the positive social skills subscale, the subscale’s psychometric properties only support a single cut-point of 13 that identifies children with low levels of positive social skills. Our study highlights the iterative, empirically-based process that cut-point development should follow and shows the value of modern test theory-based psychometric methods to inform this process.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  • Blumberg, S., Carle, A., O’Connor, K., Moore, K., & Lippman, L. (2008). Social competence: Development of an indicator for children and adolescents. Child Indicators Research, 1(2), 176–197.

    Google Scholar 

  • Blumberg, S. J., Foster, E. B. et al. (2009). Design and operation of the national survey of children’s health, 2007. Vital and Health Statistics 1.

  • Bollen, K. (1989). Structural equations with latent variables. Wiley New York.

  • Bond, T., & Fox, C. (2007). Applying the Rasch model: Fundamental measurement in the human sciences. Lawrence: Erlbaum.

  • Carle, A. C. (2009). Fitting multilevel models in complex survey data with design weights: Recommendations. BMC Medical Research Methodology, 9(49).

  • Carle, A. C., & Chassin, L. (2004). Resilience in a community sample of children of alcoholics: Its prevalence and relation to internalizing symptomatology and positive affect. Journal of Applied Developmental Psychology, 25(5), 577–595.

    Article  Google Scholar 

  • Cheung, G., & Rensvold, R. (2002). Evaluating goodness-of-fit indexes for testing measurement invariance. Structural Equation Modeling: A Multidisciplinary Journal, 9(2), 233–255.

    Article  Google Scholar 

  • Coley, R. L., & Chase-Lansdale, P. L. (1998). Adolescent pregnancy and parenthood: Recent evidence and future directions. The American Psychologist, 53(2), 152–166.

    Article  Google Scholar 

  • Embretson, S., & Hershberger, S. (1999). The new rules of measurement: What every psychologist and educator should know. Lawrence: Erlbaum.

  • Embretson, S., & Reise, S. (2000). Item response theory for psychologists. Lawrence: Erlbaum.

  • Green, B. F., Bock, D., et al. (1984). Technical guidelines for assessing computerized adaptive tests. Journal of Educational Measurement, 21(4), 347–360.

    Article  Google Scholar 

  • Hair, E., Moore, K., Garrett, S., Kinukawa, A., Lippman, L., & Michelson, E. (2005) The parent-adolescent relationship scale. What Do Children Need to Flourish? 183–202.

  • Hambleton, R. K. (1985). Item response theory. Boston: Kluwer.

    Google Scholar 

  • Hu, L., & Bentler, P. M. (1998). Fit indices in covariance structure modeling: Sensitivity to underparameterized model misspecification. Psychological Methods, 3(4), 424–453.

    Article  Google Scholar 

  • Hu, L., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling, 6(1), 1–55.

    Article  Google Scholar 

  • Jones, R. N. (2003). Racial bias in the assessment of cognitive functioning of older adults. Aging & Mental Health, 7(2), 83–102.

    Article  Google Scholar 

  • Korn, E., & Graubard, B. (2003). Estimating variance components by using survey data. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 65(1), 175–190.

    Article  Google Scholar 

  • Masten, A. S., Coatsworth, J. D., et al. (1995). The structure and coherence of competence from childhood through adolescence. Child Development, 66(6), 1635–1659.

    Article  Google Scholar 

  • McDonald, R. P. (1999). Test theory: A unified treatment. Mahwah: Erlbaum.

    Google Scholar 

  • Millsap, R. E., & Kwok, O.-M. (2004). Evaluating the impact of partial factorial invariance on selection in two populations. Psychological Methods, 9(1), 93–115.

    Article  Google Scholar 

  • Moore, K. A. (1997). Indicators of children’s well-being. In R. M. Hauser (Ed.), Population indicators of prenatal and infant health. New York: Russell Sage Foundation.

    Google Scholar 

  • Moore, K. A., Lippman, L., & Brown, B. (2004). Indicators of child well-being: The promise for positive youth development. Annals of the American Academy of Political and Social Science. Special Issue: Positive Development: Realizing the Potential of Youth. 591, 125–145.

    Google Scholar 

  • Muraki, E., & Bock, R. D. (2003). PARSCALE: IRT based test scoring and item analysis. Chicago: Scientific Software International.

    Google Scholar 

  • Muthén, B. (1984). A general structural equation model with dichotomous, ordered categorical, and continuous latent variable indicators. Psychometrika, 49(1), 115–132.

    Article  Google Scholar 

  • Muthén, B. O. (1989). Latent variable modeling in heterogeneous populations. Psychometrika, 54(4), 557–585.

    Article  Google Scholar 

  • Muthén, L. K., & Muthén, B. O. (1998). Mplus user’s guide (4th ed.). Los Angeles: Muthén & Muthén.

    Google Scholar 

  • Muthén, B. O., Kao, C-f, et al. (1991). Instructionally sensitive psychometrics: Application of a new IRT-based detection technique to mathematics achievement test items. Journal of Educational Measurement, 28(1), 1–22.

    Article  Google Scholar 

  • Nunnally, J. C., & Berstein, I. (1994). Psychometric theory. New York: McGraw-Hill.

    Google Scholar 

  • Samejima, F. (1969). Estimation of latent ability using a response pattern of graded scores. Psychometrika Monograph Supplement, 34(4, Pt. 2), 100.

    Google Scholar 

  • StataCorp L. (2007). Stata statistical software: Release 10. College Station (TX): StataCorp LP.

  • Steiger, J. (1998). A note on multiple sample extensions of the RMSEA fit index. Structural Equation Modeling: A Multidisciplinary Journal, 5(4), 411–419.

    Article  Google Scholar 

  • Takane, Y., & de Leeuw, J. (1987). On the relationship between item response theory and factor analysis of discretized variables. Psychometrika, 52(3), 393–408.

    Article  Google Scholar 

Download references

Acknowledgements

We would like to thank the US Health Resources and Services Administration, Maternal and Child Health Bureau, for making the data publicly available. Adam Carle would also like to thank Tara J. Carle and Margaret Carle whose unending support and thoughtful comments make his work possible. This paper reflects the authors’ thoughts and does not necessarily represent the official views of the authors’ employers.

Conflicts of Interest

We declare that we have no competing interests.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Adam Christopher Carle.

Additional information

Research Support: The Centers for Disease Control and Prevention (contract #200-20090M-29486) supported Dr. Carle’s work on this paper.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Carle, A.C., Blumberg, S.J., Moore, K.A. et al. Advanced Psychometric Methods for Developing and Evaluating Cut-Point-Based Indicators. Child Ind Res 4, 101–126 (2011). https://doi.org/10.1007/s12187-010-9075-1

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12187-010-9075-1

Keywords

Navigation