Skip to main content

Advertisement

Log in

The Effect of Choosing Key Versus Ordinary Schools on Student’s Mathematical Achievement in China

  • Regular Article
  • Published:
The Asia-Pacific Education Researcher Aims and scope Submit manuscript

Abstract

The current study examined the impacts on students’ cognitive performance of the key versus ordinary school system in China, using an analytic approach that combines hierarchical linear modeling with propensity score stratification. The results show that students from key schools score significantly higher on a mathematical achievement test than their counterparts in ordinary schools, after controlling for student characteristics and their family background. The specific magnitude of the school effect varies substantially across the geographic locations of the school. The advantages of key schools over ordinary schools are found to be generally greater among urban schools compared with suburban schools. The results are noteworthy as both key and ordinary schools are state-funded and the system was formed directly by policy initiatives and differential resource allocation. As such, they bear important policy implications on systematic-level school management in general.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Notes

  1. Missing data analyses showed that no systematic differences were found using the intent-to-treat sample and the available-case sample; this is possibly due to the fact that the missing rate is very low (about 0.2 %).

  2. Because there is no solid theory in school effect research, we have no way to ensure that the observed across-sector differences are clearly pre-treatment differences, rather than consequences of school effect. However, were some of the differences caused positively by school effect, the magnitude of the school effect would be reduced by balancing differences of such variables across sectors. Consequently, by including such variables in the model, conservative estimates of school effect would be obtained. If a significant school effect could still be found under this condition, our confidence in the existence of school effect is stronger as the actual magnitude of such an effect would like be larger.

  3. Although it is arguable that students’ problem solving scores are in fact affected by school effects as our data are cross-sectional, two points can be made here to justify the use of such scores as a covariate: First, PISA items on problem solving are designed to assess student’s ability to solve complex systems without resort to much subject-matter material. The essential construct measured in this test is students’ planning and logical reasoning ability which are generally considered as the core of intelligence and the modifiability of such components by school experiences are suspected. Second, as mentioned in note 2, the inclusion of problem solving scores in the model would lead to conservative estimates of school effect, were such scores positively affected by key schools. Therefore, if little school effect was reduced after adding this covariate, we could increase our confidence that our estimates of school effects are robust.

References

  • Alexander, K. L., & Palls, A. M. (1985). School sector and cognitive performance: When is a little is a little? Sociology of Education, 58, 115–128.

    Article  Google Scholar 

  • Bidwell, C., & Dreeben, R. (2006). Public and private education: Conceptualizing the distinction. In M. Hallinan (Ed.), School sector and student outcomes (pp. 9–38). Notre Dame, IN: University of Notre Dame Press.

    Google Scholar 

  • Bradley, S., & Taylor, J. (1998). The effect of school size on exam performance in secondary schools. Oxford Bulletin of Economics and Statistics, 60, 291–325.

    Article  Google Scholar 

  • Brookover, W. B., Beady, D., Flood, P., Schweitzer, J., & Wisenbaker, J. (1979). School social systems and student achievement: Schools can make a difference. New York, NY: Praeger.

    Google Scholar 

  • Bryk, A., Lee, V., & Holland, P. (1993). Catholic Schools and the common Good. Cambridge, MA: Harvard University Press.

    Google Scholar 

  • Carbonaro, W. (2006). Public-private differences in achievement among Kindergarten students: Differences in learning opportunities and student outcomes. American Journal of Education, 113, 31–65.

    Article  Google Scholar 

  • Carbonaro, W., & Covay, E. (2010). School sector and student achievement in the era of standard based reforms. Sociology of Education, 83, 160–182.

    Article  Google Scholar 

  • Cohen, J. (1988). Statistical power analysis for the behavioral sciences (3rd edn.). New York: Academic Press.

  • Coleman, J., Campbell, E., Hobson, C., McPartland, J., Mood, A., Weinfeld, F., et al. (1966). Equality of educational opportunity. Washington, DC: U.S. Department of Health, Education and Welfare.

    Google Scholar 

  • Coleman, J., & Hoffer, T. (1987). Private and public schools. New York, NY: Basic Books.

    Google Scholar 

  • Coleman, J., Hoffer, T., & Kilgore, S. (1982). Cognitive outcomes in public and private schools. Sociology of Education, 55, 65–76.

    Article  Google Scholar 

  • D’Agostino, R. B, Jr. (1998). Propensity score methods for bias reduction in the comparison of a treatment to a non-randomized control group. Statistics in Medicine, 17, 2265–2281.

    Article  Google Scholar 

  • Dearden, L., Ferri, J., & Meghir, C. (2002). The effect of school quality on educational attainment and wages. Review of Economics and Statistics, 84, 1–20.

    Article  Google Scholar 

  • Desforges, C. and Abouchaar, A. (2003). The impact of parental involvement, parental support and family education on pupil achievement and adjustment: A literature review. Report Number 433, Department of Education and Skills.

  • Edmonds, R. R. (1979). Effective schools for the urban poor. Educational Leadership, 37, 15–24.

    Google Scholar 

  • Fang, X. C. (2000). Should we stop the key school system? In Q. R. Zhong, Z. Y. Jin, & G. P. Wu (Eds.), Understanding education in China (pp. 318–323). Bejing, China: Educational Science Press.

    Google Scholar 

  • Feinstein, L., & Symons, J. (2001). Attainment in secondary school. Oxford Economics Papers, 51, 300–321.

    Article  Google Scholar 

  • Feng, D. (2007). School effectiveness and improvement research in mainland China. In T. Townsend (Ed.), International handbook of school effectiveness and improvement (pp. 287–306). Dordrecht, The Netherlands: Springer.

    Chapter  Google Scholar 

  • Graddy, K., & Stevens, M. (2005). The impact of school resources on student performance: A study of private schools in the United Kingdom. Industrial and Labor Relations Review, 58, 435–451.

    Google Scholar 

  • Greeley, A. M. (1982). Catholic high schools and minority students. New Brunswick, NJ: Transaction.

    Google Scholar 

  • Grolnick, W. S., Benjet, C., Kurowski, C. O., & Apostoleris, N. H. (1997). Predictors of parent involvement in children’s schooling. Journal of Educational Psychology, 89, 538–548.

    Article  Google Scholar 

  • Hallinan, M. (2006). School sector and student outcomes. Notre Dame, IN: University of Notre Dame Press.

    Google Scholar 

  • He, X. L. (2010). Key school: From ideation to common practice. Educational Science Research, 6, 20–23.

    Google Scholar 

  • Hoffer, T., Greeley, A., & Colemen, J. (1985). Achievement growth in Catholic and public schools. Sociology of Education, 54, 74–97.

    Article  Google Scholar 

  • Holland, P. W. (1986). Statistics and causal inference. Journal of the American Statistics Association, 81, 945–970.

    Article  Google Scholar 

  • Hong, G., & Raudenbush, S. W. (2005). Effects of Kindergarten retention policy on children’s cognitive growth in reading and mathematics. Educational Evaluation and Policy Analysis, 27, 205–224.

    Article  Google Scholar 

  • Imbens, G., & Rubin, D. B. (1997). Bayesian inference for causal effects in randomized experiments with noncompliance. Annals of Statistics, 25, 305–327.

    Article  Google Scholar 

  • Jencks, C. (1985). How much do high school student learn? Sociology of Education, 58, 128–135.

    Article  Google Scholar 

  • Jin, S. H. (2000). Educational system for the Elite and the concept of key school. Educational Research and Experiment, 4, 18–21.

    Google Scholar 

  • Li, L. (2003). Talking about education: Interview with Li Langqing. Bejing: People’s Education Press.

    Google Scholar 

  • Liu, J. (2005). Success from initiatives and efforts. In L. Ma (Ed.), A wise conversation with sixty principals (pp. 8–14). Shanghai: Shanghai Educational Publishing House.

    Google Scholar 

  • Lubienski, S., & Lubienski, C. (2006). School sector and academic achievement: A multilevel analysis of NAEP data. American Education Research Journal, 43, 651–700.

    Article  Google Scholar 

  • Ma, X. Q., Peng, W. R., & Thomas, S. (2006). Value-added research on school effect: An empirical research on secondary schools in Baoding of the Heibei province. Educational Research, 10, 74–84.

    Google Scholar 

  • McNabb, R., Sarmistha, P., & Sloane, P. (1998). Gender differences in student attainment: The case of university students in the UK. Mineo: University of Cardiff.

    Google Scholar 

  • Meredith, P., & Flashman, J. (2007). How did the statewide assessment and accountability policies of the 1990s affect instructional quality in low income schools. In A. Garmoran (Ed.), Standards-based reform and the poverty gap (pp. 45–88). Washington, DC: Brookings Institution.

    Google Scholar 

  • Ministry of Education (MOE). (1993). The ongoing reform of China’s education. Beijing: Higher Education Press.

    Google Scholar 

  • Ministry of Education (MOE). (1998). Reinforcing the development of disadvantaged schools and making every elementary and middle school works in large and medium cities. Retrieved from http://www.bjsupervision.gov.cn/zcfg/. Accessed 25 Dec 2011.

  • Ministry of Education (MOE). (2005). Further promoting the even development in nine-year compulsory education. Beijing: China Education Daily.

    Google Scholar 

  • Morgan, S. L. (2001). Counterfactuals, causal effect heterogeneity, and the Catholic school effect on learning. Sociology of Education, 74, 341–374.

    Article  Google Scholar 

  • Morgan, S. L., & Sørensen, A. (1999). Parental networks, social closure and mathematics learning: A test of Coleman’s social capital explanation of school effects. American Sociological Review, 64, 661–681.

    Article  Google Scholar 

  • Morgan, S. L., & Winship, C. (2007). Counterfactuals and causal inference: Methods and principles for social research. New York, NY: Cambridge University Press.

    Book  Google Scholar 

  • Mortimore, P. (1998). The road to improvement: Reflections on school effectiveness. Lisse: Swets & Zeitlinger Publisher.

    Google Scholar 

  • Mortimore, P., Sammons, P., Stoll, L., Lewis, D., & Econb, R. (1988). School matters: The junior years. Somerset, England: Open Books.

    Google Scholar 

  • No Child Left Behind (NCLB) Act of 2001, Pub. L. No. 107-110, § 115, Stat. 1425 (2002).

  • OECD. (2009). PISA 2009 assessment framework—Key competencies in reading mathematics and science. Paris: OECD.

    Google Scholar 

  • Raudenbush, S. W., & Bryk, A. S. (2002). Hierarchical linear models: Applications and data analysis methods (2nd ed.). Thousand Oaks, CA: Sage Publication.

    Google Scholar 

  • Raudenbush, S. W., & Willms, J. D. (1995). The estimation of school effects. Journal of Educational and Behavioral Statistics, 20, 307–335.

    Article  Google Scholar 

  • Reardon, S., Cheadle, J., & Robinson, J. (2009). The effect of Catholic schooling on math and reading development in Kindergarten through fifth grade. Journal of Research on Educational Effectiveness, 2, 45–87.

    Article  Google Scholar 

  • Rosenbaum, P. R. & Rubin, D. B. (1983). The central role  of the propensity score in observational studies for causal effects. Biometrics, 70, 41 -55.

  • Rosebaum, P. R., & Rubin, D. B. (1984). Reducing bias in observational studies using subclassification on the propensity score. Journal of the American Statistical Association, 79, 516–524.

    Article  Google Scholar 

  • Rosebaum, P. R., & Rubin, D. B. (1985). Constructing a control group using multivariate matched sampling methods. American Statistician, 39, 33–38.

    Google Scholar 

  • Rowan, B., Raudenbush, S. W., & Kang, S. J. (1991). Organizational design in high schools: A multilevel analysis. American Journal of Education, 99, 238–266.

    Article  Google Scholar 

  • Rubin, D. B. (1974). Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology, 66, 688–701.

    Article  Google Scholar 

  • Rubin, D. B. (1986). Comment: Which ifs have causal answers. Journal of the American Statistical Association, 81, 961–962.

    Google Scholar 

  • Rubin, D. B. (2000). The utility of counterfactuals for causal inference. Journal of the American Statistical Association, 95, 435–438.

    Google Scholar 

  • Rutter, M, Maughan, B., Mortimore, P., & Ouston, J. (1979). Fifteen thousand hours: Secondary schools and their effects on children. London: Open Books.

  • Smith, J., & Naylor, R. A. (2001). Determinants of degree performance in UK universities: A statistical analysis of the 1993 student cohort. Oxford Bulletin of Economics and Statistics, 63, 29–60.

    Google Scholar 

  • Sørensen, A. B., & Morgan, S. L. (2000). School effects: Theoretical and methodological issues. In M. T. Hallinan (Ed.), Handbook of the sociology of education (pp. 137–160). New York, NY: Kluwer Academic/Plenum Publishers.

    Google Scholar 

  • Teddlie, C., & Stringfield, S. (2007). A history of school effectiveness and improvement research in the USA focusing on the past quarter century. In T. Townsend (Ed.), International handbook of school effectiveness and improvement (pp. 131–166). Dordrecht, The Netherlands: Springer.

    Chapter  Google Scholar 

  • Thomas, S. (2013). Personal communication. Bristol: University of Bristol.

    Google Scholar 

  • Walford, G. (2013). The development of private and state schools in England. In A. Gürlevik, C. Palentien, & R. Heyer (Eds.), Privatschulen versus Staatliche Schulen (pp. 89–102). New York, NY: Springer.

    Chapter  Google Scholar 

  • Wang, J. M. (2009). Theory and practices of value-added educational assessment. Unpublished master thesis, Jiangxi Normal University, Nanchang, Jiangxi.

  • Weber, G. (1971). Inner-city children can be taught to read: Four successful schools. Washington, DC: Council for Basic Education.

    Google Scholar 

  • Yang, D. P. (2003). The ideal and reality of educational equity in China. Beijing, China: Beijing University Press.

    Google Scholar 

  • Yuan, Z. G. (1999). Change of educational policy in China: A case study on the equity and quality in key schools. Guangzhou, China: Guangdong Educational Press.

    Google Scholar 

  • Zagoumennov, I. (1999). School effectiveness and school improvement in the Republic of Belarus. In T. Townsend, P. Clarke, & M. Ainscow (Eds.), Third millennium schools: A world of difference in school effectiveness and school improvement (pp. 89–106). Lisse: Swets and Zeitlinger.

    Google Scholar 

  • Zhang, D. (2004). Providing good education for every student. Beijing: Education and Science Publishing House.

    Google Scholar 

Download references

Acknowledgments

This article is supported financially by the MOE of China under Grant Nos. 2009JJD880011 and 2011JJD880030 of the National Key Research Center in Social Science and Humanities. The opinions expressed here do not necessarily reflect those of the funding agency.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Youqun Ren.

Appendices

Appendix 1: Propensity Score Estimation and Stratification

Using all 20 covariates in Table 2, an iterative approach was taken to estimate each student’s propensity score via logistic models. Within each iteration, students were divided into five strata according to their propensity scores. The balance of each covariate across school sectors was examined within each stratum by Cochran–Mantel–Haenszel (CMH) χ 2 test and bar chart examination. Higher-order term of a covariate and/or interaction terms with other important covariates would be added into the initial model, if a large value of the CMH statistic was obtained after propensity score stratification. This process continues until all covariates are balanced across school sectors.

It took three iterations before the final propensity scores were obtained. The final model contained 32 terms, including four interaction terms, six quadratic terms, one cubic term, and one quartic term. Figure 2 displays the distributions of the final estimated propensity scores for key and ordinary school students on a logit scale. The two distributions overlap substantially, which implies that for most students in key schools, there is a comparable student in ordinary schools in the sense of having similar probability of getting into key schools. However, a few students in key schools have propensity scores higher than any of the ordinary school students. Similarly, a few ordinary school students have propensity scores lower than any of the key school students. For these students, there are no students in the other sector with comparable propensity scores. To take these students into account, students were classified into seven strata based on their estimated propensity scores, in which the first (and the seventh) stratum contains exclusively these students with highest (and lowest) propensity scores, and the other five strata contain key or ordinary school students with comparable propensity scores.

Fig. 2
figure 2

Distributions of estimated propensities of getting into key schools for students in different school types

The last column of Table 2 shows the results of balance check for each covariate after the stratification. As expected, there is considerably greater balance on the observed covariates after propensity score stratification. None of the 20 covariates now exhibit significant differences across school sectors after stratification. To show this is indeed the case, Fig. 3 depicts the balance within each stratum for covariate Mother’s educational level, the only variable that needs to add more higher-order terms after the second iteration. Compare to across-sector differences before stratification, greater balances are achieved within each stratum of the propensity scores.

Fig. 3
figure 3

Balance check before and after propensity score stratification for Mother’s Educational Level (MEL)

Similar process was taken to obtain refined propensity score after adding students’ problem solving scores as a covariate. Figure 4 displays the balance check after stratifying students on the re-estimated propensity scores. As expected, great balances are achieved for almost all strata.

Fig. 4
figure 4

Balance check for problem solving scores

Appendix 2: Hierarchical Linear Models Used for School Effect Estimation

After stratifying students based on their propensity scores, within-stratum school effect between key and ordinary schools were estimated using the following model

$$Y_{ij} = \gamma_{00} + \gamma_{01} {\text{Sector}}_{j} + u_{j} + e_{ij} ;\,u_{j} \sim \left( {0,\tau } \right);\,e_{ij} \sim \left( {0,\sigma^{2} } \right),$$
(1)

where \(Y_{ij}\) is the mathematics score for student \(i\) in school \(j\), \({\text{Sector}}_{j}\) is the school type (0 = ordinary school, 1 = key school), and \(\gamma_{01}\) indicates school effect on \(Y_{ij}\). In the random part of the model, \(u_{j}\) and \(e_{ij}\) are residuals at school and student level, respectively. Both follow normal distributions with mean 0 and variances \(\tau\) and \(\sigma^{2}\), respectively.

The combined school effect across strata was estimated using the following model

$$Y_{ij} = \gamma_{00} + \gamma_{01} {\text{Sector}}_{j} + \sum\limits_{h = 1}^{4} {\gamma_{0(1 + h)} {\text{Strata}}_{hij} } + u_{j} + e_{ij} ;\,u_{j} \sim \left( {0,\tau } \right);\,\,e_{ij} \sim \left( {0,\sigma^{2} } \right),$$
(2)

where \({\text{Strata}}_{hij}\) are a series of dummy variables created to indicate four of the five propensity score strata and were added to model 1 to control for selection bias.

Before stratification, school effect estimation with regression adjustment for propensity score was estimated using the following model

$$Y_{ij} = \gamma_{00} + \gamma_{01} {\text{Sector}}_{j} + \gamma_{02} {\text{Logit}}P_{ij} + e_{ij} ;\,e_{ij} \sim \left( {0,\sigma^{2} } \right),$$
(3)

where \({\text{Logit}}P_{ij}\) is the logit of propensity score for student \(i\) in school \(j\).

After stratification, Eq. 3 was used to estimate within-stratum school effects with regression adjustment for propensity score. For the combined school effect across strata, estimate with regression adjustment for propensity score was obtained by adding \({\text{Logit}}P_{ij}\) to model 2.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Yang, X., Ke, Z., Zhan, Y. et al. The Effect of Choosing Key Versus Ordinary Schools on Student’s Mathematical Achievement in China. Asia-Pacific Edu Res 23, 523–536 (2014). https://doi.org/10.1007/s40299-013-0126-5

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s40299-013-0126-5

Keywords

Navigation