Skip to main content
Top
Gepubliceerd in:

Open Access 22-08-2024

Effect of present state bias on minimal important change estimates: a simulation study

Auteurs: Berend Terluin, Piper Fromy, Andrew Trigg, Caroline B. Terwee, Jakob B. Bjorner

Gepubliceerd in: Quality of Life Research | Uitgave 11/2024

share
DELEN

Deel dit onderdeel of sectie (kopieer de link)

  • Optie A:
    Klik op de rechtermuisknop op de link en selecteer de optie “linkadres kopiëren”
  • Optie B:
    Deel de link per e-mail
insite
ZOEKEN

Abstract

Purpose

The minimal important change (MIC) in a patient-reported outcome measure is often estimated using patient-reported transition ratings as anchor. However, transition ratings are often more heavily weighted by the follow-up state than by the baseline state, a phenomenon known as “present state bias” (PSB). It is unknown if and how PSB affects the estimation of MICs using various methods.

Methods

We simulated 3240 samples in which the true MIC was simulated as the mean of individual MICs, and PSB was created by basing transition ratings on a “weighted change”, differentially weighting baseline and follow-up states. In each sample we estimated MICs based on the following methods: mean change (MC), receiver operating characteristic (ROC) analysis, predictive modeling (PM), adjusted predictive modeling (APM), longitudinal item response theory (LIRT), and longitudinal confirmatory factor analysis (LCFA). The latter two MICs were estimated with and without constraints on the transition item slope parameters (LIRT) or factor loadings (LCFA).

Results

PSB did not affect MIC estimates based on MC, ROC, and PM but these methods were biased by other factors. PSB caused imprecision in the MIC estimates based on APM, LIRT and LCFA with constraints, if the degree of PSB was substantial. However, the unconstrained LIRT- and LCFA-based MICs recovered the true MIC without bias and with high precision, independent of the degree of PSB.

Conclusion

We recommend the unconstrained LIRT- and LCFA-based MIC methods to estimate anchor-based MICs, irrespective of the degree of PSB. The APM-method is a feasible alternative if PSB is limited.
Opmerkingen

Supplementary Information

The online version contains supplementary material available at https://​doi.​org/​10.​1007/​s11136-024-03763-4.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

In the evaluation of clinical research and practice, patient-reported outcomes (PROs), such as pain or physical function, have acquired an increasingly important role [1]. Typically, questionnaires, referred to as patient-reported outcome measures (PROMs), are used to measure such PROs at multiple time points, to assess change. Because PROM change scores lack intrinsic meaning, the concept of minimal important change [MIC; also called minimal (clinically) important difference (MCID or MID), meaningful within‑individual change (MWIC), or meaningful change threshold (MCT)], was introduced to help researchers and clinicians to interpret the clinical significance of PROM change scores [2]. The MIC is the smallest within-individual PROM change score that patients consider meaningful or important. Given the assumption that patients have their own MIC thresholds, the MIC-value to be estimated in a group of patients can be conceptualized as the mean of individual MICs [35].1
Anchor-based methods to estimate an MIC use an external criterion (i.e., the anchor) as a measure of minimal importance. A frequently used anchor is the transition (or global impression of change) rating provided by patients at follow-up. A typical transition question is “Please choose the response that best describes the overall change in your <symptom/overall status/etc.> since you started taking the study medication. Much worse, A little worse, No change, A little better, Much better” [6].
Using a transition question as an anchor assumes that patients are able to assess their change between a previous state (i.e., the baseline state) and the present state (i.e., the follow-up state). This requires that patients remember their previous experience to compare it with their present experience. However, with the passage of time, human memory is fallible. The resulting “noise” can lead to transition ratings that are relatively stronger determined by the present state than by the previous state [7, 8]. This so-called “present state bias” (PSB) is, to a variable extent, likely to be common. Recently, a method was introduced to accurately estimate the degree of PSB in transition ratings [9]. Now that we are able to estimate the degree of PSB in real data, the question arises to what extent PSB affects MIC estimation. Therefore, we conducted a simulation study to assess the effect of PSB on anchor-based MICs, estimated using various methods.

Methods

Data simulation

PROM item response data

We used item response theory (IRT) to simulate multiple datasets with item responses at two time points, baseline (T1) and follow-up (T2). IRT entails a set of mathematical models describing the relationship between responses to an item of a PROM and the underlying unobserved construct (or “latent trait”, denoted by the Greek letter “theta”, \(\theta\)) that the PROM purports to measure [10]. “Theta” is also used to indicate the latent trait metric (or scale), typically with a mean of 0 and a standard deviation (SD) of 1 in estimated IRT-models. We used the graded response model (GRM) for polytomous items [11]. The GRM describes the probability of endorsing a particular item response category (or a higher response category of that item) as a function of the latent trait \(\theta\) and two item parameters, “location” and “slope”. The location parameter refers to the location on the latent trait where the probability of endorsing an item response category (or a higher category) is 50%, and the slope parameter refers to the slope of the item response function curve showing the probability of endorsing an item response option as a function of the underlying latent trait.
We simulated complete sets of item responses for 2000 “simulees” (i.e., simulated patients), based on a set of item parameters and simulated distributions of the latent traits at baseline (\({\theta }_{\text{T}1}\)) and follow-up (\({\theta }_{\text{T}2}\)). The difference between \({\theta }_{\text{T}1}\) and \({\theta }_{\text{T}2}\) is denoted “latent change” (i.e., \(\Delta \theta ={\theta }_{\text{T}2}-{\theta }_{\text{T}1}\)). We created a set of item parameters for ten items with four response options2 (see the Online Resource, Sect. 1). The latent trait and change variables were created as follows. For each simulated dataset, we created a latent trait variable for T1 (\({\theta }_{\text{T}1}\)) and a latent change variable (\(\Delta \theta\)) that was or was not correlated with \({\theta }_{\text{T}1}\). The latent trait variable at T2 (\({\theta }_{\text{T}2}\)) was then calculated by adding the \(\Delta \theta\) values to the \({\theta }_{\text{T}1}\) values.
The latent variable at T1 (\({\theta }_{\text{T}1}\)) was given an SD of 1 and a mean of − 1, 0 or 1, simulating baseline samples with relatively low, average or high mean levels of the trait.3 The latent change variable (\(\Delta \theta\)) was given an SD of 0.75, 1 or 1.25, simulating less or more variability in the latent change. The correlation between \({\theta }_{\text{T}1}\) and \(\Delta \theta\) was set at − 0.50 or 0. The mean of the latent change (\(\overline{\Delta \theta }\)) was chosen in such a way that the proportion improved (i.e., the proportion of simulees whose latent change exceeded their individual MIC—see below) was 0.2, 0.5 or 0.8.

Transition ratings

In the absence of PSB, transition ratings with 5 response options (1 = “Much worse”, 2 = “A little worse”, 3 = “No change”, 4 = “A little better”, 5 = “Much better”) were simulated as follows. First, we simulated 4 variables with means of − 1.5, − 0.5, 0.5 and 1.5 and an SD of 0.075, representing the thresholds between the transition response options on the \(\Delta \theta\) scale. The 3rd threshold (between “No change” and “A little better”) was considered to represent the individual MIC, and the mean value represented the MIC (in the \(\Delta \theta\) metric) to be estimated. Second, in each dataset, we simulated a “perceived change” variable by adding a random error variable to \(\Delta \theta\). The mean of the error variable was 0 and its SD was chosen as to create a reliability of the perceived change of 0.3 or 0.5.4 Then, the transition rating responses were derived by comparing each simulee’s perceived change with their individual thresholds. Simulees whose perceived change did not exceed the first threshold were given code “1” (“Much worse”), simulees whose perceived change exceeded the first but not the second threshold were given code “2” (“A little worse”), and so on. It should be noted that the reliability of the transition ratings, being a discretization of the perceived change, equaled the reliability of the perceived change (i.e., 0.3 or 0.5).

Present state bias

In the presence of PSB, transition ratings were simulated by creating a “weighted (latent) change” variable (\({\Delta \theta }_{w}\)) to replace the (unweighted) latent change (\(\Delta \theta\)) in the procedure to obtain transition ratings as described above. Whereas the unweighted latent change was \(\Delta \theta ={\theta }_{\text{T}2}-{\theta }_{\text{T}1}\), the following formula was used for the weighted latent change:
$$\begin{aligned} \Delta \theta_w & = q \times \left( {\theta_{{\text{T}}2} - \overline{{\theta_{{\text{T}}1} }}} \right) + \left( {1 - q} \right) \times \left( {\theta_{{\text{T}}2} - \theta_{{\text{T}}1} } \right)\quad {\text{or, equivalently:}} \\ \Delta \theta_w & = q \times \left( {\theta_{{\text{T}}2} - \overline{{\theta_{{\text{T}}1} }}} \right) + \left( {1 - q} \right) \times \Delta \theta \\ \end{aligned}$$
(1)
where \(q\) represents a variable of weights between 0 and 1, indicating the degree of PSB (0 = no PSB; 1 = complete PSB), and where \(\overline{{\theta }_{\text{T}1}}\) is the mean of the baseline state (\({\theta }_{\text{T}1}\)). Equation (1) indicates that \({\Delta \theta }_{w}\) consists of two components: a proportion \(q\) of the latent follow-up state (i.e., the present state) relative to the mean baseline state (i.e., \({\theta }_{\text{T}2}-\overline{{\theta }_{\text{T}1}}\)), and a complementary proportion \((1-q)\) of the true latent change (\(\Delta \theta\), i.e., \({\theta }_{\text{T}2}-{\theta }_{\text{T}1}\)). The mean of the baseline state (\(\overline{{\theta }_{\text{T}1}}\)) is included in the Equation for the following reason. Consider patients whose transition ratings show complete PSB and who, therefore, for rating their transition, solely rely on their present state (\({\theta }_{\text{T}2}\)) for comparison with their individual thresholds. Such patients “treat” their present state (\({\theta }_{\text{T}2}\)) as if it were their true change. However, because \({\theta }_{\text{T}2}\) itself represents a state (i.e., the end result of change) and not a change, evaluating \({\theta }_{\text{T}2}\) as if it represented change requires that people construct a reference point for their baseline state. This reference point construction is probably a vague, implicit and intuitive process. It is assumably not more than a (wild) guess about where they were at baseline, probably more or less below or above the point where they truly were at that time. If the latter assumption is true, i.e., if people are as likely to underestimate as they are to overestimate their baseline state, we may assume that the mean of the constructed individual reference points equals the mean of the true baseline state (i.e., \(\overline{{\theta }_{\text{T}1}}\)). This might be seen as an example of “wisdom of the crowd”, the phenomenon that the average of many individual estimates tends to approximate the truth [13].5
We simulated \(q\)-values as a variable with a mean and an SD, accounting for the idea that the degree of PSB probably varies across individual people. The mean \(q\)-values were varied between 0 and 1 in steps of 0.2. To obtain, for instance, a weighted change variable with a group-level (mean) PSB of 40%, variable \(q\) was simulated as a random variable with a mean of 0.4 and an SD of (0.5 − |(0.5 − 0.4)|)/4.6 Individual \(q\)-values < 0 or > 1 were constrained to 0 and 1 respectively. In this example, 40% PSB implies that the average weighted latent change consisted of 40% of \(({\theta }_{\text{T}2}-\overline{{\theta }_{\text{T}1}})\) and 60% of \(\Delta \theta\). Once \({\Delta \theta }_{w}\) was created, error was added to obtain the “perceived change” variable with the desired reliability values (0.3 or 0.5), and, subsequently, the transition ratings were created, as described above.
Table 1 summarizes the parameters that were varied across the simulated datasets. There were 648 unique combinations of parameters. Each combination was simulated five times, resulting in a total of 3240 simulated samples. Descriptive sample characteristics are summarized in Table 2.
Table 1
Parameters varying across the simulated samples
Parameter
Values
Explanation
Mean latent trait T1 (\(\overline{{\theta }_{\text{T}1}}\))
 − 1, 0, 1
\(\overline{{\theta }_{\text{T}1}}\) reflects the average level of the latent trait at T1
SD latent change (SD \(\Delta \theta\))
0.75, 1, 1.25
SD \(\Delta \theta\) reflects the variability in the latent change between T1 and T2 (\(\Delta \theta\))
Correlation between \({\theta }_{\text{T}1}\) and \(\Delta \theta\)
 − 0.50, 0.00
Values are correlation coefficients
Proportion improved
0.2, 0.5, 0.8
Proportion of simulees having a true latent change greater than their individual MIC
Reliability of the transition ratings
0.3, 0.5
Values are reliability coefficients
Present state bias
0, 0.2, 0.4, 0.6, 0.8, 1
Proportion of \({\theta }_{\text{T}2}-\overline{{\theta }_{\text{T}1}}\) that is included in the weighted latent change (\({\Delta \theta }_{w}\))
SD, Standard deviation; T1, Baseline; T2, Follow-up
Table 2
Sample characteristics across 3240 simulated samples
Characteristic
Mean (range)
Reliabilitya (T1)
0.86 (0.84, 0.88)
Mean T1 scale score
15.0 (8.4, 21.6)
SD T1 scale score
6.7 (6.2, 7.4)
Skewness T1 scale score
0.01 (− 0.83, 0.85)
Kurtosis T1 scale score
 − 0.37 (− 0.97, 0.17)
Floor effects T1 scale score
0.02 (0.00, 0.07)
Ceiling effects T1 scale score
0.02 (0.00, 0.08)
Mean T2 scale score
17.7 (5.7, 27.6)
SD T2 scale score
7.0 (3.4, 9.5)
Skewness T2 scale score
 − 0.39 (− 2.45, 1.29)
Kurtosis T2 scale score
0.13 (− 1.22, 7.42)
Floor effects T2 scale score
0.03 (0.00, 0.24)
Ceiling effects T2 scale score
0.09 (0.00, 0.44)
Mean scale change score
2.7 (− 4.2, 10.3)
SD scale change score
6.8 (4.8, 9.1)
Skewness scale change score
0.04 (− 0.44, 0.69)
Kurtosis scale change score
0.13 (− 0.69, 1.24)
Floor effects scale change score
0.00 (0.00, 0.00)
Ceiling effects scale change score
0.00 (0.00, 0.00)
SD, Standard deviation; T1, Baseline; T2, Follow-up
aCronbach’s alpha

Analysis

MIC methods

Mean change (MC) method
The MC method is the oldest anchor-based MIC method [2, 14]. It takes the mean change score of the minimally importantly changed subgroup as the MIC. In each simulated sample we calculated PROM scores (i.e., the sum of 10 item scores) at T1 and T2, and the PROM change score (T2 − T1). Then, we recorded the mean change score for the subgroup scoring “A little better” (transition rating “4”) as the mean change MIC (MICMC).
Receiver operating characteristic (ROC) method
The ROC method approaches MIC estimation as a diagnostic problem, seeking to classify improved and not-improved patients with the least amount of misclassification [1517]. In each simulated sample we performed ROC analysis using the dichotomized transition ratings (improved (codes 4–5) versus not-improved [codes 1–3)] as the “state variable” and the PROM change score as the “test variable” [18, 19]. The optimal cut-off point according to the Youden criterion (i.e., maximization of sensitivity and specificity) was recorded as the ROC-based MIC (MICROC).
Predictive modeling (PM) method
This method is based on logistic regression using the dichotomized transition ratings as the outcome and the PROM change score as the determinant [20]. The change score of interest is the one with a likelihood ratio of 1 (i.e., the change score that is equally likely to occur in the improved group as in the not-improved group). The predictive modeling MIC (MICPM) identifies about the same cut-off point as MICROC, but MICPM is much more precise [20].
Adjusted predictive modeling (APM) method
Because MICPM is biased by the proportion improved, an adjustment was developed [4] and later improved [21]. In each sample we estimated the reliability of the transition ratings according to the method described by Griffiths et al. [12], and calculated the adjusted predictive modeling MIC (MICAPM).
Longitudinal item response theory (LIRT) method
The LIRT-based MIC (MICLIRT) is based on an LIRT model of the PROM items, with the dichotomized transition item serving as an indicator of both time factors [22]. In each sample we performed LIRT analysis with the slope parameters (\({{\upalpha }}_{\text{TR}1}\) and \({{\upalpha }}_{\text{TR}2}\)) of the transition item being negatively constrained as recommended (i.e., \({{\upalpha }}_{\text{TR}1}=-{{\upalpha }}_{\text{TR}2}\)) [22]. MICLIRT was first estimated on the metric of the latent change, as the location parameter of the transition item on the latent change. MICLIRT in terms of the PROM change score was then calculated as the mean change in expected PROM score if a large sample (n = 500,000) of simulees, in \({\theta }_{\text{T}1}\) distribution similar to the simulated sample, would all improve by exactly one MIC in terms of the latent change [22]. Because constraining the slope parameters of the transition ratings (TR) causes model misfit in the presence of PSB, we also estimated MICLIRT after freeing the slope parameters. Here, MICLIRT was estimated based on \({{\upalpha }}_{\text{TR}2}\).
Longitudinal confirmatory factor analysis (LCFA) method
The LCFA-based MIC (MICLCFA) is based on an LCFA model for ordinal indicators with the PROM items at T1 and T2 loading on the latent factors at T1 and T2 respectively, and the dichotomized transition item loading on both time factors [23]. The factor loadings of the transition item (\({\uplambda }_{\text{TR}1}\) and \({\uplambda }_{\text{TR}2}\)) were negatively constrained (i.e., \({\uplambda }_{\text{TR}1}=-{\uplambda }_{\text{TR}2}\)). MICLCFA was first estimated on the metric of the latent change. MICLCFA in terms of the PROM change score was then calculated in the same way as for MICLIRT. Because constraining the factor loadings causes model misfit in the presence of PSB, we also estimated MICLCFA after freeing the TR factor loadings. Here, MICLCFA was estimated based on \({\uplambda }_{\text{TR}2}\).

The effect of PSB on the MIC estimates

The effect of PSB was evaluated by comparing the MIC estimates with the “true” MIC value (as simulated), expressing the effect in terms of bias and mean squared error (MSE). Bias represents the mean difference between the MIC estimates and the true MIC. Bias indicates the tendency to systematically over- or underestimate the true value. MSE represents the mean of the squared differences between the MIC estimates and the true value, and indicates both the bias and the (im)precision of the estimates.
Whereas we aimed to estimate the MIC in the metric of the simulated PROM, we simulated the (true) MIC in the metric of the latent change (the true latent MIC was set to 0.5 theta units). The true MIC in terms of the PROM change score was calculated for each simulated sample as the mean change in expected PROM score if a large sample (n = 500,000) of simulees, similar to the simulated sample, would all improve by exactly one MIC in terms of the latent change. This true MIC was then subtracted from the estimated MICs to obtain MIC residuals. Positive residuals indicated overestimation of the true MIC, whereas negative residuals indicated underestimation of the true MIC. To examine the effect of PSB on the MIC, the MIC residuals were plotted against the degree of PSB, and the MIC residuals were summarized in terms of bias, SD, and MSE.
Furthermore, to explore the effect of PSB on the MIC (residuals), relative to other sample characteristics, we performed multivariate linear regression, regressing the MIC residuals on the simulation parameters (see Table 1) and all of their interactions. Then, backward selection was applied to remove interactions and determinant terms explaining < 2% (adjusted R2) of the variance.

Software

All analyses were performed in the statistical program R, version 4.2.2 [24]. We used the mirt package, version 1.40 [25] to simulate the samples, estimate LIRT-based MICs in terms of the latent change, and to estimate MICs in terms of the PROM change score for the LIRT and LCFA methods. We used the packages pROC, version 1.18.4, for ROC analysis [26], lavaan, version 0.6–16, for LCFA [27], and rsimsum, version 0.11.3, for summarizing the simulation results [28]. The R-code is available in the Online Resource, Sect. 5.

Results

True MIC in terms of PROM change score

Across the simulated samples the true MIC (in terms of the PROM change score) varied between 2.56 and 3.41 (mean 3.01, SD 0.29). The true MIC appeared to be determined by the mean \({\theta }_{\text{T}1}\) (as simulated). If the baseline trait level was average (i.e., mean \({\theta }_{\text{T}1}\) = 0), the mean true MIC was 3.33 (SD 0.02), if the baseline trait level was relatively low (i.e., mean \({\theta }_{\text{T}1}\) = − 1), the mean true MIC was 3.07 (SD 0.03), and if the baseline trait level was relatively high (i.e., mean \({\theta }_{\text{T}1}\) = 1), the mean true MIC was 2.63 (SD 0.02). The true MIC varying with the baseline trait level relates to skewness of the PROM scores. The baseline PROM score was positively skewed (to the right) if the trait level was relatively low, whereas the baseline PROM score was negatively skewed (to the left) if the trait level was relatively high. Skewness implies that relatively more simulees sat at or near the extremes of the distribution, where the possibility to improve the PROM score was limited. Therefore, the mean PROM change score if all simulees improved 0.5 \(\Delta \theta\) on the latent trait, decreased if the mean baseline score was relatively low or relatively high. Moreover, in the high trait level samples the problem was aggravated by increased skewness of the follow-up score, whereas in the low trait level samples the problem was somewhat alleviated because the follow-up score was less skewed than the baseline score.

Effect of PSB on the MIC

Mean change MIC

Figure 1 provides an overview of the effect of PSB on the MIC estimates. On average, MICMC (Fig. 1, panel A) slightly overestimated the true MIC across all degrees of PSB. Table 3 confirms that MICMC was positively biased across all degrees of PSB, the bias being less than one point on a 30-points PROM scale. Given the Monte Carlo standard errors of the bias statistics, the bias was statistically significant [29]. Furthermore, Table 3 shows that the SD of the residuals and the MSE increased with increasing degrees of PSB, suggesting a decrease in the estimates’ precision.
Table 3
Performance measures of the MIC estimation by MIC method (Monte Carlo standard errors in parenthesis)
MIC method/statistic
Degree of present state bias
MICMC
0
0.2
0.4
0.6
0.8
1
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
0.92 (0.11)
0.96 (0.11)
0.90 (0.11)
0.79 (0.12)
0.63 (0.13)
0.49 (0.14)
 Standard deviation of residuals
2.49 (0.08)
2.49 (0.08)
2.66 (0.08)
2.81 (0.09)
3.03 (0.09)
3.30 (0.10)
 Mean squared error
7.02 (0.34)
7.12 (0.35)
7.87 (0.36)
8.51 (0.37)
9.57 (0.41)
11.09 (0.45)
MICROC
      
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
 − 0.59 (0.13)
 − 0.76 (0.13)
 − 0.97 (0.13)
 − 1.15 (0.14)
 − 1.32 (0.14)
 − 1.19 (0.15)
 Standard deviation of residuals
3.03 (0.09)
2.93 (0.09)
2.98 (0.09)
3.21 (0.10)
3.32 (0.10)
3.50 (0.11)
 Mean squared error
9.55 (0.44)
9.16 (0.41)
9.79 (0.44)
11.59 (0.51)
12.73 (0.53)
13.63 (0.65)
MICPM
      
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
 − 0.27 (0.12)
 − 0.27 (0.12)
 − 0.26 (0.13)
 − 0.25 (0.13)
 − 0.24 (0.14)
 − 0.24 (0.15)
 Standard deviation of residuals
2.90 (0.09)
2.89 (0.09)
2.98 (0.09)
3.09 (0.09)
3.24 (0.10)
3.40 (0.10)
 Mean squared error
8.45 (0.37)
8.41 (0.36)
8.91 (0.37)
9.58 (0.38)
10.51 (0.41)
11.57 (0.45)
MICAPM
      
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
 − 0.13 (0.02)
 − 0.12 (0.02)
 − 0.07 (0.03)
 − 0.06 (0.05)
 − 0.04 (0.07)
 − 0.05 (0.09)
 Standard deviation of residuals
0.51 (0.02)
0.52 (0.02)
0.69 (0.02)
1.08 (0.03)
1.58 (0.05)
2.15 (0.07)
 Mean squared error
0.27 (0.02)
0.29 (0.02)
0.48 (0.03)
1.16 (0.06)
2.50 (0.10)
4.64 (0.17)
MICLIRT, constrained
      
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
0.02 (0.01)
0.02 (0.02)
0.07 (0.04)
0.07 (0.06)
 − 0.00 (0.10)
0.03 (0.17)
 Standard deviation of residuals
0.33 (0.01)
0.46 (0.01)
0.83 (0.03)
1.38 (0.04)
2.23 (0.07)
3.96 (0.12)
 Mean squared error
0.11 (0.01)
0.21 (0.01)
0.69 (0.05)
1.90 (0.15)
4.96 (0.59)
15.63 (2.34)
MICLIRT, unconstrained
      
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
0.03 (0.01)
 − 0.04 (0.01)
 − 0.02 (0.01)
 − 0.03 (0.02)
 − 0.04 (0.01)
0.02 (0.02)
 Standard deviation of residuals
0.32 (0.01)
0.32 (0.01)
0.31 (0.01)
0.35 (0.01)
0.34 (0.01)
0.40 (0.01)
 Mean squared error
0.10 (0.01)
0.10 (0.01)
0.10 (0.01)
0.12 (0.01)
0.12 (0.01)
0.16 (0.01)
MICLCFA, constrained
      
 N simulations
540
540
540
540
540
536
 Mean residuals (bias)
0.04 (0.02)
0.14 (0.03)
0.29 (0.04)
0.41 (0.06)
0.53 (0.08)
0.62 (0.09)
 Standard deviation of residuals
0.35 (0.01)
0.59 (0.02)
0.97 (0.03)
1.43 (0.04)
1.82 (0.06)
2.17 (0.07)
 Mean squared error
0.12 (0.01)
0.37 (0.02)
1.03 (0.05)
2.20 (0.11)
3.58 (0.16)
5.08 (0.23)
MICLCFA, unconstrained
      
 N simulations
540
540
540
540
540
540
 Mean residuals (bias)
0.05 (0.01)
 − 0.01 (0.01)
0.02 (0.01)
0.01 (0.02)
0.01 (0.02)
0.07 (0.02)
 Standard deviation of residuals
0.34 (0.01)
0.34 (0.01)
0.33 (0.01)
0.36 (0.01)
0.35 (0.01)
0.41 (0.01)
 Mean squared error
0.12 (0.01)
0.12 (0.01)
0.11 (0.01)
0.13 (0.01)
0.12 (0.01)
0.18 (0.01)

ROC-based MIC

Figure 1 (panel B) and Table 3 show that, on average, MICROC slightly underestimated the true MIC, and that the estimate was also rather imprecise across all degrees of PSB

Predictive modeling MIC

Figure 1 (panel C) and Table 3 show that, on average, MICPM appeared to have minimal bias due to PSB, but considerable imprecision, which slightly increased with increasing PSB

Adjusted predictive modeling MIC

MICAPM appeared to have minimal bias due to PSB (Table 3). The precision of the estimate was relatively high up to PSB = 0.4, but the precision decreased with increasing PSB (Fig. 1, panel D).

LIRT-based MIC, constrained

MICLIRT showed practically no bias due to PSB, irrespective of the degree of PSB (Table 3 and Fig. 1, panel E). However, beyond PSB = 0.2, the MSE increased progressively, suggesting loss of precision of the estimate. Moreover, in the PSB > 0.6 situation, there was a sharp increase in “outlying” estimates, probably related to model misspecification. In four samples no estimates were returned due to non-convergence.

LIRT-based MIC, unconstrained

In contrast to the constrained MICLIRT, the unconstrained version appeared to recover the true MIC with practically no bias and high precision across all degrees of PSB (Fig. 1, panel F, and Table 3).

LCFA-based MIC, constrained

With increasing PSB, MICLCFA showed some minor increase in bias but a significant decrease in precision (increase in SD and MSE; Table 3 and Fig. 1, panel G).

LCFA-based MIC, unconstrained

Unlike the constrained MICLCFA, and alike the unconstrained MICLIRT, the unconstrained version of MICLCFA proved to recover the true MIC with practically no bias and high precision across all degrees of PSB (Fig. 1, panel H, and Table 3).

Effect of PSB relative to other sample characteristics

The results of the regression analyses are summarized in Table 4 (detailed results are provided in the Online Resource, Sect. 2). The unconstrained versions of MICLIRT and MICLCFA are not included in Table 4 because none of the sample characteristics (including PSB) affected these estimates. Most often, the MIC estimates were impacted by the proportion improved and \(\overline{{\theta }_{\text{T}1}}\). PSB affected the estimation of MICAPM and the constrained versions of MICLIRT and MICLCFA, mainly through interactions with other sample characteristics, but not the other MIC estimates. The reliability of the TR item did not impact the MIC estimates at all. Given the explained variances, the simulation parameters explained a considerable amount of variance (71%—92%) in the MIC estimates.
Table 4
Results of the regression analyses: simulation parameters that explained the variance in the MIC residuals, and explained variance (adjusted R2), by MIC method
MIC method
Simulation parameters
Explained variance (R2)
(\(\overline{{\theta }_{\text{T}1}}\)) (1)
SD \(\Delta \theta\) (2)
\({\theta }_{\text{T}1}\)\(\Delta \theta\) correlation (3)
Proportion improved (4)
TR reliability (5)
PSB (6)
MICMC
I: 1×4
  
M, I: 1×4
  
0.91
MICROC
I: 1×4
  
M, I: 1×4
  
0.74
MICPM
   
M
  
0.92
MICAPM
M, I: 1×3
 
I: 1×3
I: 4×6
 
M, I: 4×6
0.82
MICLIRT, constrained
 
I: 2×3×6, 2×3×4×6
I: 3×6, 2×3×6, 3×4×6, 2×3×4×6
I: 3×4×6, 2×3×4×6
 
I: 3×6, 2×3×6, 3×4×6, 2×3×4×6
0.71
MICLCFA, constrained
I: 1×6
I: 2×3×6, 2×3×4×6
I: 2×3×6, 3×4×6, 2×3×4×6
I: 4×6, 2×3×6, 3×4×6, 2×3×4×6
 
M, I: 1×6, 4×6, 2×3×6, 3×4×6, 2×3×4×6
0.83
Simulation parameters: see Table 1, M = main effect, I: interaction effect

Discussion

Effect of PSB on the MIC

PSB did not clearly impact MICMC, MICROC and MICPM but the estimates were nevertheless biased by other factors, notably the proportion improved and \(\overline{{\theta }_{\text{T}1}}\). MICAPM appeared to become less precise with increasing PSB, an effect that was mediated predominantly by the proportion improved. The constrained versions of MICLIRT and MICLCFA also clearly suffered from a decrease in precision with increasing PSB, an effect that was partly caused by model misspecification and partly mediated by the proportion improved and other sample characteristics. The constrained MICLCFA showed some progressive positive bias, up to 0.62 points, with increasing degrees of PSB, an effect that was real given the Monte Carlo standard errors (Table 3). We have no explanation why this occurred in the constrained MICLCFA and not in the constrained MICLIRT.
Our most remarkably finding was that the unconstrained versions of MICLIRT and MICLCFA appeared not to be affected by PSB at all (nor by any other sample characteristic).
To understand why the unconstrained versions of MICLIRT and MICLCFA recovered the true MIC even if PSB was complete, we need to consider the response to the transition item in the general “item factor analysis” framework [30]. We will elaborate the explanation for the LIRT framework [22], but the explanation is analogous for the LCFA framework [23, 31].
The response to the transition item as an indicator of the latent change can mathematically be described as a function \(f\left(\text{TR}\right)\)—typically a logit function in LIRT—of the latent states at T1 (\({\theta }_{\text{T}1}\)) and T2 (\({\theta }_{\text{T}2}\)):
$$f\left(\text{TR}\right)={{\upalpha }}_{\text{TR}1}\times{\theta }_{\text{T}1}+{{\upalpha }}_{\text{TR}2}\times{\theta }_{\text{T}2}+{\updelta }_{\text{TR}}$$
(2)
where \({{\upalpha }}_{\text{TR}1}\) and \({{\upalpha }}_{\text{TR}2}\) represent regression coefficients (denoted slope parameters in IRT), and \({\updelta }_{\text{TR}}\) represents the intercept [23].
In a previous study we found the following relationship between \(q\) (i.e., the degree of PSB) and the regression coefficients \({{\upalpha }}_{\text{TR}1}\) and \({{\upalpha }}_{\text{TR}2}\) (for a mathematical proof see the Online Resource, Sect. 3)7:
$$q= \frac{{{\upalpha }}_{\text{TR}1}}{{{\upalpha }}_{\text{TR}2}}+1$$
Because \({{\upalpha }}_{\text{TR}1}={{\upalpha }}_{\text{TR}2}\times(q-1)\), it follows that Eq. (2) can be rewritten as:
$$\begin{aligned} f\left( {{\text{TR}}} \right) & = {\upalpha }_{{\text{TR}}2} \times \left( {q - 1} \right) \times \theta_{{\text{T}}1} + {\upalpha }_{{\text{TR}}2} \times \theta_{{\text{T}}2} + {\updelta }_{{\text{TR}}} \quad {\text{and:}} \\ f\left( {{\text{TR}}} \right) & = {\upalpha }_{{\text{TR}}2} \times \left( {\theta_{{\text{T}}2} - \left( {1 - q} \right) \times \theta_{{\text{T}}1} } \right) + {\updelta }_{{\text{TR}}} \\ \end{aligned}$$
(3)
The expression \({\theta }_{\text{T}2}-\left(1-q\right)\times{\theta }_{\text{T}1}\), within Eq. (3), represents an alternative expression of the weighted change in Eq. (1) (for proof see the Online Resource, Sect. 4). If \(q=0\) the weighted change equals the true change (i.e., \({\theta }_{\text{T}2}-{\theta }_{\text{T}1}\)); if \(q=1\) the weighted change equals the present state (i.e., \({\theta }_{\text{T}2}\)). Equation (3) clearly indicates that the transition item is an indicator of the weighted change with slope parameter \({{\upalpha }}_{\text{TR}2}\) and intercept \({\updelta }_{\text{TR}}\), independent of the degree of PSB. This is why the MIC threshold on the latent variable underlying the transition item is correctly estimated using \({{\upalpha }}_{\text{TR}2}\) and \({\updelta }_{\text{TR}}\), irrespective of the degree of PSB (provided that \({{\upalpha }}_{\text{TR}1}\) and \({{\upalpha }}_{\text{TR}2}\) are not constrained).8
Importantly, if the degree of PSB is large, patients’ responses to the transition question may not reflect their true change, and the reported proportion improved may not agree with the true proportion improved. Nevertheless, using these (faulty) transition ratings, we are able to recover the correct MIC as if there had been no PSB.

Limitations

Simulation studies always require assumptions about how things play out in the real world. The simulation of something elusive as present state bias was quite a challenge. In particular, we made two key assumptions. First, we assumed that patients’ collective memory of their baseline state (the reference point in case of PSB) is unbiased, in the sense that patients do not collectively cherish a too favorable (or too unfavorable) memory of their baseline state. Second, we assumed that patients use the same personal thresholds to evaluate their perceived change, irrespective of the degree of PSB. If any of these assumptions are violated, the estimated MIC will be biased.9 Unfortunately, we just have to accept the two key assumptions mentioned above, because there is currently no way to test whether or not the assumptions are met in item response data. Moreover, without adopting these assumptions, it will be impossible to estimate MIC values using any of the methods examined. Perhaps, future qualitative research (cognitive interviewing) might shed more light on this matter.
Furthermore, we did not include response shift in the simulations and analyses, as we used invariant item parameters across the time points. The effect of response shift on MIC estimation needs further simulation research.

Conclusions and recommendations

PSB does not impact MICs estimated using the mean change method, the ROC method and the predictive modeling method. Yet, these methods can no longer be recommended because of bias due to the proportion improved (and baseline severity). The MIC based on adjusted predictive modeling can safely be used up until 40% PSB (i.e., PSB < 0.4). The MICs based on LIRT and LCFA provide unbiased and precise estimates, irrespective of the degree of PSB—but only if the transition item slope parameters (in LIRT) or factor loadings (in LCFA) are freely estimated. As the unconstrained MICs also perfectly recover the MIC if there is no PSB, we recommend not to apply these constraints anymore (see the Online Resource, Sect. 6, for the R-code for the recommended methods: adjusted predictive modeling (APM), unconstrained LIRT, and unconstrained LCFA).

Declarations

Conflict of interest

The authors have no relevant financial or non-financial interests to disclose.
Not applicable.

Ethical approval

Not applicable.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
share
DELEN

Deel dit onderdeel of sectie (kopieer de link)

  • Optie A:
    Klik op de rechtermuisknop op de link en selecteer de optie “linkadres kopiëren”
  • Optie B:
    Deel de link per e-mail

Onze productaanbevelingen

BSL Podotherapeut Totaal

Binnen de bundel kunt u gebruik maken van boeken, tijdschriften, e-learnings, web-tv's en uitlegvideo's. BSL Podotherapeut Totaal is overal toegankelijk; via uw PC, tablet of smartphone.

Bijlagen

Supplementary Information

Below is the link to the electronic supplementary material.
Voetnoten
1
Relevant change can occur in two directions, improvement or deterioration. For the sake of simplicity, we limit our account to improvement.
 
2
The choice for 10 items and 4 response options is arbitrary but inconsequential as to the principles of the simulation and analysis methods.
 
3
Note that higher trait levels represent “better” states, and positive latent change scores represent improvement.
 
4
The choice for 0.3 and 0.5 was based on a study of 4 empirical datasets, in which the reliability of the transition ratings appeared to vary between 0.27 and 0.48 [12].
 
5
In a fitted LIRT or LCFA model, \(\overline{{\theta }_{\text{T}1}}\) is typically set to 0, in which case \({\theta }_{\text{T}2}-\overline{{\theta }_{\text{T}1}}={\theta }_{\text{T}2}\), and the value zero (which equals \(\overline{{\theta }_{\text{T}1}}\)) becomes the natural baseline reference point. However, during the simulation procedure, we also used different values of \(\overline{{\theta }_{\text{T}1}}\) (i.e., − 1 and 1, because the item parameters were fixed) and we needed to take this into account when simulating \({\Delta \theta }_{w}\).
 
6
Simulating the SD of the q-variable to be (0.5 − |(0.5 − q)|)/4 ensures maximum variability of normally distributed individual q-values while most values fall within the 0–1 range.
 
7
We found that PSB could be quantified by \(\frac{{\uplambda }_{\text{TR}1}}{{\uplambda }_{\text{TR}2}}+1\) in an LCFA framework [9]. Under the specific specifications of the LCFA-model used for estimating the MIC, \(\frac{{\uplambda }_{\text{TR}1}}{{\uplambda }_{\text{TR}2}}\) is identical to \(\frac{{{\upalpha }}_{\text{TR}1}}{{{\upalpha }}_{\text{TR}2}}\) in the LIRT-model [23].
 
8
In the LIRT framework the MIC is estimated as \(-\frac{{\updelta }_{\text{TR}}}{{{\upalpha }}_{\text{TR}2}}\), whereas in the LCFA framework the MIC is estimated as \(\frac{{\uptau }_{\text{TR}}}{{\uplambda }_{\text{TR}2}}\), where \({\uptau }_{\text{TR}}\) represents the threshold of the transition item, and \({\uplambda }_{\text{TR}2}\) represents the factor loading of the transition item on the T2 factor.
 
9
Additional simulations revealed that, in the presence of PSB, patients’ systematic overestimation of their baseline state results in overestimation of the MIC whereas underestimation of the baseline state results in underestimation of the MIC. If patients would apply higher personal thresholds to evaluate their change in the presence of PSB than they would in the absence of it, the MIC will be overestimated. If patients would apply lower thresholds, the MIC will be underestimated.
 
Literatuur
1.
go back to reference FDA. (2023). Incorporating clinical outcome assessments into endpoints for regulatory decision-making. Silver Spring MD: US Food and Drug Administration. FDA. (2023). Incorporating clinical outcome assessments into endpoints for regulatory decision-making. Silver Spring MD: US Food and Drug Administration.
2.
go back to reference Jaeschke, R., Singer, J., & Guyatt, G. H. (1989). Measurement of health status. Ascertaining the minimal clinically important difference. Controlled Clinical Trials, 10, 407–415.CrossRefPubMed Jaeschke, R., Singer, J., & Guyatt, G. H. (1989). Measurement of health status. Ascertaining the minimal clinically important difference. Controlled Clinical Trials, 10, 407–415.CrossRefPubMed
3.
go back to reference King, M. T. (2011). A point of minimal important difference (MID): A critique of terminology and methods. Expert Review of Pharmacoeconomics and Outcomes Research, 11, 171–184.CrossRefPubMed King, M. T. (2011). A point of minimal important difference (MID): A critique of terminology and methods. Expert Review of Pharmacoeconomics and Outcomes Research, 11, 171–184.CrossRefPubMed
4.
go back to reference Terluin, B., Eekhout, I., & Terwee, C. B. (2017). The anchor-based minimal important change, based on receiver operating characteristic analysis or predictive modeling, may need to be adjusted for the proportion of improved patients. Journal of Clinical Epidemiology, 83, 90–100.CrossRefPubMed Terluin, B., Eekhout, I., & Terwee, C. B. (2017). The anchor-based minimal important change, based on receiver operating characteristic analysis or predictive modeling, may need to be adjusted for the proportion of improved patients. Journal of Clinical Epidemiology, 83, 90–100.CrossRefPubMed
5.
go back to reference Vanier, A., Sebille, V., Blanchin, M., & Hardouin, J. B. (2021). The minimal perceived change: A formal model of the responder definition according to the patient’s meaning of change for patient-reported outcome data analysis and interpretation. BMC Medical Research Methodology, 21(1), 128.CrossRefPubMedPubMedCentral Vanier, A., Sebille, V., Blanchin, M., & Hardouin, J. B. (2021). The minimal perceived change: A formal model of the responder definition according to the patient’s meaning of change for patient-reported outcome data analysis and interpretation. BMC Medical Research Methodology, 21(1), 128.CrossRefPubMedPubMedCentral
6.
go back to reference FDA. (2019). Incorporating clinical outcome assessments into endpoints for regulatory decision-making. Silver Spring MD: US Food and Drug Administration. FDA. (2019). Incorporating clinical outcome assessments into endpoints for regulatory decision-making. Silver Spring MD: US Food and Drug Administration.
7.
go back to reference Guyatt, G. H., Norman, G. R., Juniper, E. F., & Griffith, L. E. (2002). A critical look at transition ratings. Journal of Clinical Epidemiology, 55(9), 900–908.CrossRefPubMed Guyatt, G. H., Norman, G. R., Juniper, E. F., & Griffith, L. E. (2002). A critical look at transition ratings. Journal of Clinical Epidemiology, 55(9), 900–908.CrossRefPubMed
8.
go back to reference Wyrwich, K. W., & Tardino, V. M. (2006). Understanding global transition assessments. Quality of Life Research, 15(6), 995–1004.CrossRefPubMed Wyrwich, K. W., & Tardino, V. M. (2006). Understanding global transition assessments. Quality of Life Research, 15(6), 995–1004.CrossRefPubMed
9.
go back to reference Terluin, B., Griffiths, P., Trigg, A., Terwee, C. B., & Bjorner, J. B. (2022). Present state bias in transition ratings was accurately estimated in simulated and real data. Journal of Clinical Epidemiology, 143, 128–136.CrossRefPubMed Terluin, B., Griffiths, P., Trigg, A., Terwee, C. B., & Bjorner, J. B. (2022). Present state bias in transition ratings was accurately estimated in simulated and real data. Journal of Clinical Epidemiology, 143, 128–136.CrossRefPubMed
10.
go back to reference Embretson, S. E., & Reise, S. P. (2009). Item response theory for psychologists (2nd ed.). Lawrence Erlbaum. Embretson, S. E., & Reise, S. P. (2009). Item response theory for psychologists (2nd ed.). Lawrence Erlbaum.
11.
go back to reference Samejima, F. (1996). The graded response model. In W. J. van der Linden & R. Hambleton (Eds.), Handbook of modern item response theory (pp. 85–100). Springer. Samejima, F. (1996). The graded response model. In W. J. van der Linden & R. Hambleton (Eds.), Handbook of modern item response theory (pp. 85–100). Springer.
12.
go back to reference Griffiths, P., Terluin, B., Trigg, A., Schuller, W., & Bjorner, J. B. (2022). A confirmatory factor analysis approach was found to accurately estimate the reliability of transition ratings. Journal of Clinical Epidemiology, 141, 36–45.CrossRefPubMed Griffiths, P., Terluin, B., Trigg, A., Schuller, W., & Bjorner, J. B. (2022). A confirmatory factor analysis approach was found to accurately estimate the reliability of transition ratings. Journal of Clinical Epidemiology, 141, 36–45.CrossRefPubMed
13.
go back to reference Landemore, H. (2012). Collective wisdom—Old and new. In H. Landemore & J. Elster (Eds.), Collective wisdom: Principles and mechanisms. Cambridge University Press.CrossRef Landemore, H. (2012). Collective wisdom—Old and new. In H. Landemore & J. Elster (Eds.), Collective wisdom: Principles and mechanisms. Cambridge University Press.CrossRef
14.
go back to reference Hays, R. D., Farivar, S. S., & Liu, H. (2005). Approaches and recommendations for estimating minimally important differences for health-related quality of life measures. COPD, 2(1), 63–67.CrossRefPubMed Hays, R. D., Farivar, S. S., & Liu, H. (2005). Approaches and recommendations for estimating minimally important differences for health-related quality of life measures. COPD, 2(1), 63–67.CrossRefPubMed
15.
go back to reference Deyo, R. A., & Centor, R. M. (1986). Assessing the responsiveness of functional scales to clinical change: An analogy to diagnostic test performance. J Chron Dis, 39, 897–906.CrossRefPubMed Deyo, R. A., & Centor, R. M. (1986). Assessing the responsiveness of functional scales to clinical change: An analogy to diagnostic test performance. J Chron Dis, 39, 897–906.CrossRefPubMed
16.
go back to reference Stratford, P. W., Binkley, J. M., Riddle, D. L., & Guyatt, G. H. (1998). Sensitivity to change of the Roland–Morris Back Pain Questionnaire: Part 1. Physical Therapy, 78, 1186–1196.CrossRefPubMed Stratford, P. W., Binkley, J. M., Riddle, D. L., & Guyatt, G. H. (1998). Sensitivity to change of the Roland–Morris Back Pain Questionnaire: Part 1. Physical Therapy, 78, 1186–1196.CrossRefPubMed
17.
go back to reference de Vet, H. C. W., Ostelo, R. W. J. G., Terwee, C. B., van der Roer, N., Knol, D. L., Beckerman, H., Boers, M., & Bouter, L. M. (2007). Minimally important change determined by a visual method integrating an anchor-based and a distribution-based approach. Quality of Life Research, 16(1), 131–142.CrossRefPubMed de Vet, H. C. W., Ostelo, R. W. J. G., Terwee, C. B., van der Roer, N., Knol, D. L., Beckerman, H., Boers, M., & Bouter, L. M. (2007). Minimally important change determined by a visual method integrating an anchor-based and a distribution-based approach. Quality of Life Research, 16(1), 131–142.CrossRefPubMed
18.
go back to reference Greiner, M., Pfeiffer, D., & Smith, R. D. (2000). Principles and practical application of the receiver-operating characteristic analysis for diagnostic tests. Preventive Veterinary Medicine, 45(1–2), 23–41.CrossRefPubMed Greiner, M., Pfeiffer, D., & Smith, R. D. (2000). Principles and practical application of the receiver-operating characteristic analysis for diagnostic tests. Preventive Veterinary Medicine, 45(1–2), 23–41.CrossRefPubMed
19.
go back to reference Fawcett, T. (2006). An introduction to ROC analysis. Pattern Recognition Letters, 27(8), 861–874.CrossRef Fawcett, T. (2006). An introduction to ROC analysis. Pattern Recognition Letters, 27(8), 861–874.CrossRef
20.
go back to reference Terluin, B., Eekhout, I., Terwee, C. B., & de Vet, H. C. W. (2015). Minimal important change (MIC) based on a predictive modeling approach was more precise than MIC based on ROC analysis. Journal of Clinical Epidemiology, 68, 1388–1396.CrossRefPubMed Terluin, B., Eekhout, I., Terwee, C. B., & de Vet, H. C. W. (2015). Minimal important change (MIC) based on a predictive modeling approach was more precise than MIC based on ROC analysis. Journal of Clinical Epidemiology, 68, 1388–1396.CrossRefPubMed
21.
go back to reference Terluin, B., Eekhout, I., & Terwee, C. B. (2022). Improved adjusted minimal important change took reliability of transition ratings into account. Journal of Clinical Epidemiology, 148, 48–53.CrossRefPubMed Terluin, B., Eekhout, I., & Terwee, C. B. (2022). Improved adjusted minimal important change took reliability of transition ratings into account. Journal of Clinical Epidemiology, 148, 48–53.CrossRefPubMed
22.
go back to reference Bjorner, J. B., Terluin, B., Trigg, A., Hu, J., Brady, K. J. S., & Griffiths, P. (2023). Establishing thresholds for meaningful within-individual change using longitudinal item response theory. Quality of Life Research, 32(5), 1267–1276.CrossRefPubMed Bjorner, J. B., Terluin, B., Trigg, A., Hu, J., Brady, K. J. S., & Griffiths, P. (2023). Establishing thresholds for meaningful within-individual change using longitudinal item response theory. Quality of Life Research, 32(5), 1267–1276.CrossRefPubMed
23.
go back to reference Terluin, B., Trigg, A., Fromy, P., Schuller, W., Terwee, C. B., & Bjorner, J. B. (2023). Estimating anchor-based minimal important change using longitudinal confirmatory factor analysis. Quality of Life Research, 33, 963–973.CrossRefPubMed Terluin, B., Trigg, A., Fromy, P., Schuller, W., Terwee, C. B., & Bjorner, J. B. (2023). Estimating anchor-based minimal important change using longitudinal confirmatory factor analysis. Quality of Life Research, 33, 963–973.CrossRefPubMed
24.
go back to reference R Core Team. (2016). R: A language and environment for statistical computing. R Foundation for Statistical Computing. R Core Team. (2016). R: A language and environment for statistical computing. R Foundation for Statistical Computing.
25.
go back to reference Chalmers, R. P. (2012). mirt: A multidimensional item response theory package for the R environment. Journal of Statistical Software, 48(6), 1–29.CrossRef Chalmers, R. P. (2012). mirt: A multidimensional item response theory package for the R environment. Journal of Statistical Software, 48(6), 1–29.CrossRef
26.
go back to reference Robin, X., Turck, N., Hainard, A., Tiberti, N., Lisacek, F., Sanchez, J. C., & Müller, M. (2011). pROC: An open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics, 12, 77.CrossRefPubMedPubMedCentral Robin, X., Turck, N., Hainard, A., Tiberti, N., Lisacek, F., Sanchez, J. C., & Müller, M. (2011). pROC: An open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics, 12, 77.CrossRefPubMedPubMedCentral
27.
go back to reference Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48, 2.CrossRef Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48, 2.CrossRef
28.
go back to reference Gasparini, A. (2018). Rsimsum: Summarise results from Monte Carlo simulation studies. Journal of Open Source Software, 3(26), 739.CrossRef Gasparini, A. (2018). Rsimsum: Summarise results from Monte Carlo simulation studies. Journal of Open Source Software, 3(26), 739.CrossRef
29.
go back to reference Morris, T. P., White, I. R., & Crowther, M. J. (2019). Using simulation studies to evaluate statistical methods. Statistics in Medicine, 38(11), 2074–2102.CrossRefPubMedPubMedCentral Morris, T. P., White, I. R., & Crowther, M. J. (2019). Using simulation studies to evaluate statistical methods. Statistics in Medicine, 38(11), 2074–2102.CrossRefPubMedPubMedCentral
30.
31.
go back to reference Kamata, A., & Bauer, D. J. (2008). A note on the relation between factor analytic and item response theory. Structural Equation Modeling—A Multidisciplinary Journal, 15(1), 136–153.CrossRef Kamata, A., & Bauer, D. J. (2008). A note on the relation between factor analytic and item response theory. Structural Equation Modeling—A Multidisciplinary Journal, 15(1), 136–153.CrossRef
Metagegevens
Titel
Effect of present state bias on minimal important change estimates: a simulation study
Auteurs
Berend Terluin
Piper Fromy
Andrew Trigg
Caroline B. Terwee
Jakob B. Bjorner
Publicatiedatum
22-08-2024
Uitgeverij
Springer International Publishing
Gepubliceerd in
Quality of Life Research / Uitgave 11/2024
Print ISSN: 0962-9343
Elektronisch ISSN: 1573-2649
DOI
https://doi.org/10.1007/s11136-024-03763-4