56 Matching Annotations
  1. Jun 2020
    1. the top and bottom 5% of all simulation estimates will be omitted

      this seems an unusual thing to do to me. Have you seen this before? Wouldn't it mean that standard errors, coverage etc are underestimated? I'd suggest instead using all data to calculate things like standard error, but show 5-95% intervals on the plots?

      Looking at the Results, and at your code, I think this removal of the top and bottom 5% could have quite unpredictable implications on bias, SE and coverage.

      Please can we see the results without the extremes dropped?

    2. Expected outco

      recall again here - these come from the 'true' models that match the DGMs. Also important to mention that this use of 'truth' doesn't (in my opinion) bias results in our favour since all methods are able to use this 'true' expected event proportion.

      However, you could argue that the 'true' IPCW is favouring our approach - you could mention this as a limitation in the Discussion: that for finite samples one of the simpler approaches might outperform IPCW because it doesn't need to estimate the IPCW

    3. To combat this, we re-scale the values produced to be with this range and perform the regression as normal.

      we really need to get guidance from Paul Lambert on this: have you managed to get hold of him?

    4. This is a minor issue and can be dealt with by most software packages

      Presumably this is no problem at all as long as there is a weighting option (as there is in glm())

    5. the value of the KM curve at the current time is taken to be the average Observed number of events

      suggest cutting this (it's not an accurate description of what the KM estimate is, so I suggest just say we use the KM estimate as the observed proportion of events)

    6. FPFPF_P was chosen

      make it very clear that you are assuming these models are known: not attempting to estimate them - and explain why this is the right thing to do (i.e. this part of the process is not of interest)

    7. a high number of patients was chosen to improve precision of our estimates

      as here we are interested in bias rather than variability of the estimates

    8. This combines to give a simulated survival function, SSS as S(t|Z=z)=exp(−eβZtη+1η+1)S(t|Z=z)=exp⁡(−eβZtη+1η+1) S(t|Z=z) = \exp\left(-\frac{e^{\beta Z}t^{\eta+1}}{\eta+1}\right) and a simulated censoring function, ScScS_c as Sc(t|Z=z)=exp(−eγZt)

      Suggest cutting: giving the hazard functions is enough (corresponding survival functions are obvious to derive).

    9. Survival times were simulated with a baseline hazard λ0(t)=tηλ0(t)=tη\lambda_0(t) = t^{\eta} (i.e. Weibull), and a proportional hazard of eβZeβZe^{\beta Z}.

      just write down the equation all together - feels a bit strange to separate out like this

    10. Each population was simulated with three parameters: ββ\beta, γγ\gamma and ηη\eta, which defined the proportional hazards coefficients for the survival and censoring distributions and the baseline hazard function, respectively.

      suggest moving a bit later - since I'm immediately wondering what the models are - so I'd present the models first then say these will be the parameters that we vary.

    11. that they are exchangeable conditional on the measured covariates

      to be precise, exchangeable conditional on the covariates in the model used to construct the IPCW, assuming that this model is correctly specified.

    12. Inverse Probability Weighting Adjustment of the Logistic Regression Calibration-in-the-Large

      This title needs changing: doesn't capture the contents at all. How about 'Using inverse probability of censoring weights to estimate calibration-in-the-large for time-to-event models'

  2. May 2020
    1. relaxes the assumption that patients who were censored are identical to those that remain at risk

      and replaces with the assumption that they are identical/exchangeable conditional on the measured covariates.

    2. In

      I know I wrote this(!), but this para is quite hard to follow - suggest moving the points that address the first two of the three ways to where they are introduced - then introduce the third way (censoring) and say that is our focus.

    3. In these papers a fractional polynomial approach to estimating the baseline survival function (and thus being able to share it efficiently) is also provided.

      move above to where we introduce the challenge of sharing baseline hazard.

    4. this coverage is reduced compared to the previous set of results shown (approximately 75% throughout)

      but why? I would expect the coverage to be ok

    1. out performing

      the models with different number of states are modelling different outcomes, and probably answering slightly different clinical questions. So I think clinical considerations should primarily inform which of 2, 3, 4 state model is used. Then a question about whether all three need to be presented in this paper

  3. Apr 2020
    1. We did not assess the viability of these models as it was believed this assumption to make our results more understandable.

      I think it is necessary to at least check the PH assumption. For example, there might be a strong non-proportionality across gender: then it would be entirely reasonable to fit separate models by gender (e.g. as QRISK does)

    2. This timelessness of the model means it can be applied to any patient at any time during their CKD journey

      Paper needs to explain more how you can indeed apply at any point in the journey. Presume that you mean only before any state transition has occurred?