Skip to main content

External validation of predictive models of sexual, urinary, bowel and hormonal function after surgery in prostate cancer subjects

Abstract

Background

In 2020, a research group published five linear longitudinal models, predict Expanded Prostate Cancer Index Composite-26 (EPIC-26) scores post-treatment for radical prostatectomy, external beam radiotherapy and active surveillance collectively in US patients with localized prostate cancer.

Methods

Our study externally validates the five prediction models for patient reported outcomes post-surgery for localised prostate cancer. The models’ calibration, fit, variance explained and discrimination (concordance-indices) were assessed. Two Australian validation cohorts 1 and 2 years post-prostatectomy were constructed, consisting of 669 and 439 subjects, respectively (750 in total). Patient reported function in five domains post-prostatectomy: sexual, bowel, hormonal, urinary incontinence and other urinary dysfunction (irritation/obstruction). Domain function was assessed using the EPIC-26 questionnaire.

Results

1 year post-surgery, R2 was highest for the sexual domain (35%, SD = 0.02), lower for the bowel (21%, SD = 0.03) and hormone (15%, SD = 0.03) domains, and close to zero for urinary incontinence (1%, SD = 0.01) and irritation/obstruction (− 5%, SD = 0.04). Calibration slopes for these five models were 1.04 (SD = 0.04), 0.84 (SD = 0.06), 0.85 (SD = 0.06), 1.16 (SD = 0.13) and 0.45 (SD = 0.04), respectively. Calibration-in-the-large values were − 2.2 (SD = 0.6), 2.1 (SD = 0.01), 5.1 (SD = 0.1), 9.6 (SD = 0.9) and 4.0 (SD = 0.2), respectively. Concordance-indices were 0.73, 0.70, 0.70, 0.58 and 0.62, respectively (all had SD = 0.01). Mean absolute error and root mean square error were similar across the validation and development cohorts. The validation measures were largely similar at 2 years post-surgery.

Conclusions

The sexual, bowel and hormone domain models validated well and show promise for accurately predicting patient reported outcomes in a non-US surgical population. The urinary domain models validated poorly and may require recalibration or revision.

Peer Review reports

Introduction

Prostate cancer is the most common cancer globally in males with an estimated 16,741 new cases in Australia in 2020 [1, 2]. Prostatectomy is a standard of care treatment for localized prostate cancer [3]. However, prostate cancer often progresses slowly without noticeable symptoms [2]; patients with localized prostate cancer commonly die from other causes [4]. This complicates the rationale for treating prostate cancer. Treatment choices should incorporate both patient and clinical factors and avoid overtreatment of prostate cancer [4]. Prostatectomy and its associated adverse events on quality of life, including erectile dysfunction and urinary incontinence, may outweigh the benefit of treatment [4]. Hence patient reported outcomes measures (PROMs) may be equally as important as clinical outcomes such as survival. PROMs can provide a more holistic interpretation of treatment benefit and improve clinical decision making, patient-doctor communication, and patient outcomes [5,6,7].

Formal prediction models can assist clinicians in accurately predicting PROMs post-prostatectomy to determine the potential treatment benefit of prostatectomy [8]. Laviana et al. recently developed five publicly available online models to predict PROMs after three potential treatment options in pre-treatment US patients with clinical localized stage T1-T2 prostate cancer [9, 10]. The treatment options were radical prostatectomy (RP), external beam radiation therapy (EBRT) and active surveillance (AS); they are analysed collectively. The models reported good calibration and model fit [9]. External validation is important to determine if the model is useful for predicting outcomes in populations similar to, but different from, the development population and if they can be applied individually to each specific treatment [11]. We were also concerned that one treatment (RP) is historically associated with higher rates of sexual dysfunction and incontinence, and another (AS) has potentially very little incidence [12]. Given that the three therapies were analysed collectively and not individually, this study externally validates the online tool and determines its utility at 1- and 2-years post-surgery.

Materials/subjects and methods

Development models

Laviana et al. developed five longitudinal linear regression models post-treatment for localized prostate cancer [9]. Each model estimated one EPIC-26 domain score: sexual function, urinary incontinence, urinary irritation/obstruction, bowel function or hormonal function. Scores ranged from 0 to 100 where higher numbers indicate better outcomes [13]. Laviana et al. excluded the last sexual domain question (utilizing five questions instead of six); we validate the 5-question sexual domain model in the main analysis and the 6-question model as a supplementary analysis. Predictor variables were time post-treatment, primary treatment choice (RP, AS, and EBRT), age at diagnosis, race, baseline domain score, overall health, prostate specific antigen (PSA) and Gleason score. Overall health was assessed with the SF-36 Health Survey question “In general, would you say your health is Poor, Fair, Good, Very good or Excellent [14].” The exclusion criteria were receiving androgen deprivation therapy, brachytherapy or cryoablation, age > 80, PSA ≥ 50 ng/dL and non-localized prostate cancer. Laviana et al. analysed all three primary treatments choices collectively in the same models.

Validation cohort

The South Australian Prostate Cancer Clinical Outcomes Collaborative (SA-PCCOC) database is a population-based registry currently recruiting over 90% of newly diagnosed prostate cancer cases in the state of South Australia, including over 19,000 men [15]. Established in 1998, the registry collects data from collaborating public and private institutions and clinicians [15, 16]. PROMs are collected before treatment commences and at 3, 6, 12, 24 and 60 months post-primary treatment. Participants were mailed paper copies of the Expanded Prostate Cancer Index Composite-26 (EPIC-26) questionnaires and a reply-paid envelope for response.

We sought to verify the Laviana et al. models specifically for RP and followed their inclusion criteria in accordance with the transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) checklist (Supplementary Table 1) [17]. De-identified data was obtained from SA-PCCOC for post-prostatectomy patients diagnosed between January 2007 and March 2020. The registry does not record participant race; hence we assumed all subjects were Caucasian, which likely matched both the majority of SA-PCCOC patients and 74% of subjects in Laviana’s development cohort [9]. Patients were included only if they completed both a pre- and post-surgery EPIC-26 questionnaire. Post-surgery assessments were required to be within 8–16 or 20–28 months post-prostatectomy, defining a 1- and a 2-year validation cohort, respectively. If a patient completed multiple assessments per time range, only the closest to the target time was included. The 1- and 2-year survey time points are those where SA-PCCOC has the largest quantity of complete data.

SA-PCCOC data are not publicly available but can be obtained after establishing a Data Use Agreement with SA-PCCOC, which has permission to authorize data accessibility from the Southern Australian Clinical Human Research Ethics Committee (SALHN HREC). The data collected by SA-PCCOC has been approved by the SALHN HREC (protocol 307.14).

Statistical methods

Model predictive values were assessed for calibration – the agreement between the observed endpoints and predictions (calibration-in-the-large (CL) and calibration slope (CS)) – predictive accuracy (mean absolute error (MAE) and root mean square error (RMSE)), proportion of variance explained (R2) and discrimination – the effectiveness at predicting which scores would be better or worse (concordance-index, abbreviated as c-index) [11, 18]. CL was estimated as the intercept in a regression with the predicted EPIC-26 score as an offset, and CS was the estimated slope of the predicted score in a univariable regression. Theoretically perfect CL and CS values are 0 and 1, respectively. Because a large proportion of the EPIC-26 domain scores were close to either the upper (100) or lower (0) boundaries, the two calibration metrics were calculated using quantile (median) regression, with linear (mean) regression reported as a supplementary analysis. Predicted versus observed EPIC-26 scores are shown graphically locally weighted smoothing regressions superimposed over scatter plots. R2 is calculated as one minus the ratio of residual and total variance i.e. \(1-\frac{\sum_{\textrm{i}=1}^{\textrm{n}}{\left({\textrm{y}}_{\textrm{i}}-{\hat{\textrm{y}}}_{\textrm{i}}\right)}^2.}{\sum_{\textrm{i}=1}^{\textrm{n}}{\left({\textrm{y}}_{\textrm{i}}-\overline{\textrm{y}}\right)}^2}\) [18]. For predictive models, R2 can be negative [18]. MAE and RMSE are defined as MAE = \(\frac{\sum_{\textrm{i}=1}^{\textrm{n}}\left|{\hat{\textrm{y}}}_{\textrm{i}}-{\textrm{y}}_{\textrm{i}}\right|}{\textrm{n}}\) and RMSE = \(\sqrt{\frac{\sum_{\textrm{i}=1}^{\textrm{n}}{\left({\hat{\textrm{y}}}_{\textrm{i}}-{\textrm{y}}_{\textrm{i}}\right)}^2}{\textrm{n}}}\). The c-index is equal to the probability that, for two randomly selected subjects after surgery, the subject with a higher predicted score has a higher observed score instead of a lower score i.e. \(\text{P}\left(\left({\widehat{\text{y}}}_\text{i}-{\widehat{\text{y}}}_\text{j}\right)\left({\text{y}}_\text{i}-{\text{y}}_\text{j}\right)>0\vert{\widehat{\text{y}}}_\text{i}\neq{\widehat{\text{y}}}_\text{j}\right)\) [19]. CL and discrimination were reported for the first time in these prediction models. The registry database had several patients with unreported pre-surgery PSA or general health assessment; hence multiple imputation using chained equations was performed (100 datasets generated with 100 iterations each) [20]. Missing EPIC-26 domain scores were not imputed. For each validation statistic, means (standard deviations) and median (interquartile ranges) are reported across the 100 imputed datasets. Analyses were performed in R version 4.0.5 using the packages mice and quantreg [21,22,23].

Results

Validation cohorts

Between January 2007 and March 2020, the registry recorded 1848 patients having prostatectomy for localized prostate cancer, 886 of which competed both a pre- and post-surgery EPIC-26 questionnaire. 750 subjects satisfied the inclusion criteria, including 669 and 439 subjects at 1- and 2-years post-surgery, respectively (Fig. 1). Compared with the target paper’s development cohort, our study’s two validation cohorts had slightly higher, but generally similar, age and PSA values and Gleason scores (Table 1). 52% of the development cohort had a Gleason score ≤ 6, but half of the 1-year and 2-year validation cohorts had a score of 3 + 4 (49 and 50%, respectively) and these two cohorts had at least twice as large a proportion of subjects with Gleason scores of 4 + 3 or ≥ 8 compared to the development cohort.

Fig. 1
figure 1

Flow diagram presenting inclusion/exclusion criteria Abbreviations: cT: Clinical T stage, PSA: prostate-specific antigen

Table 1 Summary statistics for patient age and disease characteristics in the development and two validation cohorts

Table 2 summarizes the validation cohort EPIC-26 domain scores pre- and post-surgery. The mean pre-surgery sexual scores ranged from 61.2 to 63 and were considerably higher than their post-surgery equivalents (ranging from 29.1 to 37.1). The 1-year and 2-year pre-surgery average urinary incontinence scores of 93.3 and 93.1, respectively, were also higher than the post-surgery scores of 74.6 and 76.1, respectively. However, the 1-year and 2-year pre-surgery mean urinary irritation/obstruction scores of 86.7 and 85.8, respectively were slightly lower than the post-surgery scores of 92.4 and 93.3, respectively. The pre-surgery and post-surgery scores were similar across the bowel and hormone domains. The scores across the 1-year and 2-year cohorts were very similar; the largest difference was a post-surgery sexual score of 34 in the 1-year cohort compared to 29.1 in the 2-year cohort.

Table 2 Pre- and post-surgery mean EPIC-26 scores of the two validation cohorts

PROMs – sexual function

Of the five models, the sexual domain models exhibited the most accurate predictions (Table 3 and Fig. 2). At 1- and 2-years post prostatectomy, the 5-question sexual domain models had the highest variance explained (R2 = 0.35, SD = 0.02, respectively) and discrimination (c-index = 0.73 and 0.72, SD = 0.01 and 0.01, respectively), small bias (CL = − 2.2 and − 2.3, SD = 0.6 and 0.7 respectively), and calibration slopes close to one (CS = 1.04 and 1.08, SD = 0.04 and 0.03 respectively). The magnitude of differences between predictions and observations were large (MAE = 19.6 and 20.4, SD = 0.4 and 0.4; RMSE = 24.4 and 25.5, SD = 0.4 and 0.4, respectively) compared to other models. However, this was because the observations were further away from the minimum (0) and maximum (100) score values.

Table 3 Mean (standard deviation) of the validation statistics from 100 imputed datasets for 1- and 2-year post-prostatectomy score predictions
Fig. 2
figure 2

Predicted vs observed EPIC-26 scores Predicted vs observed EPIC-26 scores at 1 year (top row) and 2 years (bottom row) post-prostatectomy for the first imputed set. The blue solid line is a locally weighted smoothing fit with 95% confidence intervals (blue dashed lines), the black diagonal dashed line is the line of perfect prediction

PROMs – incontinence

The 1- and 2-year urinary incontinence and irritation/obstruction models produced consistently poor validation values. However, the urinary incontinence models had satisfactory calibration slopes (1-year CS = 1.16, SD = 0.13 and 2-year CS = 0.99, SD = 0.17, the latter slope being almost ideal).

PROMs – bowel and hormonal

Both the bowel and hormone models exhibited reasonable variance explained (1-year bowel and hormone: R2 = 0.21 and 0.15, SD = 0.03 and 0.03, respectively), discrimination (c-index = 0.70, SD = 0.01 for both models), bias (CL = 2.1 and 5.1, SD = 0.1 and 0.1, respectively) and calibration slopes (CS = 0.84 and 0.85, SD = 0.06 and 0.06, respectively).

All the 1-year and 2-year values were approximately similar in all domains. The model fit measures (RMSE and MAE) were similar or slightly smaller than the equivalent values in the development cohort, with the exception of the urinary incontinence domain which had slightly larger values. All models exhibited lower R2 compared to those in the development cohort. Supplementary Table 2 presents the supplementary results. The 6-question sexual domain model statistics were all similar to the 5-question statistics. The calibration statistics when estimated using linear (mean) regression instead of quantile (median) regression were similar in every domain except for urinary incontinence.

Discussion

This study externally validates a recently published online tool designed to predict patient symptoms after treatment for localized prostate cancer. The analysis shows excellent predictive performance of the model for sexual function at both 1- and 2-year assessments post-prostatectomy, reasonable performance for bowel and hormonal function, and poor predictive value of urinary incontinence and irritation/obstruction. In particular, these models demonstrated good calibration and discrimination. The online tool would be beneficial for predicting sexual, bowel and hormonal outcomes following prostatectomy. However, the hormonal predictions are not germane to surgery; as seen in Table 2, there is no difference between the three time points. Thus, using the tool to predict post RP hormonal symptoms has little clinical application despite our validation. This can also be concluded for the bowel or rectal function; the descriptive statistics for the bowel domain displayed little change in following prostatectomy. In contrast, there was a large difference in sexual function before and after surgery.

The model performance was worse in the validation cohorts than the development cohort. Although this is to be expected with validation cohorts, the discrepancy could be partially explained by a number of factors. Sexual function is particularly impeded by prostatectomy compared to other domains [12], as shown in the differences between the pre- and post-surgery scores. Our analysis only included surgery, whereas Laviana et al. also included radiotherapy and active surveillance collectively, treatment modalities associated with relatively smaller decreases in sexual scores post-treatment [9]. Surgery has been associated with worse urinary incontinence but improved urinary irritation/obstruction, which matches the validation pre- and post-surgery domain scores, compared to EBRT or AS [12, 24]. This may explain our observed poor correlation for urinary incontinence and irritation/obstruction given we examined RP alone, and the inclusion of AS and EBRT in the development cohort likely resulted in a substantial bias towards the null for surgical subjects. Although the original models considered three different treatments simultaneously, models stratified by treatment may be more specific and meaningful for the relevant therapy. Differences in RP surgical approaches between the two cohorts could also result in differences in PROM outcomes. For example, higher proportions of nerve-sparing surgeries may be associated with improved erectile function and urinary continence [4, 25]. However, the inclusion of AS and ERBT in the development cohort but not the validation cohort is likely to predispose to larger differences between the cohorts than the choice of RP surgical technique.

Differences in model performance can also result from underlying population differences. Although the development and validation cohorts’ summary statistics appear reasonably similar, the validation cohorts had on average higher Gleason scores and, to a small extent, age and PSA. These are all risk factors associated with worse outcomes [26]. RP is more likely to be offered to higher risk patients compared to AS [27], and in particular, AS patients generally have Gleason scores ≤6 [28]. Hence the validation cohorts, which only include subjects who had RP, were likely to have overall stronger risk factors compared to a cohort that includes subjects who elected for AS. Model performance can also vary because of differences in the US and South Australian healthcare systems and socioeconomic factors which have not be addressed in these analyses. Underlying population differences were likely given the lower R2 values in the validation cohorts. However, the RMSEs and MAEs were similar between the cohorts. These two statistics are less volatile to population differences and indicate a similar model fit with both data sets despite their differences [29]. This also suggests that the original models were unlikely to be overfitted to the development population.

Our study’s strengths include a large validation cohort size collected across 13 years. The registry has data on most newly diagnosed prostate cancers in South Australia [15], reflecting a high population level recruitment inherent in disease-specific registries. By validating the development models in a surgery-only cohort, this study’s results potentially better reflect the prevalence of PROMs in surgical patients, particularly for sexual and urinary symptoms. Multiple performance statistics were considered for a comprehensive model validation exploring calibration, model fit, variance explained, and discrimination.

SA-PCCOC does not record racial group, and both PSA and general health had substantial missingness, which are the most pronounced study limitations and may obscure underlying population differences between the cohorts. However multiple imputation was implemented to minimise the impact of missing data. As subjects with advanced prostate cancer were excluded, we believe it is unlikely that missing data was greatly impacted by the effect of prostate cancer on subjects’ health. Thus assuming the data was missing at random is reasonable, minimizing potential bias from the multiple imputation. Furthermore, omitting racial group would likely only have a small effect. Racial group in the development models only created a difference in the predicted scores of up to 3.8 in 99% of subjects for the sexual and urinary incontinence domains (races of “White”, “Black”, “Hispanic”, and “Asian”), and even smaller differences no greater than 2.4 for all subjects in the other three domains [9]. These differences are all below the minimum clinically important difference (MCID) for the EPIC-26 instrument [30]. However, our models may have underestimated urinary incontinence scores for the other 1% of subjects (the “Other” racial group), where their predicted score increase in this domain (8.5) exceeded the MCID. The visits assessed in this study differed to those from the original study, adding a 2-year visit but omitting visits occurring at 6 months, 3 years and 5 years. Although caution should be excised for the models’ validity in non-US populations at these timepoints, the generally good performance of these models at 1 and 2 years give evidence to suggest that the validated models will perform well across the full 5-year period analysed in the development study. A study observed similar mean EPIC-26 domain scores at 2- and 3-years in a large cohort of US patients post-prostatectomy almost identical to that of the development cohort [24].

Conclusions

The examined models perform well for predicting sexual and, to a lesser extent, bowel and hormonal symptoms 1- and 2-years post-surgery in our prostate cancer patients. However, they perform poorly in predicting urinary incontinence and irritation/obstruction. The urinary domain models may benefit from recalibration or revision to better predict PROMs post-prostatectomy in non-US populations. The models for the other three domains are likely suitable for non-US populations and should be considered for implementation into clinical practice.

Availability of data and materials

The data used for this study are available from SA-PCCOC. These data cannot be made publicly available as part of the SA-PCCOC Data Use Agreement. A request to receive this data can be made and discussed after contacting SA-PCCOC. SA-PCCOC’s contact details are listed online at https://www.prostatehealth.org.au/contact-us/.

Abbreviations

AS:

Active surveillance

c-index:

Concordance-index

CL:

Calibration-in-the-large

CS:

Calibration slope

EBRT:

External beam radiation therapy

EPIC-26:

Expanded Prostate Cancer Index Composite-26

MAE:

Mean absolute error

PROMs:

Patient reported outcomes measures

PSA:

Prostate specific antigen

RMSE:

Root mean square error

R2 :

Proportion of variance explained

RP:

Radical prostatectomy

SA-PCCOC:

South Australian Prostate Cancer Clinical Outcomes Collaborative

SD:

Standard deviation

References

  1. Cancer Australia. Prostate cancer in Australia statistics. Canberra, Australia: Australian Government 2021 Available from: https://www.canceraustralia.gov.au/affected-cancer/cancer-types/prostate-cancer/statistics

  2. Leslie SW, Soon-Sutton TL, IA R, Sajjad H, Siref LE, Prostate Cancer. StatPearls. Treasure Island (FL): StatPearls Publishing; 2023.

    Google Scholar 

  3. Sebesta EM, Anderson CB. The surgical Management of Prostate Cancer. Semin Oncol. 2017;44(5):347–57. https://doi.org/10.1053/j.seminoncol.2018.01.003.

    Article  PubMed  Google Scholar 

  4. Mottet N, Cornford P, van den Bergh RCN, Briers E, Eberli D, De Meerleer G, et al. Prostate Cancer Arnhem. Netherlands Euorp Assoc Urol. 2023; Available from: https://uroweb.org/guidelines/prostate-cancer

  5. Weldring T, Smith SMS. Article Commentary: Patient-Reported Outcomes (PROs) and Patient-Reported Outcome Measures (PROMs), vol. 6. Health Services Insights; 2013. p. HSI.S11093. https://doi.org/10.4137/hsi.s11093.

    Book  Google Scholar 

  6. Nelson EC, Eftimovska E, Lind C, Hager A, Wasson JH, Lindblad S. Patient reported outcome measures in practice. BMJ. 2015;350(feb10 14):g7818-g. https://doi.org/10.1136/bmj.g7818.

    Article  Google Scholar 

  7. Kingsley C, Patel S. Patient-reported outcome measures and patient-reported experience measures. BJA Educ. 2017;17(4):137–44. https://doi.org/10.1093/bjaed/mkw060.

    Article  Google Scholar 

  8. Walz J, Gallina A, Perrotte P, Jeldres C, Trinh Q-D, Hutterer GC, et al. Clinicians are poor raters of life-expectancy before radical prostatectomy or definitive radiotherapy for localized prostate cancer. BJU Int. 2007;100(6):1254–8. https://doi.org/10.1111/j.1464-410x.2007.07130.x.

    Article  PubMed  Google Scholar 

  9. Laviana AA, Zhao Z, Huang L-C, Koyama T, Conwill R, Hoffman K, et al. Development and internal validation of a web-based tool to predict sexual, urinary, and bowel function longitudinally after radiation therapy, surgery, or observation. Eur Urol. 2020;78(2):248–55. https://doi.org/10.1016/j.eururo.2020.02.007.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  10. Laviana AA, Zhao Z, Huang L-C, Koyama T, Conwill R, Hoffman K, et al. A Prediction Model for Domain Scores for Prostate Cancer Patients. Nashville, Tennessee: USA: CEASAR Investigators. Available from: https://statez.shinyapps.io/PCDSPred/

  11. Steyerberg EW, Vergouwe Y. Towards better clinical prediction models: seven steps for development and an ABCD for validation. Eur Heart J. 2014;35(29):1925–31. https://doi.org/10.1093/eurheartj/ehu207.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Sanda MG, Dunn RL, Michalski J, Sandler HM, Northouse L, Hembroff L, et al. Quality of life and satisfaction with outcome among Prostate-Cancer survivors. N Engl J Med. 2008;358(12):1250–61. https://doi.org/10.1056/nejmoa074311.

    Article  CAS  PubMed  Google Scholar 

  13. Sanda MG, Wei JT, Litwin MS. Scoring Instructions for the Expanded Prostate cancer Index Composite Short Form (EPIC-26). The University of Michigan. Available from: https://medicine.umich.edu/sites/default/files/content/downloads/Scoring%20Instructions%20for%20the%20EPIC%2026.pdf

  14. Ware JE, Snow KK, Kosinski M, Gandek B. SF-36 health survey: manual and interpretation guide. Boston, Massachusetts: The Health Institute, New England Medical Center; 1993.

    Google Scholar 

  15. South Australian Prostate Cancer Clinical Outcomes Collaborative. About Us. Adelaide, Australia: [cited 2020 27 November]. Available from: https://www.prostatehealth.org.au/about-us/

  16. South Australian Prostate Cancer Clinical Outcomes Collaborative. List of Our Collaborators. Adelaide, Australia: [cited 2021 23 December]. Available from: https://www.prostatehealth.org.au/collaborators/list-of-our-collaborators/

  17. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. Ann Intern Med. 2015;162(1):55. https://doi.org/10.7326/m14-0697.

    Article  PubMed  Google Scholar 

  18. Kvalseth TO. Cautionary note about R2. Am Stat. 1985;39(4):279–85. https://doi.org/10.2307/2683704.

    Article  Google Scholar 

  19. Harrell FE. Evaluating the yield of medical tests. JAMA: the. J Am Med Assoc. 1982;247(18):2543 https://doi.org/10.1001/jama.1982.03320430047030.

  20. White IR, Royston P, Wood AM. Multiple imputation using chained equations: issues and guidance for practice. Stat Med. 2011;30(4):377–99. https://doi.org/10.1002/sim.4067.

    Article  PubMed  Google Scholar 

  21. R Core Team. R: a language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2021.

    Google Scholar 

  22. van Buuren S, Groothuis-Oudshoorn K. Multivariate imputation by chained equations in R. J Stat Softw. 2011;45(3) https://doi.org/10.18637/jss.v045.i03.

  23. Koenker R. quantreg: Quantile Regression. 2021. Available from: https://CRAN.R-project.org/package=quantreg.

  24. Barocas DA, Alvarez J, Resnick MJ, Koyama T, Hoffman KE, Tyson MD, et al. Association between radiation therapy, surgery, or observation for localized Prostate Cancer and patient-reported outcomes after 3 years. J Am Med Assoc. 2017;317(11):1126. https://doi.org/10.1001/jama.2017.1704.

    Article  Google Scholar 

  25. Haga N, Miyazaki T, Tsubouchi K, Okabe Y, Shibayama K, Emoto D, et al. Comprehensive approach for preserving cavernous nerves and erectile function after radical prostatectomy in the era of robotic surgery. Int J Urol. 2021;28(4):360–8. https://doi.org/10.1111/iju.14491.

    Article  PubMed  Google Scholar 

  26. Srigley JR, Amin M, Boccon-Gibod L, Egevad L, Epstein JI, Humphrey PA, et al. Prognostic and predictive factors in prostate cancer: historical perspectives and recent international consensus initiatives. Scand J Urol Nephrol. 2005;39(sup216):8–19. https://doi.org/10.1080/03008880510030914.

    Article  Google Scholar 

  27. InformedHealth.org. Localized prostate cancer: Low-risk prostate cancer: Active surveillance or treatment? Cologne, Germany: Institute for Quality and Efficiency in Health Care. [updated 12 March 2020; cited 2020 26 December]. Available from: https://www-ncbi-nlm-nih-gov.proxy.library.adelaide.edu.au/books/NBK487255/

  28. Sathianathen NJ, Konety BR, Crook J, Saad F, Lawrentschuk N. Landmarks in prostate cancer. Nat Rev Urol. 2018;15(10):627–42. https://doi.org/10.1038/s41585-018-0060-7.

    Article  PubMed  Google Scholar 

  29. Alexander DLJ, Tropsha A, Winkler DA. Beware of R2: simple, unambiguous assessment of the prediction accuracy of QSAR and QSPR models. J Chem Inf Model. 2015;55(7):1316–22. https://doi.org/10.1021/acs.jcim.5b00206.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  30. Skolarus TA, Dunn RL, Sanda MG, Chang P, Greenfield TK, Litwin MS, et al. Minimally important difference for the expanded Prostate Cancer index composite short form. Urol. 2015;85(1):101–6. https://doi.org/10.1016/j.urology.2014.08.044.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors acknowledge SA-PCCOC and its data custodians Robyn McGeachie and Scott Walsh for providing the data for analysis. The data sets are the property of SA-PCCOC. We also thank Aaron A. Laviana for providing descriptive results (demographics) for their validation cohort only consisting of subjects who had surgery without external beam radiation therapy.

Funding

The authors have no relevant financial or non-financial interests to disclose. The SA-PCCOC database is funded by The Movember Foundation, The Hospital Research Foundation, Mundi Pharma and Genesis Care. Funding is not specific to this project.

Author information

Authors and Affiliations

Authors

Contributions

MAB: Conceptualization, Methodology, Software, Validation, Formal analysis, Investigation, Writing – Original Draft, Writing – Review & Editing, Visualization. MEO: Conceptualization, Methodology, Investigation, Resources, Writing – Original Draft, Writing – Review & Editing, Visualization, Supervision, Project administration. KLM: Writing – Review & Editing. ADV: Conceptualization, Methodology, Software, Validation, Formal analysis, Investigation, Writing – Original Draft, Writing – Review & Editing, Visualization, Supervision.

Corresponding author

Correspondence to Matthew A. Borg.

Ethics declarations

Ethics approval and consent to participate

Data from the SA-PCCOC database was used for this study. The database has ethics approval from the Southern Adelaide Clinical Human Research Ethics Committee which extends to this study. Informed consent, in the form of an ethics-approved opt-out recruitment system, was obtained from all individual subjects and/or their legal guardian(s) included in the study. All methods were carried out in accordance with relevant guidelines and regulations.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1: Supplementary Table 1.

TRIPOD Checklist: Prediction Model Validation. Supplementary Table 2. Validation statistics for the five prediction models across the 100 imputed datasets of the 1- and 2-year post-prostatectomy cohorts.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Borg, M.A., O’Callaghan, M.E., Moretti, K.L. et al. External validation of predictive models of sexual, urinary, bowel and hormonal function after surgery in prostate cancer subjects. BMC Urol 24, 2 (2024). https://doi.org/10.1186/s12894-023-01373-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12894-023-01373-9

Keywords