Logo Logo
Hilfe
Hilfe
Switch Language to English

Schalk, Daniel; Rehms, Raphael ORCID logoORCID: https://orcid.org/0000-0002-3996-0194; Hoffmann, Verena S. ORCID logoORCID: https://orcid.org/0000-0003-1478-5390; Bischl, Bernd ORCID logoORCID: https://orcid.org/0000-0001-6002-6980 und Mansmann, Ulrich ORCID logoORCID: https://orcid.org/0000-0002-9955-8906 (2024): Distributed non-disclosive validation of predictive models by a modified ROC-GLM. In: BMC Medical Research Methodology, Bd. 24, 190 [PDF, 4MB]

Abstract

Background : Distributed statistical analyses provide a promising approach for privacy protection when analyzing data distributed over several databases. Instead of directly operating on data, the analyst receives anonymous summary statistics, which are combined into an aggregated result. Further, in discrimination model (prognosis, diagnosis, etc.) development, it is key to evaluate a trained model w.r.t. to its prognostic or predictive performance on new independent data. For binary classification, quantifying discrimination uses the receiver operating characteristics (ROC) and its area under the curve (AUC) as aggregation measure. We are interested to calculate both as well as basic indicators of calibration-in-the-large for a binary classification task using a distributed and privacy-preserving approach. Methods : We employ DataSHIELD as the technology to carry out distributed analyses, and we use a newly developed algorithm to validate the prediction score by conducting distributed and privacy-preserving ROC analysis. Calibration curves are constructed from mean values over sites. The determination of ROC and its AUC is based on a generalized linear model (GLM) approximation of the true ROC curve, the ROC-GLM, as well as on ideas of differential privacy (DP). DP adds noise (quantified by the ℓ2 sensitivity Δ2(f^)) to the data and enables a global handling of placement numbers. The impact of DP parameters was studied by simulations. Results : In our simulation scenario, the true and distributed AUC measures differ by ΔAUC<0.01 depending heavily on the choice of the differential privacy parameters. It is recommended to check the accuracy of the distributed AUC estimator in specific simulation scenarios along with a reasonable choice of DP parameters. Here, the accuracy of the distributed AUC estimator may be impaired by too much artificial noise added from DP. Conclusions : The applicability of our algorithms depends on the ℓ2 sensitivity Δ2(f^) of the underlying statistical/predictive model. The simulations carried out have shown that the approximation error is acceptable for the majority of simulated cases. For models with high Δ2(f^), the privacy parameters must be set accordingly higher to ensure sufficient privacy protection, which affects the approximation error. This work shows that complex measures, as the AUC, are applicable for validation in distributed setups while preserving an individual’s privacy.

Dokument bearbeiten Dokument bearbeiten