Privacy-preserving AUC Computation in Distributed Machine Learning with PHT-meDIC

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Ensuring privacy in distributed machine learning while computing the Area Under the Curve (AUC) is a significant challenge because pooling sensitive test data is often not allowed. Although cryptographic methods can address some of these concerns, they may compromise either scalability or accuracy. In this paper, we present two privacy-preserving solutions for secure AUC computation across multiple institutions: (1) an exact global AUC method that handles ties in prediction scores and scales linearly with the number of samples, and (2) an approximation method that substantially reduces runtime while maintaining acceptable accuracy. Our protocols leverage a combination of homomorphic encryption (modified Paillier), symmetric and asymmetric cryptography, and randomized encoding to preserve the confidentiality of true labels and model predictions. We integrate these methods into the Personal Health Train (PHT)-meDIC platform, a distributed machine learning environment designed for healthcare, to demonstrate their correctness and feasibility. Results using both real-world and synthetic datasets confirm the accuracy of our approach: the exact method computes the true AUC without revealing private inputs, and the approximation provides a balanced trade-off between computational efficiency and precision. All relevant code is publicly available at https://github.com/PHT-meDIC/PP-AUC , facilitating straightforward adoption and further development within broader distributed learning ecosystems.

Author summary

A commonly used metric to evaluate the performance of machine learning models is the Area Under the Curve (AUC). Calculating the AUC in distributed machine learning settings is challenging because data cannot be shared between institutions due to privacy concerns. To address this, we developed two privacy-preserving methods: one that calculates the exact AUC securely and another that provides faster approximations with high accuracy. These methods use advanced encryption techniques to protect sensitive data while enabling secure collaboration. We tested them in a real-world healthcare platform called PHT-meDIC and demonstrated their effectiveness. The code is publicly available at https://github.com/PHT-meDIC/PP-AUC to support wider adoption.

Article activity feed