Synthical logo
Synthical
Your space
Profile
Activity
Favorites
Folders
Feeds
From arXiv

Privacy-aware Early Detection of COVID-19 through Adversarial Training

Early detection of COVID-19 is an ongoing area of research that can help with triage, monitoring and general health assessment of potential patients and may reduce operational strain on hospitals that cope with the coronavirus pandemic. Different machine learning techniques have been used in the literature to detect coronavirus using routine clinical data (blood tests, and vital signs). Data breaches and information leakage when using these models can bring reputational damage and cause legal issues for hospitals. In spite of this, protecting healthcare models against leakage of potentially sensitive information is an understudied research area. In this work, we examine two machine learning approaches, intended to predict a patient's COVID-19 status using routinely collected and readily available clinical data. We employ adversarial training to explore robust deep learning architectures that protect attributes related to demographic information about the patients. The two models we examine in this work are intended to preserve sensitive information against adversarial attacks and information leakage. In a series of experiments using datasets from the Oxford University Hospitals, Bedfordshire Hospitals NHS Foundation Trust, University Hospitals Birmingham NHS Foundation Trust, and Portsmouth Hospitals University NHS Trust we train and test two neural networks that predict PCR test results using information from basic laboratory blood tests, and vital signs performed on a patients' arrival to hospital. We assess the level of privacy each one of the models can provide and show the efficacy and robustness of our proposed architectures against a comparable baseline. One of our main contributions is that we specifically target the development of effective COVID-19 detection models with built-in mechanisms in order to selectively protect sensitive attributes against adversarial attacks.
Upvote
Simplify
Published on January 9, 2022
Copy BibTeX
Loading...
Comments
Summary
There is no AI-powered summary yet, because we do not have a budget to generate summaries for all articles.
1. Buy subscription
We will thank you for helping thousands of people to save their time at the top of the generated summary.
If you buy our subscription, you will be able to summarize multiple articles.
Pay $undefined
≈10 summaries
Pay $undefined
≈60 summaries
2. Share on socials
If this article gets to top-5 in trends, we'll summarize it for free.
Copy link
Content
Summary