Towards Privacy and Utility in Tourette TIC Detection Through Pretraining Based on Publicly Available Video Data of Healthy Subjects
Research output: Contribution to journal › Conference article › Contributed › peer-review
Contributors
Abstract
Data privacy is typically particularly difficult to achieve in medical applications of machine learning, despite its importance in this area. The datasets are often small, which is why machine learning models such as neural networks tend to memorize information about the training data. This allows confidential and sensitive information about patients to be extracted from the model. Further challenging achieving data privacy is that the best possible utility must be ensured. In this work, we aim to detect tics based on video data of patients with Gilles de la Tourette syndrome. Facial landmarks were used as a lower-dimensional representation of the video data. Through membership inference attacks, we show that training a simple neural network directly on sensitive training data leaks information about the training data, and that this can be prevented by suitable pretraining on a large amount of unlabeled public data of healthy subjects. The proposed approach can not only reduce the attack accuracy to 52.15 %, but also achieves a high tic detection accuracy of 86.53 %.
Details
Original language | English |
---|---|
Journal | International Conference on Acoustics, Speech, and Signal Processing (ICASSP) |
Publication status | Published - 2023 |
Peer-reviewed | Yes |
Conference
Title | 48th IEEE International Conference on Acoustics, Speech and Signal Processing |
---|---|
Subtitle | Signal Processing in the AI era |
Abbreviated title | ICASSP 2023 |
Conference number | 48 |
Duration | 4 - 10 June 2023 |
Website | |
Location | Rodos Palace Luxury Convention Resort |
City | Rhodes Island |
Country | Greece |
External IDs
ORCID | /0000-0002-2989-9561/work/151981738 |
---|
Keywords
ASJC Scopus subject areas
Keywords
- contrastive learning, membership inference, pretraining, privacy-preserving, representation learning