Skip to content

(MLHC'23) UDAMA: Unsupervised Domain Adaptation through Multi-discriminator Adversarial Training with Noisy Labels Improves Cardio-fitness Prediction

Notifications You must be signed in to change notification settings

yvonneywu/UDAMA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

📖 UDAMA: Unsupervised Domain Adaptation through Multi-discriminator Adversarial Training with Noisy Labels Improves Cardio-fitness Prediction

This repository contains the implementation code for the paper:

header image

Abstract

Deep learning models have shown great promise in various healthcare monitoring applications. However, most healthcare datasets with high-quality (gold-standard) labels are small-scale, as directly collecting ground truth is often costly and time-consuming. As a result, models developed and validated on small-scale datasets often suffer from overfitting and do not generalize well to unseen scenarios. At the same time, large amounts of imprecise (silver-standard) labeled data, annotated by approximate methods with the help of modern wearables and in the absence of ground truth validation, are starting to emerge. However, due to measurement differences, this data displays significant label distribution shifts, which motivates the use of domain adaptation. To this end, we introduce UDAMA, a method with two key components: Unsupervised Domain Adaptation and Multidiscriminator Adversarial Training, where we pre-train on the silver-standard data and employ adversarial adaptation with the gold-standard data along with two domain discriminators. In particular, we showcase the practical potential of UDAMA by applying it to Cardio-respiratory fitness (CRF) prediction. CRF is a crucial determinant of metabolic disease and mortality, and it presents labels with various levels of noise (goldand silver-standard), making it challenging to establish an accurate prediction model. Our results show promising performance by alleviating distribution shifts in various label shift settings. Additionally, by using data from two free-living cohort studies (Fenland and BBVS), we show that UDAMA consistently outperforms up to 12% compared to competitive transfer learning and state-of-the-art domain adaptation models, paving the way for leveraging noisy labeled data to improve fitness estimation at scale.

Data

We use data from the Fenland Study and the Biobank Validation Study. We cannot publicly share this data, but it is available from the MRC Epidemiology Unit at the University of Cambridge upon reasonable request.

To facilitate easier testing of our code, we provide small samples with the same vector shapes and naming conventions and put them into /example_data. The input vector of activity for both pre-training and fine-tuning is a 3D tensor of dimensions [samples, timesteps, features] while the output heart rate is an 1D vector of [samples]. In particular, in /example_data we provide X = [2, 600, 26] and y = [2].

Getting started

For development, we used Python 3.9.7 and TensorFlow 2.4.1. The main libraries needed to execute our code are as follows:

  • keras 2.4.3
  • matplotlib 3.5.1
  • pandas 1.2.2
  • scikit-learn 1.1.2
  • numpy 1.19.5
  • tqdm==4.62.3

Run the code

To run the experiments:

python 10_run_exp.py  -r 1 -f 3 -a1 0.9 -a2 0.1

Pre-trained models

We provide the best pre-trained model and its weights in the folder /example_model/. This model can be used directly to extract embeddings with 10_run_exp.py and subsequently perform domain adaptation.

Citation

Yu Wu, Dimitris Spathis, Hong Jia, Ignacio Perez-Pozuelo, Tomas Gonzales, Soren Brage, Nicholas Wareham, Cecilia Mascolo. "UDAMA: Unsupervised Domain Adaptation through Multi-discriminator Adversarial Training with Noisy Labels Improves Cardio-fitness Prediction" In Machine Learning for Healthcare Conference. New York, USA, 2023

About

(MLHC'23) UDAMA: Unsupervised Domain Adaptation through Multi-discriminator Adversarial Training with Noisy Labels Improves Cardio-fitness Prediction

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages