Fair representation learning of electronic health records

From ISLAB/CAISR
Revision as of 10:01, 3 October 2022 by Ececal (Talk | contribs)

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Title Fair representation learning of electronic health records
Summary Fair representation learning of electronic health records
Keywords fair machine learning, bias, fairness
TimeFrame Fall 2022
References Dullerud, N., Roth, K., Hamidieh, K., Papernot, N. and Ghassemi, M., 2022. Is fairness only metric deep? evaluating and addressing subgroup gaps in deep metric learning. arXiv preprint arXiv:2203.12748.

Reddy, C., Sharma, D., Mehri, S., Romero-Soriano, A., Shabanian, S. and Honari, S., 2021, June. Benchmarking bias mitigation algorithms in representation learning through fairness metrics. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1).

Yuan, Y., Xun, G., Suo, Q., Jia, K. and Zhang, A., 2019. Wave2vec: Deep representation learning for clinical temporal data. Neurocomputing, 324, pp.31-42.

Prerequisites
Author
Supervisor Ali Amirahmadi, Ece Calikus, Kobra Etminani
Level Master
Status Open


Deep representation learning methods have shown promising performance in different domains, including NLP, image analysis, and healthcare modeling. In healthcare, researchers focus on mitigating data sparsity and high dimensionality and modeling the complex short and long-term dependencies in electric health records (EHR by different representation learning methods. These models are used as feature extractors for few-shot learning and various other downstream tasks. Ensuring fairness in machine learning is extremely important to achieve health equity across different groups in society. Different approaches and frameworks address bias and fairness issues, such as anti-classification, parity, and calibration.

The ultimate goal of this project is to analyze the effect of different EHR representation learning methods on fairness and look for representations that are agnostic to the patients’ sensitive attributes or have low subgroup gaps in downstream tasks.

Potential datasets: https://mimic.mit.edu/docs/about/ https://docs.nightingalescience.org/