Predicting depression risk in cancer patients with multimodal data Passed
Thursday May 25, 2023 11:30 - 11:35 G3
Lecturer: Anne De Hond
Track: MIE: Natural Language Processing
When patients with cancer develop depression, it is often left untreated. We developed a prediction model for depression risk within the first month after starting cancer treatment using machine learning and Natural Language Processing (NLP) models. The LASSO logistic regression model based on structured data performed well, whereas the NLP model based on only clinician notes did poorly. After further validation, prediction models for depression risk could lead to earlier identification and treatment of vulnerable patients, ultimately improving cancer care and treatment adherence.
Language
English
Seminar type
On site only
Objective of lecture
Inspiration
Level of knowledge
Intermediate
Target audience
Technicians/IT/Developers
Researchers
Students
Care professionals
Healthcare professionals
Keyword
Innovation/research
Conference
MIE
Authors
Anne de Hond, Marieke van Buchem, Claudio Fanconi, Ilse Kant, Ewout W Steyerberg, Tina Hernandez-Boussard
Lecturers
Anne De Hond Lecturer
PhD Candidate
Leiden University Medical Center
Anne obtained her master's degree in Econometrics and Management Science from the Erasmus University Rotterdam. Her econometrics studies piqued her interest in data modelling for the healthcare sector. Anne started her PhD research in 2018 at the Erasmus School of Health Policy and Management where she studied adaptation to disability and quality of life assessment. After a year and a half, her research interests pivoted towards artificial intelligence for clinical prediction algorithms. She continued her PhD research in 2019 at the Leiden University Medical Center under the supervision of prof. dr. Ewout Steyerberg and dr. Ilse Kant. During her PhD, she collaborated with prof. dr. Tina Hernandez-Boussard at Stanford University, where she studied multi-modal prediction models and algorithmic fairness. Anne's research interests are fairness for AI models, validation of AI for healthcare practice, and explainable AI.