UbiWell Lab

The Ubiquitous Computing for Health and Well-being (UbiWell) Lab is an interdisciplinary research group at the Khoury College of Computer Sciences and the Bouvé College of Health Sciences at Northeastern University.
We work on developing data-driven solutions to enable effective sensing and interventions for mental- and behavioral-health outcomes with mobile and ubiquitous technologies.

Research Areas

Our interdisciplinary team works at the intersection of mobile/wearable sensing, data science, human-centered computing, and behavioral science.

We work on exploring and advancing the complete "lifecycle" of mental- and behavioral-health sensing and intervention, which includes (a) accurately sensing and detecting a mental or behavioral health condition, like stress and opioid use; (b) after detecting a particular condition, determining the right time to deliver the intervention or support, such that the user is most likely to be receptive to the interventions provided; and (c) choosing the best intervention delivery mechanism and modality to ensure just-in-time delivery and reachability.

Sensing to intervention lifecycle

A simplified representation of the sensing to intervention lifecycle.

Current Projects

Causal modeling for physiological stress

Stress predictions from physiological signals

We are working on various projects to leverage multimodal data and understand the contextual and behavioral factors that lead to physiological stress, and evaluate its association with the perception of stress.

Predicting relapse during OUD treatment

Stress predictions from physiological signals

We are working on a longitudinal study to detect at-risk indicators, e.g., stress, craving, and mood, among patients undergoing Opioid Use Disorder (OUD) treatment, using passively collected contextual and sensor data from smartphones and wearables.

States-of-receptivity for digital health interventions

Stress predictions from physiological signals

We have multiple projects currently underway to better evaluate the contexts where people are willing and able to engage with and use digital health interventions. These range from behavior-change interventions in free-living situations to interventions during specific scenarios, e.g., driving.

News

Show all

Recent Publications

Show all
CardioAI: A Multimodal AI-based System to Support Symptom Monitoring and Risk Prediction of Cancer Treatment-Induced Cardiotoxicity
Siyi Wu, Weidan Cao, Shihan Fu, Bingsheng Yao, Ziqi Yang, Changchang Yin, Varun Mishra, Daniel Addison, Ping Zhang, Dakuo Wang
Proceedings of the Conference on Human Factors in Computing Systems (CHI) 2025
Show details
Feasibility and Utility of Multimodal Micro Ecological Momentary Assessment on a Smartwatch
Ha Le, Veronika Potter, Rithika Lakshminarayanan, Varun Mishra, Stephen Intille
Proceedings of the Conference on Human Factors in Computing Systems (CHI) 2025
Show details
A Longitudinal Observational Study with Ecological Momentary Assessment and Deep Learning to Predict Non-Prescribed Opioid Use, Treatment Retention, and Medication Nonadherence among Persons Receiving Medication Treatment for Opioid Use Disorder
Michael V. Heinz, George D. Price, Avijit Singh, Sukanya Bhattacharya, Ching-Hua Chen, Asma Asyyed, Monique B. Does, Saeed Hassanpour, Emily Hichborn, David Kotz, Chantal A. {Lambert-Harris}, Zhiguo Li, Bethany McLeman, Varun Mishra, Catherine Stanger, Geetha Subramaniam, Weiyi Wu, Cynthia I. Campbell, Lisa A. Marsch, Nicholas C. Jacobson
Journal of Substance Use and Addiction Treatment 2025
Show details
AI on My Shoulder: Supporting Emotional Labor in Front-Office Roles with an LLM-based Empathetic Coworker
Vedant Das Swain, Qiuyue "Joy" Zhong, Jash Rajesh Parekh, Yechan Jeon, Roy Zimmermann, Mary P Czerwinski, Jina Suh, Varun Mishra, Koustuv Saha, Javier Hernandez
Proceedings of the Conference on Human Factors in Computing Systems (CHI) 2025
Show details
Extending Stress Detection Reproducibility to Consumer Wearable Sensors
Ohida Binte Amin, Varun Mishra, Tinashe M. Tapera, Robert Volpe, Aarti Sathyanarayana
arXiv 2025
Show details
Show all