The rise of digital health technologies has provided individuals with unprecedented access to biometric data and health insights. However, excess monitoring may contribute to fatigue, anxiety, and information overload, sometimes reducing engagement and worsening outcomes. This article explores how artificial intelligence-enabled assistants might help address this challenge by filtering, contextualizing, and personalizing health information, potentially supporting informed self-management while mitigating some unintended harms of digital health technologies.
Introduction
The proliferation of digital health technologies offers individuals extensive access to their health information through fitness trackers, continuous monitors, and self-diagnostic platforms1. Increasingly, these provide insights into possible current or future diseases and actions that can be taken by individuals to address or avoid these1,2. While these tools generate valuable biometric data, medical insight, and potentially enable greater control over one’s well-being, they also present a paradox: more information does not necessarily yield better health outcomes, and there is evidence, at least for some users, that they can experience digital health fatigue, anxiety, and information overload3,4. This can potentially result in increased cognitive and emotional burden and decreased engagement with healthcare – and it may even ultimately result in compromised health outcomes3,4,5. This article explores how artificial intelligence-enabled assistants might balance informed health management alongside cognitive burden, maximizing digital health benefits while mitigating their potential harms.
The digital health information dilemma
Digital health fatigue, in this article, refers to the diminishing returns and growing burden experienced by users bombarded with continuous streams of health metrics and nudges from monitoring technologies. Terms like ‘cyberchondria’ have also emerged, referring to health anxiety fueled by excessive digital self-tracking5. Recent research suggests that for select populations, interactive digital health tools and continuous health monitoring may impose significant cognitive and emotional burdens, including heightened anxiety from misinterpreting normal physiological fluctuations, and emergence of maladaptive tracking behaviors that transform well-intentioned health management into rigid, distressing preoccupations with metrics and targets6,7. For instance, users may misinterpret smartwatch electrocardiogram (ECG) alerts as signs of cardiac disease or become anxious when wearable sleep metrics suggest poor rest despite feeling well. Recent studies have suggested that information overload from digital health tools can lead to user fatigue and skepticism; ultimately resulting in reduced engagement, tool abandonment, and poorer health outcomes as users disengage from both the technology and often the healthy behaviors these platforms were designed to encourage6,7,8,9.
This phenomenon in part parallels concerns in traditional healthcare regarding medical (i.e., healthcare practitioner-initiated) ‘overtesting’ or low-value testing - where unnecessary diagnostic procedures generate false positives, incidental findings, and patient stress10. Some alerts or abnormal readings from wearables can serve as analogous “false alarms,” where benign variations trigger unnecessary concern or follow-up. In the digital health context, this issue potentially operates at a larger scale and frequency, with consumer-oriented tools continuously generating metrics and alerts that often involve interpretation without clinical or expert oversight11.
Navigating health information overload
Current approaches
Current digital health technologies employ several strategies to attempt to manage the overwhelming volume of health data generated by modern monitoring systems. These information management strategies include, but are not limited to, approaches such as: reducing data volume through filtering, improving data presentation through consolidation and visualization, and, more recently, AI-mediated interpretation (Fig. 1)4,12,13.
Filtering often involves personalized threshold systems that allow users to customize when and how they receive health-related notifications. For instance, Fitbit allows users to set personalized heart rate alerts based on their individual norms rather than using average population-based thresholds14, theoretically reducing alert fatigue while maintaining sensitivity to physiologic changes.
Data consolidation and visualization strategies, exemplified by platforms like Apple Health Kit, have also tried to address the cognitive burden of monitoring disparate health parameters15. These systems can synthesize health metrics from various devices or sensors, such as heart rate data from a watch, sleep patterns from wearables, and glucose readings from continuous monitors, into unified, intuitive dashboards that allow users to track their health holistically rather than through fragmented applications.
However, amongst other challenges, these strategies still place the burden on users to interpret what their health data actually means and determine appropriate responses to concerning patterns.
The promise of AI health companions
Artificial intelligence-driven health companions have been proposed as a potential solution to the digital health information dilemma. Proponents have suggested that these assistive AI systems could serve as intelligent mediators between users and their health information, helping to filter, contextualize, and personalize health data in ways that promote understanding without overwhelming users with every information point. They could function to notify individuals about health concerns and suggest interventions at a level of detail and based on thresholds predetermined by that individual, or, perhaps, determined by the system to be appropriate for that individual.
Several promising technologies may help enable these AI health companions. Large language models (LLMs) adapted for healthcare applications represent one such technology. Recent models like Google’s Personal Health LLM (PH-LLM) demonstrate how these models can be fine-tuned to interpret raw sensor data from wearables and generate personalized health insights16. For instance, by distilling complex data into actionable recommendations, these systems can improve sleep and fitness behaviors with expert-level performance. Across 857 expert-curated cases constructed from real-world wearable data, PH-LLM’s recommendations were rated comparable to human experts for fitness and only slightly below expert ratings for sleep, though still receiving the top score 73% of the time in the sleep domain. By summarizing dense streams of wearable metrics into clear, prioritized guidance, such systems may also help users navigate information overload and reduce the digital health fatigue that often arises from constant, uninterpreted data influx. Similarly, Google’s recent Personal Health Agent (PHA) has introduced a multi-agent health companion framework composed of a data science agent (analyzing personal and population health data), a domain expert agent (contextualizing findings with medical knowledge), and a health-coach agent (e.g., supporting behavior change through evidence-based and motivational strategies), illustrating how layered generative architectures can operationalize the filtering and personalization functions envisioned for AI health companions17,18,19.
In parallel, recent research demonstrates that LLMs can be fine-tuned to help patients navigate and comprehend their medical records, translating complex medical terminology into accessible explanations and reducing information overload20. For instance, a patient reviewing their oncology clinic note could be presented with a list of unfamiliar but clinically important terms, such as ‘rituximab’ or ‘large cell lymphoma’, each linked to plain-language explanations, helping them focus on the key elements of their care plan rather than being overwhelmed by dense jargon.
In clinical practice, LLMs have demonstrated the ability to translate continuous glucose sensor data into a concise, plain-language two-week overview; these generated summaries were judged by clinicians to have high accuracy, completeness, and safety ratings21.
Taken together, early evidence suggests LLM-powered companions can convert and triage high volumes of incoming data into patient-specific guidance, potentially reducing cognitive load while still surfacing clinically important insights - provided they meet standards for safety validation, privacy and security, and ongoing user and clinical oversight.
Building better systems for the future: considerations and opportunities
Technical frameworks and implementation
LLM-based systems for managing health information overload require technical architectures that can ingest, process, and then selectively deliver personalized health insights, rather than simply presenting users with raw data streams from multiple devices.
Data ingestion solutions have included, but are not limited to: API connections that pull data from multiple health devices and platforms (including electronic health records through standards like FHIR), natural language processing of PDFs and clinical reports, neural network adapters that convert sensor measurements into structured formats compatible with language models, and hybrid approaches that combine automated data collection with manual user input for contextual information16,22,23,24. Health data processing approaches for LLMs often involve data standardization methods such as normalization to individual baselines, population percentiles, or statistical scaling techniques, along with temporal aggregation strategies that may compute rolling averages, trend analysis, or pattern recognition over configurable time windows (such as 7-day heart rate averages or 30-day sleep pattern analysis)16,24,25,26. Robust LLM-based technologies primarily focused on filtering and selectively delivering health information have yet to emerge. However, other AI-driven approaches for triaging health data have included knowledge-based filtering systems (using domain expertise or user-determined rules to screen incoming health data), contextual prefiltering algorithms that filter information based on user circumstances (such as suppressing non-urgent notifications during designated sleep hours), condition-driven signal selection (prioritizing metrics tied to a person’s diagnoses or historical trends) or combinations of these approaches26,27,28. Finally, LLM systems can surface relevant health data to users in their raw form, or as in the PH-LLM, in structured responses, such as summaries, and SMART (Specific, Measurable, Achievable, Relevant, Time-bound) recommendations16,26.
User-centered design considerations
Effective health information filtering systems must navigate fundamental tensions between being helpful and being intrusive29. Should these systems autonomously filter out information they deem irrelevant, risking suppression of data users might actually want to see? Should they make health-related decisions independently, or limit themselves to presenting curated information while preserving user agency? A further critical design challenge involves determining when or if systems should escalate concerns to healthcare professionals versus empowering users to make those decisions themselves. These philosophical questions in part underscore the need for transparent user control over filtering parameters, clear communication about what information has been suppressed and why, and mechanisms that allow users to adjust the balance between comprehensive monitoring and cognitive simplicity based on their individual health literacy and preferences for medical autonomy30.
Governance and oversight
In implementing AI-driven solutions for digital health, ensuring robust governance and oversight is essential. Regulatory bodies may face challenges as these systems can blur traditional boundaries between medical devices and consumer products, necessitating innovation in oversight approaches31. One such governance challenge centers on maintaining appropriate human medical expertise and oversight in increasingly automated systems, ensuring AI augments rather than compromises clinical judgment while still delivering on the promise of reduced information burden.
Conclusion
Artificial intelligence may offer solutions to the paradox of digital health information overload, potentially balancing informed health management with reduced cognitive burden. More research and collaboration between consumers, clinicians, and developers is needed to optimize these systems and ensure they truly enhance health outcomes. Ultimately, the success of AI health companions will depend as much on technical sophistication as on trust. Patients and clinicians must be assured that these systems are accurate, safe, and transparent, with clear accountability and meaningful oversight. As these technologies continue to evolve, their thoughtful implementation offers an opportunity to reshape digital health into a domain where information genuinely empowers, rather than burdens, its users.
Data availability
No datasets were generated or analyzed during the current study.
References
Huhn, S. et al. The impact of wearable technologies in health research: scoping review. JMIR Mhealth Uhealth 10, e34384 (2022).
Yeung, A. W. K. et al. The promise of digital healthcare technologies. Front. Public Health 11, 1196596 (2023).
Rosman, L. et al. Wearable devices, health care use, and psychological well-being in patients with atrial fibrillation. J. Am. Heart Assoc. 13, e033750 (2024).
Arnold, M., Goldschmitt, M. & Rigotti, T. Dealing with information overload: a comprehensive review. Front. Psychol. 14, 1122200 (2023).
Schenkel, S. K., Jungmann, S. M., Gropalis, M. & Witthöft, M. Conceptualizations of cyberchondria and relations to the anxiety spectrum: systematic review and meta-analysis. J. Med. Internet Res. 23, e27835 (2021).
Zhong, L., Cao, J. & Xue, F. The paradox of convenience: how information overload in mHealth apps leads to medical service overuse. Front. Public Health 12, 1408998 (2024).
Rosman, L., Gehi, A. & Lampert, R. When smartwatches contribute to health anxiety in patients with atrial fibrillation. Cardiovasc. Digit. Health J. 1, 9–10 (2020).
Jakicic, J. M. et al. Effect of wearable technology combined with a lifestyle intervention on long-term weight loss: the IDEA randomized clinical trial. JAMA 316, 1161–1171 (2016).
O’Kane, M. J. et al. Efficacy of self monitoring of blood glucose in patients with newly diagnosed type 2 diabetes (ESMON study): randomised controlled trial. BMJ 336, 1174–1177 (2008).
Müskens, J. L. J. M., Kool, R. B., van Dulmen, S. A. & Westert, G. P. Overuse of diagnostic testing in healthcare: a systematic review. BMJ Qual. Saf. 31, 54–63 (2022).
Grosman-Rimon, L. & Wegier, P. With advancement in health technology comes great responsibility - Ethical and safety considerations for using digital health technology: a narrative review. Medicine 103, e39136 (2024).
Khairat, S. S. et al. The impact of visualization dashboards on quality of care and clinician satisfaction: integrative literature review. JMIR Hum. Factors 5, e22 (2018).
van Rossum, M. C. et al. Adaptive threshold-based alarm strategies for continuous vital signs monitoring. J. Clin. Monit. Comput. 36, 407–417 (2022).
Radin, J. M., Wineinger, N. E., Topol, E. J. & Steinhubl, S. R. Harnessing wearable device data to improve state-level real-time surveillance of influenza-like illness in the USA: a population-based study. Lancet Digit. Health 2, e85–e93 (2020).
Truslow, J. et al. Understanding activity and physiology at scale: the Apple Heart & Movement Study. npj Digit. Med. 7, 242 (2024).
Khasentino, J. et al. A personal health large language model for sleep and fitness coaching. Nat. Med. https://doi.org/10.1038/s41591-025-03888-0 (2025).
Heydari, A. A. et al. The anatomy of a personal health agent. Preprint at https://arxiv.org/abs/2508.20148 (2025).
Gottweis, J. et al. Towards an AI co-scientist. Preprint at https://arxiv.org/abs/2502.18864 (2025).
Tu, T. et al. Towards conversational diagnostic artificial intelligence. Nature 642, 442–450 (2025).
Jang, W. S. et al. Enhancing LLMs for identifying and prioritizing important medical jargons from electronic health record notes utilizing data augmentation [Preprint]. Preprint at https://arxiv.org/abs/2502.16022.
Healey, E., Tan, A. L. M., Flint, K. L., Ruiz, J. L. & Kohane, I. A case study on using a large language model to analyze continuous glucose monitoring data. Sci. Rep. 15, 1143 (2025).
Griffin, A. C. et al. Clinical, technical, and implementation characteristics of real-world health applications using FHIR. JAMIA Open 5, ooac077 (2022).
Qin, Y. et al. ToolLLM: Facilitating large language models to master 16000+ real-world APIs. Preprint at https://arxiv.org/abs/2307.16789.
Ferrara, E. Large language models for wearable sensor-based human activity recognition, health monitoring, and behavioral modeling: a survey of early trends, datasets, and challenges. Sensors 24, 5045 (2024).
Mishra, T. et al. Pre-symptomatic detection of COVID-19 from smartwatch data. Nat. Biomed. Eng. 4, 1208–1220 (2020).
Fang, C. M., et al. PhysioLLM: supporting personalized health insights with wearables and large language models. Preprint at https://arxiv.org/abs/2406.19283 (2024).
De Croon, R. et al. Health recommender systems: systematic review. J. Med. Internet Res. 23, e18035 (2021).
Wang, B. et al. A systematic literature review on integrating AI-powered smart glasses into digital health management for proactive healthcare solutions. npj Digit. Med. 8, 410 (2025).
van Velsen, L., Ludden, G. & Grünloh, C. The limitations of user-and human-centered design in an eHealth context and how to move beyond them. J. Med. Internet Res. 24, e37341 (2022).
Bouderhem, R. Shaping the future of AI in healthcare through ethics and governance. Humanit. Soc. Sci. Commun. 11, 416 (2024).
Freyer, O. et al. Overcoming regulatory barriers to the implementation of AI agents in healthcare. Nat. Med. https://doi.org/10.1038/s41591-025-03841-1 (2025).
Acknowledgements
None. This work received no funding.
Author information
Authors and Affiliations
Contributions
A.M. and S.G. developed the concept of the manuscript. A.M. wrote the first draft of the manuscript. A.M. and S.G. contributed to the writing, interpretation of the content, and editing of the manuscript, revising it critically for important intellectual content. A.M. and S.G. have read and approved the completed version. A.M. and S.G. take accountability for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.
Corresponding author
Ethics declarations
Competing interests
A.M. declares no nonfinancial interests and no competing financial interests. S.G. declares a nonfinancial interest as an Advisory Group member of the EY-coordinated “Study on Regulatory Governance and Innovation in the field of Medical Devices” conducted on behalf of the DG SANTE of the European Commission. S.G. declares the following competing financial interests: he has or has had consulting relationships with Una Health GmbH, Lindus Health Ltd., Flo Ltd., ICURA ApS, Rock Health Inc., Thymia Ltd., FORUM Institut für Management GmbH, High-Tech Gründerfonds Management GmbH, DG SANTE, Prova Health Ltd, Haleon Plc., and Ada Health GmbH, and holds share options in Ada Health GmbH. A.M., S.G. are News and Views Editors for npj Digital Medicine. A.M. and S.G. played no role in the internal review or decision to publish this News and Views article.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Mahajan, A., Gilbert, S. Do we need AI guardians to protect us from health information overload?. npj Digit. Med. 8, 632 (2025). https://doi.org/10.1038/s41746-025-02093-0
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41746-025-02093-0
