Introduction

The proliferation of digital health technologies offers individuals extensive access to their health information through fitness trackers, continuous monitors, and self-diagnostic platforms1. Increasingly, these provide insights into possible current or future diseases and actions that can be taken by individuals to address or avoid these1,2. While these tools generate valuable biometric data, medical insight, and potentially enable greater control over one’s well-being, they also present a paradox: more information does not necessarily yield better health outcomes, and there is evidence, at least for some users, that they can experience digital health fatigue, anxiety, and information overload3,4. This can potentially result in increased cognitive and emotional burden and decreased engagement with healthcare – and it may even ultimately result in compromised health outcomes3,4,5. This article explores how artificial intelligence-enabled assistants might balance informed health management alongside cognitive burden, maximizing digital health benefits while mitigating their potential harms.

The digital health information dilemma

Digital health fatigue, in this article, refers to the diminishing returns and growing burden experienced by users bombarded with continuous streams of health metrics and nudges from monitoring technologies. Terms like ‘cyberchondria’ have also emerged, referring to health anxiety fueled by excessive digital self-tracking5. Recent research suggests that for select populations, interactive digital health tools and continuous health monitoring may impose significant cognitive and emotional burdens, including heightened anxiety from misinterpreting normal physiological fluctuations, and emergence of maladaptive tracking behaviors that transform well-intentioned health management into rigid, distressing preoccupations with metrics and targets6,7. For instance, users may misinterpret smartwatch electrocardiogram (ECG) alerts as signs of cardiac disease or become anxious when wearable sleep metrics suggest poor rest despite feeling well. Recent studies have suggested that information overload from digital health tools can lead to user fatigue and skepticism; ultimately resulting in reduced engagement, tool abandonment, and poorer health outcomes as users disengage from both the technology and often the healthy behaviors these platforms were designed to encourage6,7,8,9.

This phenomenon in part parallels concerns in traditional healthcare regarding medical (i.e., healthcare practitioner-initiated) ‘overtesting’ or low-value testing - where unnecessary diagnostic procedures generate false positives, incidental findings, and patient stress10. Some alerts or abnormal readings from wearables can serve as analogous “false alarms,” where benign variations trigger unnecessary concern or follow-up. In the digital health context, this issue potentially operates at a larger scale and frequency, with consumer-oriented tools continuously generating metrics and alerts that often involve interpretation without clinical or expert oversight11.

Navigating health information overload

Current approaches

Current digital health technologies employ several strategies to attempt to manage the overwhelming volume of health data generated by modern monitoring systems. These information management strategies include, but are not limited to, approaches such as: reducing data volume through filtering, improving data presentation through consolidation and visualization, and, more recently, AI-mediated interpretation (Fig. 1)4,12,13.

Fig. 1
figure 1

Selected applications of a personal AI ‘guardian’ filtering health data into user-friendly outputs.

Filtering often involves personalized threshold systems that allow users to customize when and how they receive health-related notifications. For instance, Fitbit allows users to set personalized heart rate alerts based on their individual norms rather than using average population-based thresholds14, theoretically reducing alert fatigue while maintaining sensitivity to physiologic changes.

Data consolidation and visualization strategies, exemplified by platforms like Apple Health Kit, have also tried to address the cognitive burden of monitoring disparate health parameters15. These systems can synthesize health metrics from various devices or sensors, such as heart rate data from a watch, sleep patterns from wearables, and glucose readings from continuous monitors, into unified, intuitive dashboards that allow users to track their health holistically rather than through fragmented applications.

However, amongst other challenges, these strategies still place the burden on users to interpret what their health data actually means and determine appropriate responses to concerning patterns.

The promise of AI health companions

Artificial intelligence-driven health companions have been proposed as a potential solution to the digital health information dilemma. Proponents have suggested that these assistive AI systems could serve as intelligent mediators between users and their health information, helping to filter, contextualize, and personalize health data in ways that promote understanding without overwhelming users with every information point. They could function to notify individuals about health concerns and suggest interventions at a level of detail and based on thresholds predetermined by that individual, or, perhaps, determined by the system to be appropriate for that individual.

Several promising technologies may help enable these AI health companions. Large language models (LLMs) adapted for healthcare applications represent one such technology. Recent models like Google’s Personal Health LLM (PH-LLM) demonstrate how these models can be fine-tuned to interpret raw sensor data from wearables and generate personalized health insights16. For instance, by distilling complex data into actionable recommendations, these systems can improve sleep and fitness behaviors with expert-level performance. Across 857 expert-curated cases constructed from real-world wearable data, PH-LLM’s recommendations were rated comparable to human experts for fitness and only slightly below expert ratings for sleep, though still receiving the top score 73% of the time in the sleep domain. By summarizing dense streams of wearable metrics into clear, prioritized guidance, such systems may also help users navigate information overload and reduce the digital health fatigue that often arises from constant, uninterpreted data influx. Similarly, Google’s recent Personal Health Agent (PHA) has introduced a multi-agent health companion framework composed of a data science agent (analyzing personal and population health data), a domain expert agent (contextualizing findings with medical knowledge), and a health-coach agent (e.g., supporting behavior change through evidence-based and motivational strategies), illustrating how layered generative architectures can operationalize the filtering and personalization functions envisioned for AI health companions17,18,19.

In parallel, recent research demonstrates that LLMs can be fine-tuned to help patients navigate and comprehend their medical records, translating complex medical terminology into accessible explanations and reducing information overload20. For instance, a patient reviewing their oncology clinic note could be presented with a list of unfamiliar but clinically important terms, such as ‘rituximab’ or ‘large cell lymphoma’, each linked to plain-language explanations, helping them focus on the key elements of their care plan rather than being overwhelmed by dense jargon.

In clinical practice, LLMs have demonstrated the ability to translate continuous glucose sensor data into a concise, plain-language two-week overview; these generated summaries were judged by clinicians to have high accuracy, completeness, and safety ratings21.

Taken together, early evidence suggests LLM-powered companions can convert and triage high volumes of incoming data into patient-specific guidance, potentially reducing cognitive load while still surfacing clinically important insights - provided they meet standards for safety validation, privacy and security, and ongoing user and clinical oversight.

Building better systems for the future: considerations and opportunities

Technical frameworks and implementation

LLM-based systems for managing health information overload require technical architectures that can ingest, process, and then selectively deliver personalized health insights, rather than simply presenting users with raw data streams from multiple devices.

Data ingestion solutions have included, but are not limited to: API connections that pull data from multiple health devices and platforms (including electronic health records through standards like FHIR), natural language processing of PDFs and clinical reports, neural network adapters that convert sensor measurements into structured formats compatible with language models, and hybrid approaches that combine automated data collection with manual user input for contextual information16,22,23,24. Health data processing approaches for LLMs often involve data standardization methods such as normalization to individual baselines, population percentiles, or statistical scaling techniques, along with temporal aggregation strategies that may compute rolling averages, trend analysis, or pattern recognition over configurable time windows (such as 7-day heart rate averages or 30-day sleep pattern analysis)16,24,25,26. Robust LLM-based technologies primarily focused on filtering and selectively delivering health information have yet to emerge. However, other AI-driven approaches for triaging health data have included knowledge-based filtering systems (using domain expertise or user-determined rules to screen incoming health data), contextual prefiltering algorithms that filter information based on user circumstances (such as suppressing non-urgent notifications during designated sleep hours), condition-driven signal selection (prioritizing metrics tied to a person’s diagnoses or historical trends) or combinations of these approaches26,27,28. Finally, LLM systems can surface relevant health data to users in their raw form, or as in the PH-LLM, in structured responses, such as summaries, and SMART (Specific, Measurable, Achievable, Relevant, Time-bound) recommendations16,26.

User-centered design considerations

Effective health information filtering systems must navigate fundamental tensions between being helpful and being intrusive29. Should these systems autonomously filter out information they deem irrelevant, risking suppression of data users might actually want to see? Should they make health-related decisions independently, or limit themselves to presenting curated information while preserving user agency? A further critical design challenge involves determining when or if systems should escalate concerns to healthcare professionals versus empowering users to make those decisions themselves. These philosophical questions in part underscore the need for transparent user control over filtering parameters, clear communication about what information has been suppressed and why, and mechanisms that allow users to adjust the balance between comprehensive monitoring and cognitive simplicity based on their individual health literacy and preferences for medical autonomy30.

Governance and oversight

In implementing AI-driven solutions for digital health, ensuring robust governance and oversight is essential. Regulatory bodies may face challenges as these systems can blur traditional boundaries between medical devices and consumer products, necessitating innovation in oversight approaches31. One such governance challenge centers on maintaining appropriate human medical expertise and oversight in increasingly automated systems, ensuring AI augments rather than compromises clinical judgment while still delivering on the promise of reduced information burden.

Conclusion

Artificial intelligence may offer solutions to the paradox of digital health information overload, potentially balancing informed health management with reduced cognitive burden. More research and collaboration between consumers, clinicians, and developers is needed to optimize these systems and ensure they truly enhance health outcomes. Ultimately, the success of AI health companions will depend as much on technical sophistication as on trust. Patients and clinicians must be assured that these systems are accurate, safe, and transparent, with clear accountability and meaningful oversight. As these technologies continue to evolve, their thoughtful implementation offers an opportunity to reshape digital health into a domain where information genuinely empowers, rather than burdens, its users.