Introduction

Bees play a fundamental role in maintaining ecosystems through pollination, contributing to biodiversity, forest regeneration, and agricultural sustainability. Approximately 75% of the world’s crops that produce fruits and seeds for human consumption depend, at least partially, on pollinators for sustained production and quality. Beekeeping, or apiculture, encompasses the management of bee colonies to optimize pollination and honey production1.

During the past two decades, the beekeeping industry has experienced a significant decline in honey yields per colony2,3, severely affecting both productivity and sustainability. For example, in France, honey production in 2024 is estimated at 20,000 tonnes, a sharp drop from 33,900 tonnes in 2023, according to ADA France (Federation of the Regional Beekeeping Development Associations)4. According to ADA France, honey production in 2024 is projected to decrease by 27% compared to the average output recorded between 2021 and 2023. This downturn raises pressing concerns about colony health, hive productivity, and the long-term viability of apiculture. It also underscores the growing need for data-centric tools to help beekeepers effectively monitor and manage their operations.

Recent advances in information technology, such as the Internet of Things (IoT) and artificial intelligence (AI), have opened up new possibilities to improve decision making in agriculture, including beekeeping applications5,6. The Food and Agriculture Organization (FAO) has highlighted the strategic role of big data and machine learning in enabling predictive diagnostics and optimizing hive management practices1.

At the European level, Verbeke et al.7 examined the adoption of Digital Beehive Monitoring Technology (DBMT) through a survey of 844 beekeepers in 18 European countries. Their study provides insight into beekeeper characteristics, current usage patterns, perceived benefits, and key factors influencing technology adoption. The findings reveal that 79.1% of the beekeepers had not yet implemented any form of digital monitoring, while 20.9% used it to a limited extent, mainly to track hive weight. Beekeepers primarily expect DBMT to help with hive management and decision making, followed by improving colony health, reducing winter losses, and saving time, all contributing to more efficient and effective beekeeping practices. However, only about a quarter of respondents foresee a reduction in operational costs, suggesting that they take into account the additional investment required for adopting this technology.

In France, the French public bee institute (ITSAP) surveyed over 400 beekeepers to assess their interest in various functionalities that could be integrated into a digital beekeeping platform8. Figure 1 presents the results of this survey.

Fig. 1
figure 1

Beekeepers’ expectations regarding the functionalities to be integrated into a farm management tool (extracted and translated from8).

The results suggest that beekeepers prioritize features involving straightforward data representation, such as visual dashboards and summary tables. This underscores the central role of usability in the adoption and effectiveness of digital tools within apicultural contexts. Furthermore, there was notable enthusiasm for functionalities enabling real-time alerts and data sharing, reflecting a clear demand for collaborative platforms that support anticipatory and collective hive management. Beekeepers appear willing to modernize their practices and embrace digital tools, including data entry on dedicated platforms and the use of IoT devices9. The survey reveals that many already record their data on a computer or other IT devices, and that half of them express little reluctance toward adopting a suitable digital application.

From their perspective, such an application should prioritize user-friendliness and provide straightforward functionalities, allowing easy access to historical data accumulated during hive management. In addition, the application should incorporate alert features, such as weather updates and health event notifications, to facilitate timely information sharing. Ultimately, a participatory approach in both the platform’s development and the definition of its terms of use appears to be the most effective way to create a reliable and well-adapted digital tool for beekeepers.

These expectations highlight the need for a comprehensive and user-friendly digital solution that supports both real-time monitoring and collaborative hive management. To meet these needs, this article introduces BeeViz, a monitoring platform that integrates traditional apicultural knowledge with modern data-driven decision support systems. Unlike existing solutions, which often provide fragmented or isolated data visualization tools, BeeViz integrates continuous data collection, predictive analytics, and anomaly detection to help beekeepers manage real-time hives and long-term strategic planning. The system not only allows beekeepers to track hive dynamics monitoring parameters such as internal temperature, humidity, and weight but also detects early warning signs of anomalies through machine learning-based time-series analysis. This approach enables the anticipation of colony stress, food shortages, or environmental hazards, helping beekeepers take proactive measures rather than reacting to adverse conditions.

In addition to monitoring individual hives, the system paves the way for collaborative learning approaches by taking advantage of a network of connected hives to constantly refine predictive models. Unlike traditional threshold-based alerts, which rely on fixed rules, our system can dynamically adapt by incorporating live data from multiple sources, improving accuracy over time. Furthermore, expanding the monitored parameters to include CO2 levels and foraging activity could provide deeper insights into colony metabolism and environmental interactions.

This paper is structured as follows: “Related work” provides a review of existing monitoring and anomaly detection techniques, “Our approach to prediction and anomaly detection” details the proposed system architecture, “Implementation, testing, and visualization” presents implementation and experimental results, “Discussion and future research directions” addresses the limitations of our study and presents avenues for future research, finally, “Conclusion” presents conclusions.

Related work

Hive monitoring approaches

Hive monitoring in precision apiculture has evolved from manual inspection to data-driven techniques. Traditional methods rely on qualitative observations recorded by beekeepers, often on paper or hive covers10. These practices, while valuable, limit longitudinal tracking and advanced analysis.

Modern approaches fall into two main categories: computer vision–based systems and sensor-based time series monitoring. Figure 2 summarizes this classification.

Fig. 2
figure 2

Taxonomy of hive monitoring approaches.

The former includes visual and acoustic analysis using cameras and microphones. For example, Tashakkori et al.11 tracked bee movements using threshold-based algorithms, while Bjerge et al.12 used ultraviolet (UV) light and convolutional neural networks (CNNs) to detect varroa mites. Magnier et al.13 modeled bee flight trajectories to infer environmental stress, and Kriouile et al.6 applied Mask R-CNN (Region-based Convolutional Neural Network) for object detection inside hives.

Sensor-based systems, which underpin our work, use embedded devices to monitor internal hive parameters such as temperature, humidity, and weight. These systems enable real-time data acquisition and remote monitoring14,15. Studies have shown that these metrics are closely linked to colony health and behavior. For instance, internal temperature regulation is essential for brood development16, while excessive humidity can degrade honey quality and promote mold17,18. Hive weight variations can signal nectar flow, swarming, or seasonal transitions19.

Time series forecasting in beekeeping

Forecasting hive conditions using time series data has gained traction in recent years. Approaches vary by modeling technique and prediction target.

Robustillo et al.20 compared four statistical models : vector autoregressive (VAR), time-varying VAR (tvVAR), dynamic linear model (DLM), and generalized additive model (GAM) to forecast temperature, humidity, and weight. The tvVAR model showed the best performance across hives. Erdal and Tolga21 used AutoRegressive Integrated Moving Average (ARIMA) to predict national honey production in Turkey based on historical data.

Deep learning models offer greater flexibility in capturing nonlinear patterns. Anwar et al.22 developed a bidirectional long short-term memory (Bi-LSTM) with attention to estimate cumulative hive weight using internal and external sensor data. Meireles Filho et al.23 compared long short-term memory (LSTM), gated recurrent unit (GRU), and CNN architectures to forecast annual honey production, finding LSTM to be the most accurate.

These studies demonstrate the potential of time series forecasting to support proactive hive management, though most focus on isolated metrics or lack integration with anomaly detection.

Anomaly detection in beekeeping

Anomaly detection is critical for identifying disruptive events such as swarming, predator intrusion, or hive damage. Approaches fall into two categories: rule-based systems and AI-driven models.

Rule-based systems define thresholds or patterns manually, often validated statistically24,25. While effective in specific scenarios, they lack adaptability and scalability.Recent work by Chen et al.26 introduced an edge-computing system using acoustic signals for multiclass anomaly classification, highlighting the growing interest in multimodal detection.

AI-based methods include both non-regressive and regressive models. Davidson et al.27 evaluated deep recurrent AutoEncoders (AEs), isolation forests, and one-class Support Vector Machines (SVMs) for detecting swarming events using temperature data. Their results showed that AutoEncoders performed best in multivariate settings, while rule-based methods excelled in univariate contexts. In the context of precision apiculture, they identified three main categories of anomalies detectable through sensor data:

  • Swarming events: this phenomenon is an important behavioral anomaly for beekeepers, as it can lead to complete exhaustion of a colony, described by a sudden drop in hive weight. It leads to a reduction in honey production and requires immediate action to gather new colonies.

  • Sensor-related anomalies which may arise from hardware failure or calibration issues, leading to erroneous or misleading measurements.

  • Opening the hive: where manual intervention by the beekeeper can temporarily distort sensor readings, especially those related to internal temperature and weight.

Senger et al.28 used ARIMA and Seasonal ARIMA (SARIMA) models combined with Generalized Extreme Studentized Deviate (GESD) testing to detect anomalies based on residuals, achieving up to 75% detection accuracy. Further expanding on these classifications, they reported two additional types of anomalies observed in monitored hives:

  • Predator intrusions, such as mice entering hives during cold seasons, disrupting the bees’ natural clustering behavior used for thermoregulation.

  • Physical damage to the beehive, often caused by environmental factors like strong winds or the invasion of ants or wasps. These events can compromise hive integrity and may be detectable via spikes in humidity or irregular weight data.

Discussion

While the literature offers valuable insights into hive monitoring, forecasting, and anomaly detection, most approaches treat these components separately. Few systems integrate predictive analytics and anomaly detection into a unified, user-oriented platform. Moreover, generalization across hives remains a challenge due to environmental variability and data heterogeneity.

Our work addresses these gaps by combining time series forecasting and anomaly detection in a modular system designed for real-time hive management. By integrating sensor data, intelligent processing, and user-friendly visualization, BeeViz aims to support both individual and collaborative decision-making in precision apiculture.

Our approach to prediction and anomaly detection

In this section, we first introduce the main functionalities of the system, followed by its conceptual architecture. We then present the intelligent time series processing techniques and the approaches employed for anomaly detection.

Main functionalities

Monitoring smart beehives presents a number of technical and operational challenges, particularly in ensuring timely and accurate detection of anomalies that may compromise colony health. To support this goal, the system incorporates a real-time anomaly detection mechanism that alerts beekeepers to unusual conditions as they arise. As illustrated in Fig. 3, the platform provides several key features:

  • View real-time sensor data for three critical parameters: internal temperature, relative humidity, and hive weight.

  • Selection and management of individual hives, including historical data and triggered alerts.

  • Access to short-term predictive insights based on time-series forecasting for the selected hive.

  • The option to pause data collection and anomaly detection temporarily, in order to avoid false positives during interventions such as hive inspections or honey extraction.

  • Customization of alert sensitivity thresholds to adapt detection granularity to specific operational contexts.

Fig. 3
figure 3

use case diagram of BeeViz.

Proposed architecture

Effective hive management increasingly requires the deployment of connected devices and advanced data processing frameworks. In this work, we design a modular system architecture composed of six functional units, each responsible for a specific step in the data workflow, from sensor acquisition to predictive analysis. The system enables continuous data capture from smart hives, secure cloud-based storage, and real-time processing using time-series forecasting techniques. This architecture is designed to support informed decision-making by offering real-time monitoring capabilities and early detection of anomalies, thereby helping beekeepers respond proactively to environmental and colony-specific challenges.

The six component groups are (Fig. 4):

  1. 1.

    Connected hives: each user-owned hive is equipped with embedded sensors that continuously capture and transmit internal parameters—temperature, humidity, and weight—in real time.

  2. 2.

    Cloud database: real-time data collected from the hives is securely stored in a cloud-based infrastructure, ensuring both accessibility and scalability for long-term data management.

  3. 3.

    Intelligent Time Series Processing Systems: these modules analyze the incoming data streams to perform anomaly detection and generate short-term forecasts for the three monitored metrics.

  4. 4.

    Learning Database: this repository aggregates historical data, serving as a foundation for training, validating, and testing predictive models and anomaly detection algorithms.

  5. 5.

    Web-based Dashboard: the dashboard offers beekeepers an intuitive interface to visualize live hive data, forecast trends, and real-time alerts generated by the system.

  6. 6.

    Communication APIs: acting as communication gateways, the APIs facilitate data exchange between intelligent modules, the database management system, and the hive sensors, enabling real-time updates on the dashboard.

Fig. 4
figure 4

Conceptual architecture.

Designing intelligent time series processing

We implement a supervised learning approach that will take our two-column database as input:

  • Timestamp : contains the time points at which observations were recorded; the distance between two timestamps of a metric is determined according to the resampling step determined previously.

  • Y: contains the values (temperature, humidity, or weight) recorded at each corresponding timestamp.

We have tested the performance and predictive capacity of two types of architecture in our training database.

  1. 1.

    Recurrent neural networks (RNNs): RNNs are a class of artificial neural networks specifically designed to handle sequential data. Their architecture allows them to capture temporal dependencies by maintaining an internal memory of past input, which is essential for modeling time series where the order of observations impacts future values. This ability to leverage historical context makes RNNs particularly suitable for forecasting applications involving dynamic and evolving data patterns.

  2. 2.

    Facebook Prophet: Facebook Prophet is an open-source forecasting tool developed by Facebook’s data science team. It is specially designed for time series forecasting and is particularly useful for commercial and operational forecasting tasks. The Prophet procedure is basically an additive regression model.

Approaches used to design anomaly detection

Since our training database does not contain information on existing anomalies, we cannot use a classical supervised learning approach to train our models, and we cannot measure the precision of the models developed or select the best model. For this reason, we decided to let the user choose the anomaly model for each metric based on sensitivity. The sensitivity of the models developed is measured in terms of the number of anomalies it was able to detect in the training database. We have chosen three popular approaches to time series anomaly detection:

  • Predictive confidence level approach: This approach relies on the best-performing time series forecasting model previously trained using the procedure detailed in Algorithm 1. A confidence interval is constructed based on the model’s prediction and its Mean Absolute Percentage Error (MAPE), which has already been computed. This interval allows for the dynamic identification of anomalous values by flagging any observation that deviates significantly from the expected range.

  • Statistical profile approach: This approach has been used effectively in the economic and financial sectors. We will then try to project it onto our learning database.

  • Unsupervised clustering approach: One of the difficulties common to clustering algorithms for anomaly detection is defining the number of clusters, which is required by most clustering algorithms as a hyperparameter.

Although there are many techniques for estimating the number of clusters in tabular data, it is not possible to dynamically estimate the number of clusters for time series. In fact, we opt for two clustering algorithms that showed significant efficiency in our training database without the need to specify the number of clusters: (1) Density-Based Spatial Clustering of Applications with Noise (DBSCAN) and (2) isolation forest.

Implementation, testing, and visualization

Dataset

Several research projects have used open source data to develop their own version of the monitoring system. Metidji et al.29 compared several datasets, indicating the advantages and disadvantages of each. The dataset proposed by the authors, based on HOBOS, contains data without outliers or missing data. This dataset provides uninterrupted data for several years, covering the three most important parameters (temperature, humidity, and weight). We used this dataset in our study.

To ensure reproducibility and model robustness, we applied a structured data preprocessing pipeline. The main steps are summarized in Table 1. A detailed description of each step is available in our companion paper 29. Data normalization was applied at the global level across the entire hive dataset, rather than separately for each hive.

Table 1 Summary of preprocessing steps applied to the HOBOS dataset.

The dataset is made up of two hives located in Germany and contains two years of data. Figures 5 and 6 show the distribution between training, validation, and test data.

Fig. 5
figure 5

Learning data distribution for the prediction.

Fig. 6
figure 6

Learning data distribution for the anomalies detection.

For both prediction and anomaly detection tasks, we systematically applied a leave-one-hive-out strategy, where one hive was reserved exclusively for testing the trained models. This approach guarantees that:

  1. 1.

    Prediction: the model does not overfit to the data of the training hive and can generate forecasts with comparable accuracy on previously unseen hives. To ensure a rigorous evaluation of the model, we verified several key aspects of the experimental setup. The training set (730 days) and validation set (130 days) were strictly separated in time, with no overlap, ensuring temporal independence between the learning and validation phases. The test set, although derived from a different hive, covers a similar time period as the training and validation data (860 days). This configuration allows us to assess the model’s ability to generalize to a different hive under comparable environmental conditions.

  2. 2.

    Anomaly detection: the detection models maintain consistent sensitivity across the reference (training) and test hives. The training phase was conducted using data from a single hive, while the test phase relied on data from a second, distinct hive. This choice was motivated by the absence of labeled anomaly data for validation purposes as detailed in the “Approaches used to design anomaly detection”. By using a separate hive for testing, we ensured that the evaluation of the model’s ability to detect anomalies was performed on entirely unseen data, both in terms of location and hive-specific dynamics. As explained in the “Implementation of the anomaly detection system”, for instance, the confidence interval–based method for humidity anomaly detection achieves nearly identical sensitivity values for both hives (38% vs. 41%). This demonstrates that the proposed approach reliably produces a high-sensitivity anomaly detector applicable across different hives.

Implementation of the prediction system

Prediction model training algorithm

To develop an accurate and robust prediction system for hive monitoring, we implemented a model training algorithm that explores different time series forecasting techniques. The process, outlined in Algorithm 1, focuses on two primary forecasting models: Recurrent Neural Networks (RNNs) and Prophet. The system aims to predict crucial parameters of the hive, including internal temperature, humidity, and weight, using historical data.

The training process follows a structured pipeline:

  • Technology and variable selection: the system assesses multiple forecasting methods—specifically RNN and Prophet—in key environmental variables, including internal temperature, humidity, and hive weight.

  • Data preparation: for each method-variable pair, the input data are preprocessed to ensure quality, involving normalization procedures and the handling of missing or inconsistent values.

  • Model initialization and evaluation: an initial predictive model is built for each configuration. If early evaluation metrics are satisfactory, the model is refined further; otherwise, the process shifts to alternative techniques.

  • Optimization and validation: promising models undergo hyperparameter tuning and are validated against unseen data. Forecast accuracy is assessed by comparing actual versus predicted trends, often supported by visual inspection.

  • Model selection and storage: if optimization yields improved results, the improved model is retained for future deployment; otherwise, the best available baseline is selected. Models that fail to meet performance thresholds are discarded in favor of other approaches.

The RNN models were implemented using the Huber loss function and a fixed random seed of 2 to ensure reproducibility. No early-stopping criterion was applied. All models used Batch Normalization with a batch size of 32. We allow the user to specify their desired prediction horizon directly from the dashboard (see “The dashboard”).

This approach ensures a systematic evaluation of different forecasting methodologies while adapting to the complexities of hive behavior. The final models selected through this process will be integrated into the monitoring system, enabling accurate short-term predictions that support beekeepers in proactive decision-making.

Algorithm 1
figure a

Prediction model training algorithm.

Optimizing initial models with grid search

After identifying the most important hyperparameters for each technology (see Table 2), we set the initial values of these hyperparameters to build \(M_{0T}\), the initial model.

After evaluating the initial model, we adopted an optimization technique to select the optimal hyperparameters values defined with the aim of minimizing the value of the chosen error metric. To achieve this, we adopted the Grid Search technique.

It is a technique used in machine learning and model selection to find the optimal hyperparameters for a given model. Hyperparameters are parameters that are not learned during training but are defined prior to training and influence the learning process. The grid search consists of systematically testing different combinations of hyperparameter values to determine the set of values that gives the best model performance.

The term “Grid Search” refers to the process of creating a grid of hyperparameter values and exhaustively searching all possible combinations of these values to find the best combination. Each combination is evaluated using an error metric selected from the validation DB. The combination that produces the best performance is then selected as the optimal set of hyperparameters for the model.

Table 2 summarizes the hyperparameters chosen for each technology and the optimal values possible for each.

Table 2 List of hyper-parameters to be manipulated in grid search.

Table 3 presents the optimal parameters selected following a series of validation experiments, with final values chosen for each RNN model according to the specific characteristics of each parameter(weight, humidity and temperature).

Table 3 Final chosen values for each parameter-specific RNN model.

Model results during the validation period

Table 4 shows the evaluation of the initial models \(M_{0}\) and the results after optimization of the humidity, weight, and temperature models. As mentioned in “Designing intelligent time series processing”, we use the RNN and Facebook prophet architectures. The database used here is the validation database of the Fig. 5.

We note that the initial temperature model \(M_{0RNN}\) gave us satisfactory results that make its optimization unnecessary, especially as this may lead to model overlearning in the training data. In addition, the Grid Search technique produced a considerably improved version of our model for humidity and weight, as shown in Table 4.

For the Prophet-type of architecture, Grid Search gave a very significant optimization in the models initially designed, but none of them was ideally able to follow the local trend variation in values.

We note a significant difference between the prediction quality of the two versions, where the RNNs predictions seem more sophisticated and realistic. We will therefore adopt \(M_{0RNN}\) for the temperature and \(M_{optimRNN}\) for weight and humidity.

Table 4 RNN & PROPHET optimization.

We now present the accuracy of the forecasting systems developed using the test dataset. Table 5 and Fig. 7 summarize and compare the MAE of the three final models for the three metrics. The results show that the errors observed on the test dataset (originating from a different hive than the training data) are comparable to those obtained during validation. This provides a form of cross-hive validation and suggests that the models may exhibit a degree of scalability. However, we acknowledge that broader cross-site validation, involving hives from different geographic locations and environmental conditions, is necessary to fully assess the generalizability and robustness of the proposed system.

Fig. 7
figure 7

Forecast results in the test DB.

Table 5 Summary of MAE error rates for the three forecasting models.

Implementation of the anomaly detection system

This section explains the development of the four anomaly detection models using the algorithms introduced in “Approaches used to design anomaly detection”, and then presents their level of sensitivity as a function of their detection capability.

Regarding humidity and the training DB, we note that the predictive confidence level approach is the most sensitive algorithm (38.33%), while DBSCAN is the least sensitive (6.81%), and the others have moderate sensitivity (around 20–25%). For temperature, we note that the Isolation Forest approach is the most sensitive algorithm (20.06%), while DBSCAN is the least sensitive (0.28%), and the others have moderate sensitivity (around 8–16%). For weight, we note that the predictive confidence level approach is the most sensitive algorithm (34.16%), while DBSCAN is the least sensitive (1.83%), and the others have moderate sensitivity (around 21–25%).

Table 6 summarizes and compares the sensitivity rates of the four models across the two databases and shows that, in most cases, the results are relatively consistent. This suggests that the models may generalize reasonably well across similar hive contexts, although further validation is needed to confirm their robustness in more diverse settings.

Table 6 Sensitivity of anomaly detection models.

While most anomaly detection models demonstrated consistent sensitivity across training and test datasets, the Isolation Forest algorithm showed a significant disparity (21.83% vs. 66.51%), as highlighted in table 6. This inconsistency suggests a lack of generalization and potential overfitting to the training hive data. As a result, we excluded Isolation Forest from the final prototype. These findings emphasize the importance of hive-specific calibration when deploying anomaly detection models in real-world settings. In our current implementation, each model is trained on data from a single hive, which improves reliability but limits generalization. To address this, future work will explore more robust evaluation strategies, such as hive-specific cross-validation folds, to better assess model adaptability. We recommend to investigate transfer learning techniques (e.g., fine-tuning on subsets of hive data) to enhance model performance across heterogeneous apiaries. These directions aim to ensure that the system remains reliable and adaptable in diverse operational contexts.

The dashboard

Since connected beehives are not yet fed by bee colonies that can generate data close to reality, we have replaced connected beehives in our prototype with an API that will play the role of connected beehives by periodically and continuously broadcasting values for internal humidity, temperature, and weight that simulate real values.

To generate random numbers with the same distribution as the actual hive data, we used a function to generate Gaussian random numbers, i.e. with a normal distribution. This function will take as input the mean and standard deviation of the metric. The BeeViz interface (Figs. 8, 9 and 10) provides a visual dashboard for hive monitoring, displaying temperature, humidity, and weight data over a selected time range (e.g., May to September). Each graph includes historical values (solid blue lines), predictions (dashed blue lines), statistical profiles (yellow areas), and predictive confidence intervals (green areas). Orange warning icons highlight detected anomalies. Users can configure the view by selecting the hive, adjusting the prediction horizon and sensitivity, and toggling the display of predictions and anomalies (Figs. 9 and 10).

Fig. 8
figure 8

Dashboard: visualization interface.

Fig. 9
figure 9

Dashboard: prediction configuration interface.

Fig. 10
figure 10

Dashboard: anomaly configuration interface.

Since our predictive models are based on iterative forecasting rather than direct multi-step prediction, we allow the user to specify their desired prediction horizon directly from the dashboard. Based on this input, the system recursively generates forecasts day by day (predicting N + 1, then N + 2, etc.) until the user’s chosen horizon.

Software environment

To ensure reproducibility and modularity, the BeeViz system was developed using a multi-layered software stack supporting data preprocessing, model training, anomaly detection, and dashboard deployment.

Machine learning components were developed in Python, while the web services and dashboard relied on JavaScript technologies.

For AI model development, TensorFlow and Scikit-learn were used to implement and train the forecasting and anomaly detection models, with Pandas and Matplotlib supporting data preprocessing, time-series manipulation, and visualization. The trained models were deployed through Flask microservices, exposing REST APIs that enable real-time inference and anomaly monitoring.

On the application side, Node.js and Express.js handled backend APIs and communication with connected hives and the database. The web dashboard, implemented in React.js with Redux and ApexCharts, provides interactive visualization of sensor data, forecasts, and alerts.

Socket.IO was used to enable real-time bidirectional data streaming between the backend and the dashboard.

All data were stored in a MongoDB NoSQL database using a JSON-based schema, allowing efficient management of sensor records, predictions, and anomaly logs. The overall architecture ensures seamless integration between the AI services, database, and visualization components, enabling reliable real-time monitoring and decision support for connected beehives. The exact software components and versions used in the experiments are reported in Table 7.

Table 7 Software components and exact versions used in the experiments.

All components are open-source and publicly available. The system was trained and tested on a workstation equipped with an Intel Core i7-11800H CPU, 64 GB RAM, and an NVIDIA GeForce RTX 3080 GPU (8 GB VRAM).

Discussion and future research directions

Several avenues for future research could further enhance BeeViz’s capabilities. First, expanding the system to visualize, predict, and detect anomalies in additional metrics, such as CO2 emissions, which indicate the metabolic state of the colony, or flight activity and external temperature, which influence internal hive conditions, would provide a more comprehensive assessment of hive health.

Second, given that our prediction and anomaly detection models are designed for general applicability, a collaborative learning approach could be explored. Instead of relying solely on the HOBOS dataset, the system could continuously improve its machine learning models by incorporating real-time data from connected hives. This federated approach would allow for progressive refinement and adaptation of the models based on diverse environmental conditions and hive behaviors. Finally, a promising direction would be to analyze correlations between multiple hive parameters (e.g., humidity, temperature, and weight) to optimize predictive accuracy and anomaly detection. Understanding these interdependencies could lead to more robust and reliable insights for beekeepers30.

Beyond the scope of this study, several recent works offer promising directions to enhance the robustness and intelligence of monitoring systems. For instance, DeepAutoD 31 introduces a deep learning framework for unpacking protected Android applications, which could inspire secure data handling mechanisms in embedded hive sensors. Tan et al.32 leverage graph neural networks to infer BGP (Border Gateway Protocol) community attributes, an approach that could be adapted to model interactions across connected hives. Additionally, targeted adversarial example generation 33 and semantic-aware malware detection 34 suggest new ways to improve the resilience of predictive models against environmental noise. Research on community detection in graphs 35 and deep learning model security 36 further highlights the importance of designing adaptive, collaborative systems capable of evolving with real-world data. These contributions represent valuable foundations for future iterations of the BeeViz platform. By addressing these research directions, BeeViz could further strengthen data-driven hive management, contributing to more sustainable and informed beekeeping practices.

While the current system uses hive-specific models to ensure reliability, future work will explore collaborative learning and transfer learning strategies to improve generalization across diverse environments.

Conclusion

This paper presents BeeViz, a monitoring system designed to support precision beekeeping through continuous data collection, visualization, and intelligent analysis. The system ensures seamless communication with connected hives, enabling real-time acquisition and storage of key hive parameters such as internal temperature, humidity, and weight. Beyond simple data monitoring, BeeViz integrates predictive modeling and anomaly detection techniques to improve hive management. A comprehensive review of existing monitoring approaches has been conducted, highlighting how collected data can be leveraged for decision support in apiculture.

BeeViz generates web-based visualizations, allowing beekeepers to track hive dynamics over time and detect early warning signs of potential problems. In addition, the system incorporates advanced time-series analysis for anomaly detection and short-term forecasting, empowering beekeepers with actionable insights to anticipate and mitigate risks. By integrating intelligent analytics, BeeViz contributes to more efficient and proactive hive management.