Introduction

The overconsumption of traditional energy resources is leading to their depletion, which in turn contributes to increased pollution and global warming. Consequently, the depletion of these resources and the resulting pollution have created a growing demand for renewable energy sources1,2. Organic photovoltaic (OPV) cells have gained significant attention in recent years due to their potential for producing cost-effective, lightweight, and flexible solar energy solutions. Despite achieving power conversion efficiencies (PCE) exceeding 18% in laboratory settings3,4, challenges such as stability and long-term degradation continue to hinder their widespread adoption. Among the various alternatives, organic solar cells stand out as a promising option due to their affordability and flexibility. In particular, inverted organic solar cells (iOSCs) are of special interest because of their enhanced stability, especially when fabricated using materials like PTB7-Th as the active layer. However, the PCE of iOSCs tends to degrade over time, underscoring the need for accurate predictive models to forecast their performance over extended periods5,6.

As a third-generation photovoltaic technology, organic solar cells utilize an active layer composed of organic molecules with a π conjugated structure. When exposed to sunlight, this layer generates excitons (electron–hole pairs), which are subsequently separated into free electrons and holes, creating an electrical current7,8. Although the PCE — defined as the ratio of electrical power delivered by a solar cell to the incident solar energy — of organic solar cells currently lags behind that of inorganic cells, researchers have made significant progress since the discovery of plastic solar cells in 19868 and6. Various structural designs and materials are being explored to enhance their efficiency.

The problem of degradation in inverted solar cells is a major challenge that affects their long-term performance10. Exposure to environmental factors such as humidity and high temperatures leads to the deterioration of materials used in these cells, resulting in a decline in their efficiency11. Therefore, predicting the degradation of these cells is crucial, as it helps identify weaknesses and improve cell design, contributing to enhanced performance and sustainability12. When discussing the PCE of solar cells, it is considered one of the most important characteristics in evaluating the performance of these cells13. Predicting PCE is vital for the development and optimization of this technology14. As organic solar cells develop, predicting their PCE is essential for comparing them with inorganic cells and tracking their progress. Unlike parameters such as short-circuit current (Jsc) and open-circuit voltage (Voc), which require extensive experimental data for accurate modeling, PCE can guide improvement efforts more directly, saving time and resources. PCE is calculated using a formula that includes Jsc, Voc, fill factor (FF), and incident light power8. Traditional methods for predicting the performance of organic solar cells involve evaluating the electrical and structural characteristics of the devices and simulating their current density–voltage (J-V) curves15. While simple equivalent circuit models can aid in comparative analysis, they do not offer detailed insights into the nuanced performance characteristics of the devices16. These traditional methods are also often time-consuming and costly, as they require specific parameters to achieve accurate predictions5. Although analytical models and more complex simulations provide a deeper understanding of charge transport mechanisms, they are challenging to implement due to the difficulties in obtaining the necessary parameters17.

Given the remarkable success of Machine Learning (ML) techniques in various engineering fields— such as medical imaging18, solar energy prediction19, traffic flow prediction20, financial market predictions21, energy demand forecasting22, and solar cell efficiency prediction23—we aimed to address these challenges by initially leveraging the conventional double exponential model, commonly used to simulate the degradation of organic solar cells24,25. Specifically, we proposed optimizing the double exponential model using the Levenberg–Marquardt algorithm, rather than the Nonlinear Least Squares optimization typically used in the literature. Additionally, we introduced four ML-based techniques, namely LSTM, MLP, Support Vector Machine (SVM) and Random Forest (RF), for monitoring the degradation of iOSCs through the prediction of the PCE. Our goal is to predict the quality of iOSCs by selecting certain experimental points and attempting to predict the remaining points, thus enhancing the efficiency and effectiveness of the organic solar cell technology. The results show the high accuracy of our proposed technique, as evidenced by the calculation of the error between the predicted and experimentally measured PCE values.

Methodology

Many researchers focus on studying, developing, and optimizing the structure of organic solar cells before they are manufactured. They then subject these cells to various influencing factors to evaluate their performance characteristics, efficiency, and lifespan. In addition, they compile comprehensive databases for use in various scientific studies and research.

In this work, we have used a database obtained from an experimental work by Sacramento et al.10 studying the stability and the degradation analysis of iOSCs formed by PTB7-Th: PC70BM Blend with Different Stacks of Electron Transport Layers (ETLs). These iOSCs were fabricated in a nitrogen (N2) environment.

Device structure and fabrication

Figure 1 shows the proposed device structure for the iOSCs studied by A. Sacramento et al.10, which is composed of the following layers: (ITO)/(ETL)/(PTB7-Th:PC70BM)/(V2O5)/(Ag). Five types of configurations for the ETLs have been studied: PFN/LiF (type I), LiF/PFN (type II), PFN (type III), LiF (type IV), and without an ETL (type V). The iOSCs were fabricated, and their J-V characteristics were measured as reported elsewhere10,12,26,27.

Fig. 1
figure 1

Device structure of iOSCs.

Five groups of iOSCs were formed to study the behavior of their parameters over time, which were:

  • Group I: ITO/PFN/LiF/PTB7-Th: PC70BM / V2O5/Ag,

  • Group II: ITO/LiF/PFN/PTB7-Th: PC70BM / V2O5/Ag,

  • Group III: ITO/PFN/PTB7 Th: PC70BM/ V2O5/Ag,

  • Group IV: ITO/LiF/PTB7-Th: PC70BM/ V2O5/Ag,

  • Group V: ITO/PTB7-Th: PC70BM/V2O5/Ag.

Layers description

The different layers involved in the iOSC structure are described as follows:

  • ITO (indium tin oxide): Typically used as a transparent conductive electrode due to its high transparency and conductivity. It’s often used in displays, solar cells, and other optoelectronic devices.

  • PFN(Poly[(9.9-bis(3’-(N,N-dimethylaminto)propyl)-2,7-fluorene)-alt-2,7-(9,9-dioctylfluorene)]): A polymer used as an electron-blocking layer or hole-transporting layer in organic electronic devices. It helps improve the device’s efficiency and performance.

  • LiF (lithium fluoride): is another material used as an ETL in organic solar cells, often in combination with other layers. The search results show that the performance of OPV cells with LiF as the sole ETL is not as good as when LiF is used in an ETL stack, such as (LiF/PFN) or (PFN/LiF)28.

  • PTB7-TH: PC70BM: A blend of organic semiconductors commonly used in organic solar cells. PTB7-Th is a polymer donor material, while PC70BM is a fullerene acceptor material. Together, they form the active layer responsible for converting light into electricity.

  • HTL (hole-transport layer): This layer facilitates the transport of positive charges (holes)from the active layer to the anode. This configuration uses V2O5 (Vanadium Pentoxide) as the HTL, serving as a control/reference structure.

  • Ag (silver): Often used as a reflective electrode or cathode in organic electronic devices. In this configuration, it’s likely serving as the cathode, collecting electrons generated in the active layer.

Optimization of layers structure in OSCs

The optimization of the ETL, Hole Transport Layer (HTL), and the active layer is crucial for enhancing the performance and longevity of organic solar cells (OSCs). ETL materials’ choice and stacking order significantly influence the stability and efficiency of OSCs. Studies show that using Poly[(9,9-bis(3’-(N,N-dimethylamino)propyl)-2,7-fluorene)-alt-2,7-(9,9-dioctylfluorene)] (PFN) in combination with a thin layer of Lithium Fluoride (LiF) improves stability compared to other configurations, with the PFN/LiF stack providing the best long-term performance29. For instance, cells with this stack exhibited a TS80 of 201 h, demonstrating the importance of optimizing ETL structures for durability and efficiency30.

The HTL, such as Vanadium Pentoxide (V2O5), plays a vital role in facilitating hole transport from the active layer to the anode31. By establishing a reference structure with V2O5, researchers can evaluate the impact of various ETL configurations on the performance of iOSCs32. Optimizing the HTL alongside the ETL is essential for achieving long-term stability and performance33.

The active layer’s morphology, thickness, and material composition are critical for maximizing charge separation and transport34. Fine-tuning these parameters can lead to significant improvements in PCE35. For example, a well-defined fibril network morphology can enhance charge transport and reduce recombination losses, achieving PCEs as high as 16.5%36. Additionally, controlling the active layer thickness is vital, as it affects key electrical parameters like short-circuit current density (Jsc) and open-circuit voltage (Voc)12.

Overall, careful optimization of the ETL, HTL, and active layers is essential for enhancing the efficiency, stability, and overall performance of iOSCs, ultimately contributing to their practical application and longevity.

Power conversion efficiency

The PCE is one of the most important parameters for evaluating the performance of solar cells37. It provides a measure of the electrical power that can be generated by a solar cell compared to the incident light power3. The PCE is generally expressed as a percentage and is calculated using the formula38:

$$PCE=\frac{{J}_{sc}*{V}_{oc}*FF}{{\text{P}}_{\text{in}}}$$
(1)

where Pin is the power of the incident light39.

Alternatively, the PCE can be determined directly from the current density–voltage (J-V) characteristics of the solar cell, taking into account the measured Voc ,Jsc and FF40. This equation highlights that all three parameters (Voc,Jsc, and FF) can limit the overall device performance38. Optimizing these key parameters through careful design and engineering of the solar cell structure and materials is crucial for achieving high-efficiency photovoltaic devices17.

Figure 2 illustrates the degradation of the (PCE) of (iOSCs), Groupe 1: (ITO/PFN/Lif/PTB7-Th: PC70BM/V2O5/Ag) over time. Initially, the PCE exhibits a rapid decrease during the first few hundred hours of operation, a phenomenon known as "burn-in"41. This initial decay is attributed to photochemical reactions occurring between the various layers of the cell42. Following this "burn-in" phase, the PCE experiences a slower exponential decay, which aligns with observations reported in the literature for organic solar cells based on the PTB7-Th:PC70BM blend11 Over time, PCE degradation poses a significant challenge to the commercialization of iOSCs. Understanding and minimizing degradation mechanisms, particularly during the "burn-in" phase, is essential for enhancing the long-term stability of these photovoltaic devices.

Fig. 2
figure 2

Degradation of the PCE over time for group I.

Degradation mechanisms in inverted organic solar cells

Degradation refers to the decline in the performance of organic solar cells beyond the initial "burn-in" phase41. This decline can be caused by environmental factors such as oxidation and humidity, morphological changes in the active and transport layers, and interactions between the different layers of the device43. In summary, while "burn-in" represents an initial degradation phenomenon, overall degradation denotes the long-term performance decline of iOSCs42. Optimizing the structure of the device layers is crucial for mitigating both types of degradation44.

Model development

The present work is divided into two main contributions. First, we propose a double exponential model based on Levenberg–Marquardt algorithm rather than the simple Nonlinear Least Squares model used by Sacramento et al.10 to predict the PCE of iOSC. Second, we introduce four different ML techniques namely MLP, LSTM, SVM and RF to predict the PCE of iOSC. In the following subsections, we provide a theoretical background of these proposed methods.

Multi-layer perceptron

An MLP is a non-dynamic kind of artificial neural networks that is commonly used in ML for supervised learning problems, including fitting, prediction and classification problems. As can be seen from Fig. 3, the MLP consists of three main layers: an input layer, a hidden layer and an output layer. The input layer receives the 0external excitation signals and fore propagates them to the hidden layer neurons. The hidden layer consists of a given number of hidden layers, with given number of neurons in each, that act as the MLP’s actual processing engine. The output layer generates a decision or prediction corresponding to the input signals. The output of each neuron is updated according to Eq. (3):

$${y}_{j}=f\left(\left({\sum }_{i}{w}_{ij}*{X}_{ij}\right)-{b}_{j}\right)$$
(2)

where \({y}_{j}\) is the jth neuron output. \({X}_{ij}\) is the excitation coming from the neurons of the previous layer i. \({w}_{ij}\) is the weights linked the neurons of the previous layer i to the neuron j. \({b}_{j}\) is the bias. \(f\) is the activation function. It can be a sigmoid, a hyperbolic tangent or a rectified linear unit function. Scientists often apply MLP to supervised learning tasks using training on a set of input–output pairs helping to extract the relation between those inputs and outputs. The training process involves adjusting the weights and biases of the model to minimize the error between computed and desired outputs. Different optimization algorithms like gradient descent or its variants along with an activation function are used to adjust weights and bias relative to the error.

Fig. 3
figure 3

Generic architecture of MLP.

In this study, we have chosen to work on the prediction of the PCE of iOSCs based on MLP as it has shown good prediction capabilities for complex systems and patterns due to its ability to learn and store knowledge (in the form of weights) for future predictions45. MLPs are powerful models that have been and still are used for a wide range of engineering tasks21 including solar energy applications22,46,47.

Long short-term memory

LSTM is a kind of dynamic ANNs. It belongs to the family of recurrent neural networks (RNN) initially designed to deal with time-dependent tasks48. LSTM was proposed by Hochreiter and Schmidhuber in 1997 essentially to overcome the limitations of learning long-term dependencies based on traditional RNNs due to the vanishing and exploding gradient problems usually accompany the well-known Back Propagation Through Time (BPTT) learning algorithm49. LSTMs can learn and remember over extended sequences through their sophisticated memory cells. As shown by Fig. 4, the memory cell is the core of the LSTM. It maintains a cell state ct updatable via a series of operations that selectively add or remove information. LSTMs has also special neural units that regulate the information flow within the cell called gates. LSTMs employ several kinds of gates. Forget Gates ft select the information to be discarded from the cell state. Input Gates it updates the cell state by injecting new information. Output Gates ot select the information from the cell state to be used for the output production. Activation functions like the sigmoid and hyperbolic tangent are typically used by LSTM to control the information flow and to update the state of the memory cell.

Fig. 4
figure 4

Generic architecture of LSTM50.

The LSTM architecture can capture long-term dependencies in input data sequences via the ability to keep relevant information for long periods of time. This property renders LSTMs a popular solution in many engineering tasks where learning long-term dependencies is essential such as natural language processing51, time series prediction52, solar energy53,54, to name but a few.

Support vector machine

Support Vector Machine (SVM) is a robust algorithm frequently used to tackle non-linear problems, even when the dataset size is small. It is applicable to both classification and regression tasks, with the regression variant referred to as Support Vector Regression (SVR).

For a dataset containing N samples of input–output pairs:

\(D=\left\{\left({x}_{1},{y}_{1}\right),\left({x}_{2},{y}_{2}\right),\dots ,\left({x}_{N},{y}_{N}\right)\right\}\epsilon {R}^{k}xR\). The regression function can be described as: \(f({X}_{i})={W}^{T}.{X}_{i}+b, w\in {R}^{K }\) (3).

  1. o

    W is the normal vector to the hyperplane,

  2. o

    b is the bias term (the distance of the hyperplane from the origin),

  3. o

    \({x}_{i}\)​ is the input vector.

SVM for non-linear problems uses a non-linear mapping function that transforms the input data into a higher-dimensional feature space, making it easier to separate the data linearly in that space. The optimization problem for SVM is:

$$\text{min}\left(\frac{1}{2}{\Vert w\Vert }^{2}+C{\sum }_{i=1}^{M}\tau i\right)$$
(4)

subjects to boundary conditions

$${y}_{i}\left({w}^{T}.\varphi \left({x}_{i}\right)+b\right)\ge 1-\tau i, \tau i\ge 0$$
(5)

where:

  • \(\tau\)i ​ is a slack variable allowing some misclassification,

  • C is the regularization parameter that controls the trade-off between fitting the training data and maintaining model simplicity. A larger C tends to increase model accuracy on training data but may cause overfitting, while a smaller C results in a simpler model that might misclassify some points.

The non-linear transformation φ (xi) can be replaced with kernel functions, such as K (xi, xj), which compute the necessary values in the input space, making SVM computationally efficient for non-linear problems.

Random forests

The RF is a highly adaptable supervised ML algorithm, introduced by Leo Breiman and Adele Cutler in 200155. Its popularity stems from its ease of use and flexibility, which have led to widespread adoption in both classification and regression tasks. At its core, RF leverages ensemble learning, a powerful technique where multiple models (in this case, decision trees (DT)) are combined to solve complex problems and enhance model performance.

While DTs are a fundamental component of many supervised learning algorithms, they are often susceptible to issues such as bias and overfitting, where the model becomes too tailored to the training data and struggles with generalization. However, when DTs are aggregated into an ensemble using the Random Forest method, the algorithm mitigates these issues. By utilizing a large number of decision trees that are uncorrelated with one another, the Random Forest algorithm achieves significantly more accurate and reliable predictions.

A flow chart of the RF is shown in Fig. 5. Each DT in the forest is built using a random subset of the dataset and a random selection of features at each split. This randomization injects variability into the individual trees, making the model less prone to overfitting and improving its predictive power. During the prediction phase, RF combines the results of all trees through voting (in classification tasks) or averaging (in regression tasks). This collaborative approach allows the algorithm to arrive at stable, precise predictions by drawing from the collective wisdom of multiple trees.

Fig. 5
figure 5

The flowchart of random forest.

To effectively train a RF model, there are three key hyperparameters that need to be set: node size, which controls the depth of the trees; the number of trees, which determines how many trees will be included in the forest; and the number of features sampled, which dictates how many features are considered at each split. Properly tuning these parameters is critical for optimizing the model’s performance.

Random Forests are highly regarded for their ability to handle complex datasets, effectively reduce overfitting, and provide robust predictions across various environments. They are also capable of dealing with missing data and scaling to large datasets. Another key advantage is their ability to determine feature importance, offering insights into which variables contribute most to the model’s predictions.

These strengths make Random Forests applicable across a wide range of fields. In finance, they are used for tasks like customer evaluation and fraud detection56. In healthcare, they are employed in areas such as computational biology, biomarker discovery, and sequence annotation57, demonstrating their versatility across diverse engineering and scientific applications.

Double exponential model

The double exponential model, often used in curve fitting and data modeling, can be expressed as:

$$f\left(x\right)=A{e}^{-Bx}+C{e}^{-Dx}$$
(6)

where: A, B, C, and D are parameters to be determined through fitting, x represents the independent variable and e denotes the base of the natural logarithm. Figure 6 shows an example of this model for different values of A, B, C, and D parameters.

Fig. 6
figure 6

The double exponential model for different values of A, B, C and D parameters.

Fitting the double exponential model to a given data involves finding the best values for these parameters that minimize the difference between the model’s predictions and the actual data points.

A general approach to fit the double exponential model to a given data consists of:

  • Initial Parameter Estimation: One might start with random initialization of the parameters A, B, C, and D to begin the fitting process. The initialization can be also based on the data or any prior knowledge of the system.

  • Curve Fitting Algorithm: A curve fitting algorithm or optimization technique is used to adjust the parameters of the double exponential model. Popular optimization methods include: Nonlinear Least Squares, Gradient Descent or Levenberg–Marquardt Algorithm24.

Experimental setup

Used dataset

Our study focused on leveraging the experimental data obtained by Sacramento et al.10 by examining the degradation of inverted organic solar cells (iOSCs) using a PFN/LiF stacking arrangement that showed improved cell stability. Key performance parameters were evaluated, including open circuit voltage (Voc), short circuit current density (Jsc), fill factor (FF), and power conversion efficiency (PCE). These parameters were extracted from current density–voltage (J-V) measurements conducted under both dark and light conditions over a period of 7,110 h10. The measurements were performed at room temperature (23°C) with an average relative humidity (RH) of 54%, using a Keithley 2450 source measurement unit. The samples were tested using a solar simulator under the AM 1.5G spectrum, with a light intensity of 100 mW/cm2 (1 sun), calibrated with a Solarex Corporation-certified monocrystalline silicon photodiode58.

As shown in Table 1, the dataset is organized in a tabular format with columns representing the performance parameters of iOSCs over time during degradation testing. This comprehensive dataset enables a thorough analysis of the degradation mechanisms affecting the solar cells’ long-term performance. The data spans a wide timeframe, allowing for detailed tracking of how each parameter evolves during the degradation process.

Table 1 Performance parameters of (iOSCs) over time during degradation testing.

Figure 7 shows the normalized performance parameters for iOSCs with respect to their initial values as a function of time. These parameters include Jsc, Voc, FF, and PCE. The normalization allows for a clearer comparison of the degradation rates of each parameter over time, highlighting the stability and long-term performance of the iOSCs. The data demonstrate the varying rates of degradation, crucial for understanding and improving the longevity and efficiency of iOSCs. In our study, we focus particularly on the degradation of the time-dependent power conversion efficiency (PCE), which is critical for evaluating the performance of organic solar cells over time.

Fig. 7
figure 7

Normalized performance parameters for (iOSCs), regarding their initial values as a function of time.

It turns out that in our study, we are interested in the degradation of the PCE due to its significance in determining the efficiency of organic solar cells. It is important to note that the dataset is complete and representative, as it was collected under controlled climatic conditions, including constant temperature, humidity, and radiation. This ensures that the degradation observed in the data is purely due to time and internal factors rather than external environmental fluctuations, allowing for more accurate modeling and analysis of iOSCs’ long-term performance.

Data pre-processing

In order to improve the proposed model’s learning process and performance, we use two main preprocessing operations in this study. First, we handled missing values by fitting the original PCE using a double exponential model, and then imputing the missing points, allowing the model to work with a complete dataset and avoid biases caused by missing data points. Second, we applied data scaling to ensure that all values are on a comparable scale, which is highly beneficial for ML training. Details of these preprocessing operations are presented in this section.

Data imputation

In our experiments, the original dataset consisted of 58 measured PCE values over 7100 h, but there were several significant gaps in the measurements. One of the most critical gaps occurred between the 2,500-h mark and the 6500-h mark, leaving a missing interval of approximately 4,000 h. This absence of data could have introduced uncertainty into the analysis and compromised the accuracy of any conclusions drawn regarding the long-term degradation of the iOSC devices.

To address the issue of missing data, several techniques are typically used, such as linear interpolation, extrapolation, mean or median imputation, and forward/backward fill, among others. However, in our work, we employed an imputation technique based on fitting the original data to a double exponential model, which closely matched the observed degradation trend, as will be discussed in Section "Data fitting". By applying this model, we were able to generate 102 additional data points within the 4,000-h gap and other smaller gaps in the dataset. This allowed us to increase our data to 160 values and to maintain a continuous dataset while ensuring the imputed points followed the same degradation behavior as the original observations.

Figure 8 illustrates the original and imputed data, along with the double exponential fit. The imputation method allowed us to bridge the gaps in the dataset, particularly in the critical range between 2,500 and 6,500 h, ensuring the continuity of the data for subsequent analysis and modeling. This approach enhanced the robustness of our analysis, making it possible to better evaluate the long-term stability and performance of the iOSCs.

Fig. 8
figure 8

Imputation of the missing PCE data.

Data normalisation

To prepare the data for analysis and model training, the PCE values are scaled between 0 and 1 using the following formula:

$$PC{E}_{Normalized}=\frac{PCE-\text{ min}(\text{PCE})}{\text{max} \left(PCE\right)-\text{min}(PCE)}$$
(7)

where PCE and \(PC{E}_{Normalized}\) represent the original and the normalized values of the PCE, respectively. The \(max\) and \(min\) operators are the maximum and minimum PCE values in the within the dataset.

Scaling the PCE values ensures that all values are on a comparable scale, which is important for machine learning-based methods. Additionally, scaling can improve the convergence speed of optimization algorithms by reducing the likelihood of skewed gradients, leading to more efficient ML training processes.

Training process

Our leveraged dataset contains 160 PCE values recorded over approximately 7,100 h, with irregular time steps. To ensure a fair evaluation of the ML techniques used for predicting future PCE values based on the learned history of PCE degradation, we applied a cross-validation approach based on the rolling block technique. The rolling technique, also known as a rolling or sliding window technique, is commonly used in time-series analysis to train models. It involves creating a dynamic or shifting window of data, where each window contains a subset of sequential observations. The model is trained on the data within this window, and predictions are made for subsequent points. After making predictions, the window shifts forward by one or more-time steps, and the process repeats.

This method offers several advantages, particularly for time-dependent data like PCE values. First, it helps maintain the temporal structure of the dataset by ensuring that training is done on sequential data, which is essential for capturing degradation patterns over time. Additionally, the rolling technique allows the model to learn from different portions of the dataset while avoiding data leakage between training and testing sets. It also ensures that the model is tested on unseen data, which is crucial for evaluating the model’s generalization ability.

In our experiments, we used training blocks of 80 values and test blocks of 20 values. Table 2 illustrates the various windows used for training and testing across the four machine learning techniques applied in this study.

Table 2 The division of the data into training and test sets.

Performance evaluation

To evaluate the performance of our proposed models for the PCE prediction of the iOSCs degradation, we have used an objective evaluation based on various statistical metrics to measure the accuracy and reliability of the model’s predictions. These metrics include Mean Squared Error (MSE), Root Mean Squared Error (rMSE), Relative Root Mean Squared Error (rrMSE), Relative Absolute Error (RAE), Residual Standard Error (RSE), Residual Sum of Squares (RSS), and the Coefficient of Determination (R2). Below is an explanation of each metric59,60,61,62,63:

Mean squared error (MSE): The average of the squared differences between the actual and predicted values.

$$MSE=\frac{1}{N}\sum_{i=1}^{N}{\left({y}_{i}-\widehat{y}\right)}^{2}$$
(8)

Lower values of MSE indicate better model performance. This metric is sensitive to outliers due to the squaring of errors.

Root mean squared error (rMSE): The square root of the mean squared error:

$$rMSE=\sqrt{MSE}=\sqrt{\frac{1}{N}\sum_{i=1}^{N}{\left({y}_{i}-\widehat{y}\right)}^{2}}$$
(9)

It represents the standard deviation of the residuals, making it easier to interpret as it is in the same units as the target variable.

Relative root mean squared error (rrMSE): rMSE normalized by the mean of the actual values:

$$rrMSE=\frac{rMSE}{\overline{y} }$$
(10)

It provides insight into how large the residuals are relative to the magnitude of the actual values, offering a normalized view of prediction error.

Relative absolute error (RAE): Measures the sum of absolute errors relative to the sum of absolute errors of a baseline model (such as the mean of actual values):

$$RAE=\frac{\sum_{i=1}^{N}\left|{y}_{i}-\widehat{{y}_{i}}\right|}{\sum_{i=1}^{N}\left|{y}_{i}-\overline{y }\right|}$$
(11)

It indicates the proportion of error in the model compared to a baseline model. Lower values of RAE suggest better model performance.

Residual standard error (RSE): Estimates the standard deviation of the residuals from the model, It is defined by the following formula:

$$RSE=\frac{\sum_{i=1}^{N}{\left({y}_{i}-\widehat{{y}_{i}}\right)}^{2}}{\sum_{i=1}^{N}{\left({y}_{i}-\overline{y }\right)}^{2}}$$
(12)

It Reflects the model’s goodness of fit; lower values indicate a better fit of the model to the data.

Residual sum of squares (RSS): The sum of the squared residuals, representing the total deviation of the predictions from the actual values:

$$RSS=\sum_{i=1}^{N}{\left({y}_{i}-\widehat{{y}_{i}}\right)}^{2}$$
(13)

It provides a measure of the total error in the model’s predictions. Lower values indicate a better fit.

Coefficient of determination (R2): The proportion of variance in the dependent variable that is predictable from the independent variables:

$$R^{2} = 1 - \frac{{\sum\nolimits_{i = 1}^{N} {\left( {y_{i} - \widehat{{y_{i} }}} \right)^{2} } }}{{\sum\nolimits_{i = 1}^{N} {\left( {y_{i} - \overline{y}} \right)^{2} } }}$$
(14)

R2 Ranges from 0 to 1, with higher values indicating a better fit of the model to the data. An R2 of 1 signifies a perfect fit.

Results

In this section, we present the results obtained from this study. First, we begin by discussing the results of data fitting achieved using our leveraged double exponential model. Following that, we will present the findings from applying four different machine learning techniques—namely, MLP, LSTM, SVM, and RF to predict the PCE of the studied iOSC structure.

Data fitting

According to Sacramento et al. 10, the degradation of the PCE over the time follows a mathematical model based on two-term exponential function:

$$\frac{PCE\left(t\right)}{PCE\left(0\right)}={A}_{1}{e}^{\frac{-t}{{T}_{1}}}+{A}_{2}{e}^{\frac{-t}{{T}_{2}}}$$
(15)

where PCE(0) is the PCE at time t = 0 h, t is the time, (T1, T2) are time constants and (A1, A2) are weights of the degradation. This model for PCE degradation in solar systems is widely accepted in the literature24,25, The parameters (A1, A2) and (T1, T2) are typically determined by minimizing the least squares error between the measured and modeled PCE values based on this double exponential model.

For instance, Sacramento et al.10 optimized the time constants and degradation weights for the data presented in Section "Used dataset", yielding the following values: A1 = 0.45, A2 = 1- A1 = 0.55, T1 = 220 h and T2 = 6000 h. In this work, we utilized the same model but employed the Levenberg–Marquardt algorithm to optimize the double exponential parameters, as opposed to the Nonlinear Least Squares method used by Sacramento et al.10. Our optimization resulted in the following parameter values: A1 = 3.9510, A2 = 2.6580, T1 = 529.3806 and T2 = 1637.2.

Figure 9 compares the measured PCE data with the predictions made using the double exponential models proposed by Sacramento et al. and in this study. Overall, both models fit the measured PCE data well. However, it is evident from the figure that our proposed model, based on the Levenberg–Marquardt optimization, more closely aligns with the measured PCE data compared to the model by Sacramento et al., which used the Nonlinear Least Squares method. Another observation is that in both cases, the surface of the double exponential model is smooth and cannot capture the fluctuations in the measured PCE data.

Fig. 9
figure 9

Fitting of the PCE over the time based on the double exponential model (a) PCE in function of time(b) Normalized PCE in function of logarithmic scaled time.

Data prediction

Hyper parameters setting

In this section, we analyze the impact of hyperparameter configurations on the performance of various ML architectures, highlighting the importance of hyper parameters like architecture size and training algorithms in achieving optimal predictive accuracy.

MLP

Figure 10 shows the performance of MLP using the Levenberg–Marquardt (LM) training algorithm across different architectures. For single-layer architectures, the best performance occurs with 5 neurons, achieving an MSE of 0.082 and a high correlation coefficient (r) of 0.967. As the number of neurons increases, performance declines, with 50 neurons showing a significant drop (MSE = 0.831, r = 0.894) and the largest architectures (100 and 200 neurons) performing poorly (MSE > 1.4, r < 0.73).

Fig. 10
figure 10

MLP performance in function of different architectures for Levenberg Marquardt training algorithm.

For multi-layer architectures, the architecture with two hidden layers, five neurons each (5,5) performs the best (MSE = 0.072, r = 0.978), while larger architectures like50,50 result in reduced accuracy (MSE = 0.449, r = 0.878). This indicates that smaller, well-balanced architectures outperform larger, more complex ones due to the risk of overfitting.

Figure 11 illustrates the performance of MLP using the Bayesian Regularization Backpropagation (BR) training algorithm across various architectures. For single-layer architectures, the best performance is achieved with 5 neurons, yielding an MSE of 0.068 and a high correlation coefficient (r = 0.973). As the number of neurons increases, the performance worsens, with architectures of 100 and 200 neurons showing significant increases in MSE (1.445 and 11.416, respectively) and lower r values (< 0.61), suggesting overfitting or diminishing returns with larger architectures.

Fig. 11
figure 11

MLP performance in function of different architectures for Bayesian regularization backpropagation training algorithm.

In multi-layer architectures, the10,10 architecture performs the best (MSE = 0.074, r = 0.970), closely followed by5,5 neurons. However, larger architectures like 30,30 show a considerable drop in performance (MSE = 0.310, r = 0.901). This pattern indicates that smaller and moderate-sized architectures with the BR algorithm provide more robust generalization and better accuracy, while larger models may struggle with overfitting or poor convergence.

Figure 12 illustrates the performance of the MLP using the Gradient Descent (GD) training algorithm for various architectures. For single-layer architectures, performance generally declines as the number of neurons increases. The best performance is observed with 5 neurons, yielding a relatively low MSE of 0.129 and a high correlation coefficient (r = 0.973). However, as the number of neurons increases, the MSE significantly worsens, with 100 and 200 neurons showing MSE values of 4.06 and 8.28, respectively, accompanied by a substantial drop in correlation (r < 0.65), indicating poor model generalization and potential overfitting.

Fig. 12
figure 12

MLP performance in function of different architectures for Gradient descent.

For multi-layer architectures, a similar trend is observed. Architectures with5,5 and10,10 neurons perform relatively well, with low MSE (0.192 and 0.144) and high correlation (r = 0.934 and 0.965, respectively). However, larger architectures, such as50,50 and [100, 100], show significantly higher MSE and a sharp decline in correlation (r = 0.857 and 0.748), confirming the diminishing returns with increasing model complexity. Overall, this highlights that smaller architectures are more effective with Gradient Descent, while larger models struggle with convergence and generalization.

Figure 13 illustrates the performance of the MLP utilizing the Resilient Backpropagation (Rprop) training algorithm across various architectures. For single-layer configurations, the optimal performance is achieved with 5 neurons, yielding a low Mean Squared Error (MSE) of 0.095 and a high correlation coefficient (r = 0.977). As the number of neurons increases, performance deteriorates, with 50 neurons showing a significant increase in MSE (0.824) and a drop in correlation (r = 0.794). The largest architectures, specifically those with 100 and 200 neurons, demonstrate poor performance with MSE values exceeding 4.67 and 10.21, and correlation coefficients below 0.69.

Fig. 13
figure 13

MLP performance in function of different architectures for resilient backpropagation training algorithm.

In the case of multi-layer architectures, the configuration of5,5 neurons yields the best results (MSE = 0.095, r = 0.976), while larger setups, such as50,50, show reduced accuracy (MSE = 0.209, r = 0.928). These findings emphasize that smaller, balanced architectures are more effective with the Rprop algorithm, as they minimize the risk of overfitting compared to larger, more complex structures.

LSTM

Figure 14 presents the relationship between the number of MaxEpochs and three key performance metrics: MSE, RAE, and r. As the MaxEpochs increases, MSE and RAE decrease steadily until reaching their lowest values at 250 epochs, where the model achieves optimal performance with an MSE of 0.054 and an RAE of 0.072. The correlation coefficient (r) peaks at 250 epochs with a value of 0.983, reflecting the highest predictive accuracy. Beyond 250 epochs, MSE and RAE begin to rise, while r slightly decreases, indicating overfitting or diminishing returns in model training. This analysis demonstrates that 250 MaxEpochs offer the best balance between error reduction and model accuracy.

Fig. 14
figure 14

LSTM prediction performance in function of the number of training epochs.

Figure 15 illustrates the relationship between MSE, RAE, and correlation coefficient (r) across three data rows. MSE and RAE are plotted on a logarithmic scale, revealing significant differences in error across the rows. Row 1 achieves the lowest MSE (0.0541) and RAE (0.0729), coupled with the highest correlation (r = 0.9835), indicating optimal model performance. In contrast, Rows 2 and 3 show larger MSE and RAE values, with reduced correlation coefficients, reflecting a decline in model accuracy and efficiency. The semi-log scale highlights the disparities in error magnitude, demonstrating the clear superiority of Row 1 in terms of error minimization and correlation strength. This analysis justifies Row 1 as the most accurate model configuration.

Fig. 15
figure 15

LSTM prediction performance in function of the learning rate.

Figure 16 showcases the performance of the LSTM model as the number of hidden units increases. The MSE and RAE curves reflect a consistent decrease in error until 200 hidden units, where the model performs optimally (MSE = 0.0541, RAE = 0.0729). Beyond 200 hidden units, both metrics rise again, signaling over fitting. The correlation coefficient (r) reaches its maximum at 200 hidden units (r = 0.9835), further confirming this as the most efficient configuration. Thus, 200 hidden units offer the best balance between accuracy and model complexity.

Fig. 16
figure 16

LSTM prediction performance in function of the number of hidden units.

Figure 17 compares the performance of three optimizers—SGDM, RMSprop, and Adam—based on MSE, RAE, and R. Adam emerges as the top performer, achieving the lowest MSE (0.0541) and RAE (0.0729), and the highest correlation (R = 0.9835), indicating accurate predictions and minimal error. SGDM follows with a moderate MSE of 0.1727 and R of 0.9500, while RMSprop performs the worst, with an MSE of 0.3056 and R of 0.9386. These results highlight Adam’s superior convergence and efficiency in this context. The clear performance gap between optimizers emphasizes the importance of selecting the right training algorithm to maximize model accuracy and minimize errors.

Fig. 17
figure 17

LSTM prediction performance in function of the optimizer.

SVM

This sub-section studies the impact of SVM hyperparameters tunning on the performance of PCE prediction. As shown in Fig. 18, we examined the effect of Box Constraint and Kernel Scale on RMSE for different values of Epsilon. The results show that the performance of the model, in terms of prediction accuracy, is highly dependent on the interaction of these parameters. For an Epsilon of 0.001 and 0.01, the study finds that a higher Box Constraint of 10 and a Kernel Scale of 10 consistently yield the best performance, with the lowest RMSE values. Specifically, for Epsilon = 0.01, this combination produces an RMSE of 0.3227.

Fig. 18
figure 18

The impact of SVM hyperparameters tunning on the performance of PCE prediction.

In contrast, when Epsilon increases to 0.1, a moderate Box Constraint of 1 performs best, yielding the lowest RMSE of 0.3065. This result suggests that for larger Epsilon values, which introduce more tolerance for errors, a smaller Box Constraint is necessary to avoid overfitting. Overall, the results indicate that Box Constraint of 10 and Kernel Scale of 10 form the best architecture for smaller Epsilon values, while for larger Epsilon, adjusting the Box Constraint downwards improves the model’s generalisation and reduces errors.

The most optimal SVM architecture among all experiments corresponds to an RMSE of 0.3065, achieved with a Box Constraint and Kernel Scale of 1 for an Epsilon of 0.1. This configuration strikes an ideal balance between fitting the data closely and maintaining sufficient flexibility in the decision boundary, effectively preventing overfitting while ensuring high prediction accuracy.

RF

This sub-section investigates the impact of RF hyperparameter tuning on the performance of PCE prediction. As shown in Fig. 19, we examined the effect of the number of trees and minimum leaf size on RMSE for different values of the maximum number of splits.

Fig. 19
figure 19

The impact of RF hyper parameters tunning on the performance of PCE prediction.

Each panel of the figure corresponds to a value of maximum number of splits. It appears clearly that the three panels are similar regardless the value of this latter. Therefore, the maximum number of splits seems to do not have a big influence on the performance of the prediction. So, a small value of maximum number of splits are preferable as they decrease the complexity of the model. Increasing the number of trees leads to higher RMSE values, indicating that overfitting may occur with too many trees. For instance, for a maximum number of splits of 10, the lowest RMSE value of 0.3541 is observed with 5 trees and a minimum leaf size of 5. This suggests that small tree numbers combined with a lower leaf size yield better results, preventing the model from becoming too complex and improving generalization.

Across all experiments, the most optimal RF architecture corresponds to a configuration of 5 trees and a minimum leaf size of 5 for a maximum split of 5, yielding the lowest overall RMSE of 0.3492. This setup combines between low model complexity and high prediction accuracy, preventing overfitting while maintaining a high level of generalisation.

Comparative study

Figure 20 shows the fit of PCE values over time, comparing measured data against predictions from our four proposed ML models: LSTM, MLP, SVM, and RF. Each model’s predicted performance is compared against the actual measured PCE data, highlighting how well each ML technique can capture the degradation trends over time.

Fig. 20
figure 20

Fitting of the PCE over time based on the proposed ML models. (a) PCE in function of time (b) Normalized PCE in function of logarithmic scaled time.

The LSTM model follows the general trend of the measured PCE values closely throughout most of the time series. It captures both the initial sharp decline and the steady long-term degradation. While LSTM maintains a reasonable fit across the entire timeline, it slightly underestimates the values toward the later stages of degradation, particularly between 5,000 and 6,000 h.

The MLP model shows a similar fit to LSTM in the early stages (up to around 2,000 h). However, it slightly overestimates the PCE values in the mid-range (2,000 to 2,500 h). MLP’s performance declines more noticeably in the long-term prediction (after 6,000 h), where it fails to fully capture the rate of degradation.

The SVM model performs well during the early stage (up to 200 h), tracking the measured data relatively closely. However, it shows a notable overestimation in the early to mid-range (between 200 and 1,500 h) and then diverges slightly after 6,000 h, where it underestimates the PCE values.

The RF model diverges significantly from the measured data, particularly in the early stages of degradation (up to around 500 h), where it shows an oversimplified, flat prediction. RF is unable to capture the fluctuations in the data and performs the worst among the models, particularly in the early and mid-range, although it improves slightly in the later stages.

LSTM continues to be one of the best-performing models for time-series prediction in this study, particularly because it can handle sequential data and retain information over long periods. Its ability to model long-term dependencies makes it effective in capturing the overall degradation trend, although some fine-tuning might be necessary for more accurate long-term predictions.

The MLP model captures the general trend of the PCE degradation but struggles with accuracy over longer time intervals. MLP, being a simpler feedforward neural network, lacks the temporal dynamics handling capabilities of LSTM, which explains its weaker performance in the later stages.

SVM performs reasonably well during the initial stage but struggles with capturing the full complexity of the degradation in the later stages. This model, while effective in capturing some general patterns, may not be flexible enough for long-term time-series forecasting with fluctuating patterns.

RF, as expected, performs the worst in this context, particularly due to its limitations in capturing the sequential and temporal dynamics inherent in time-series data. It oversimplifies the degradation pattern and fails to adapt to the fluctuations in the measured data, especially in the early and mid-range.

Furthermore, Table 3 compares the measured and predicted PCE data for an iOSC cell using different ML models, including LSTM, MLP, SVM, and RF, across MSE, rMSE, rrMSE, RAE, RSE, RSS and R-squared evaluation metrics. The MSE and rMSE values indicate that LSTM provides the closest fit to the measured data, with the lowest error (MSE = 0.0541, rMSE = 0.2223) and the highest R-squared value of 0.9835. This suggests that LSTM effectively captures the long-term degradation pattern of the PCE values, making it the most accurate model. The RF, on the other hand, shows the highest error values (MSE = 0.1610, rMSE = 0.3723) and the lowest R-value of 0.9274, indicating that it struggles to generalize over the lifetime of the iOSC cell, likely due to its inability to capture the sequential nature of the data effectively.

Table 3 A comparison between the measured and the predicted PCE data of the studied iOSC cell using different ML-based prediction techniques over all the lifetime of the cell.

From a broader perspective, the results highlight the strengths of LSTM and MLP models in time-series predictions, where long-term dependencies and temporal trends are crucial, such as in the degradation behavior of PCE over time. The SVM and RF, although capable of capturing general trends, perform less favorably because they lack mechanisms to retain memory over long sequences. This suggests that for predicting the performance and degradation of OPV cells over their lifetime, models like LSTM that can model sequential dependencies are more effective compared to non-sequential techniques like SVM and RF, which are better suited for tasks with less temporal complexity.

Furthermore, Table 4 summarizes the 95% confidence intervals (CIs) and coverage probabilities of four LSTM, MLP, SVM, and RF models for predicting PCE in iOSCs. Each model’s performance is evaluated based on the mean lower bound, mean upper bound, mean CI width, and coverage probability.

Table 4 A comparison between LSTM, MLP, SVM, and RF Models for PCE Prediction in terms of Bootstrapped Confidence Intervals Statistics.

The LSTM model demonstrates the narrowest CI with a mean CI width of 0.5035. This indicates that LSTM’s predictions are relatively more confident compared to the other models. Despite this narrow width, LSTM maintains a high coverage probability of 96.27%, which means the model’s CIs capture almost all actual PCE values. This balance of narrow CIs and high coverage suggests that LSTM is well-suited to making precise PCE predictions while maintaining high reliability. The LSTM model’s ability to capture temporal dependencies in the data likely contributes to this strong performance, particularly in modeling sequential data.

The MLP and SVM models show a slightly wider mean CI width of 0.5919 and 0.5916, respectively, compared to LSTM. Their coverage probabilities of 95.65% and 95.03% are also slightly lower but still respectable, indicating that MLP and SVM remain reliable models for PCE prediction. Their wider CIs suggest that they are slightly less confident in their predictions, which may be due to the models’ simpler architectures compared to LSTM. They lack to capture time dependencies, which could explain the minor drop in certainty. However, they still manage to perform reasonably well in terms of coverage.

The RF model presents the widest mean CI width, at 1.8229, indicating the highest level of uncertainty in its predictions. Its coverage probability of 94.41% is still respectable, but the much wider CIs suggest that RF is less certain about its predictions compared to the other models. The broader intervals are likely a result of the model’s ensemble nature, which combines multiple decision trees to make predictions. This ensemble method introduces more variability in the predictions, leading to higher uncertainty. However, RF’s ability to maintain reasonable coverage suggests that while it is cautious in its predictions (as shown by the wide CIs), it still captures a significant portion of the actual PCE values.

Figure 21 shows the predicted vs. actual PCE values with the 95% bootstrapped CIs provide a visual representation of each model’s uncertainty and performance. These graphs further support the insights gained from the table, offering a clear view of how well each model captures uncertainty.

Fig. 21
figure 21

Predicted vs. actual PCE with bootstrapped confidence intervals for (a) LSTM, (b) MLP, (c) SVM and (d) RF.

The LSTM graph visually confirms the model’s high confidence in its predictions, with narrow CIs closely surrounding the predicted values. Most of the actual PCE values fall well within the CIs, supporting the high coverage probability of 96.27%. The narrow CIs reflect LSTM’s ability to make precise predictions, likely due to its capacity to learn complex temporal patterns in the data. This makes LSTM a highly reliable model for PCE prediction, especially when prediction accuracy is critical.

The MLP and SVM graphs show slightly wider confidence intervals compared to LSTM, which is consistent with the table results. While the CIs are somewhat broader, the graphs still indicates that most actual PCE values are captured within the prediction intervals, reflecting the models’ coverage of 95.65% and 95.03 for MLP and SVM, respectively. The slightly lower confidences compared to LSTM are evident in the slightly wider bands, which may be attributed to MLP’s and SVM simpler architecture, making them less adept at capturing subtle patterns in the data. Nonetheless, these models still perform reasonably well, providing reliable PCE predictions.

The RF graph displays the widest CIs among the four models, reflecting the high level of uncertainty in RF’s predictions. The wide intervals suggest that RF is more conservative in its predictions, providing a broader range to account for potential variability in the data. This is consistent with the table’s result of the largest mean CI width (1.8229) for RF. Despite the broader intervals, the RF model still captures a reasonable percentage of actual PCE values (94.41% coverage), indicating that while the model is cautious, it is still able to provide relatively reliable predictions.

The results of our study indicate a clear advantage of using LSTM and MLP neural networks over SVM and RF for predicting the PCE of iOSCs. The LSTM model outperformed all the other reported models in this study for all the used performance metrics. This significant performance difference can be attributed to its architecture, which is specifically designed to retain information over long sequences. This capability is particularly beneficial for time-series data, where the efficiency of a solar cell can change over time due to various factors such as material degradation, environmental conditions, and operational stresses. By capturing these long-term dependencies, the LSTM model can provide more accurate predictions, which is evident from the lower error metrics compared to other studied prediction models.

The MLP model ranked second with a slight difference after the LSTM and outperformed SVM and RF models in this study for all the used performance metrics. This performance underscores the ability of MLP networks to handle good prediction capabilities for complex systems and patterns due to its ability to learn and store knowledge (in the form of weights) for future predictions. MLPs are highly effective for prediction tasks due to their ability to model complex non-linear relationships in data. In addition, MLP layers and non-linear activation functions enable it to capture intricate dependencies between input features and outputs.

Furthermore, the search for optimal parameters of the double exponential model using the Levenberg–Marquardt optimization algorithm is highly efficient and outperforms the traditional double exponential model that relies on the Nonlinear Least Squares optimization algorithm. The Levenberg–Marquardt algorithm combines the benefits of both gradient descent and the Gauss–Newton method, providing faster convergence and better handling of complex, non-linear relationships. This results in more accurate parameter estimation and improved model performance in fitting data compared to the Nonlinear Least Squares approach.

The application of advanced AI techniques such as LSTM and MLP in predicting the PCE of organic solar cells represents a significant advancement in the field. Traditional methods of optimizing PCE are labour-intensive and time-consuming, often relying on trial-and-error approaches. By leveraging AI, researchers can predict the efficiency of various material and device configurations before physical testing, thus saving considerable time and resources.

The findings of this study highlight the potential of AI to accelerate the development and optimization of organic solar cells. The high accuracy of the LSTM and MLP models, in particular, suggests that AI can provide valuable insights into the factors that influence PCE, guiding researchers towards the most promising materials and configurations. This predictive capability is crucial for advancing the efficiency and commercial viability of organic solar cells, ultimately contributing to the broader goal of sustainable energy generation.

Future research directions

Despite the promising results, there are several areas for future research that could further improve the prediction accuracy and utility of AI models in this context:

  • Integration of environmental data: Incorporating environmental data such as temperature, humidity, and exposure to pollutants could provide a more comprehensive understanding of the factors affecting PCE and enhance the model’s predictive accuracy.

  • Long-term stability data: Including data on the long-term stability and degradation of solar cells can help in predicting the lifespan and performance of solar cells over extended periods, providing more valuable insights for researchers and manufacturers.

  • Advanced AI techniques: Exploring other advanced AI techniques, such as convolutional neural networks (CNNs) or hybrid models that combine multiple AI approaches, could potentially yield even more accurate predictions.

  • Material and structural innovations: Research into new materials and innovative structural designs for organic solar cells can be guided by AI predictions, accelerating the discovery of high-efficiency configurations.

  • Real-world testing and validation: Implementing real-world testing and validation of AI predictions can help refine the models and ensure their practical applicability in diverse environmental conditions and usage scenarios.

Conclusion

This research demonstrates that both LSTM and MLP neural networks are viable tools for predicting the PCE of iOSCs. The application of these AI techniques can significantly accelerate the optimization of organic solar cells, reducing the reliance on extensive experimental trials. By leveraging the predictive capabilities of AI, researchers can focus their efforts on the most promising materials and device configurations, ultimately advancing the development of high-efficiency organic solar cells.

The findings underscore the potential of artificial intelligence in transforming the research and development landscape of organic photovoltaics. The high accuracy achieved by the LSTM and MLP models, with R2 values of 0.9675 and 0.9562, respectively, suggests that temporal dependencies and nonlinear relationships in the data can be effectively captured, leading to precise predictions. This is particularly beneficial for long-term performance assessment and optimization of solar cells, where traditional methods fall short due to their labour-intensive and time-consuming nature.

Data Availability Statement

The datasets used and/or analysed during the current study available from the corresponding author on reasonable request.