Introduction

Medical imaging plays a critical role in modern healthcare, providing vital insights into the internal structures of the human body for the diagnosis and treatment of diseases such as neurological, cardiovascular, and oncological disorders. The advancement of digital imaging technologies, including magnetic resonance imaging (MRI), computed tomography (CT), and positron emission tomography (PET) has enabled high-resolution visualization of anatomical and physiological features1,2. However, a persistent challenge in medical imaging is the presence of noise i.e. random variations in pixel intensity that degrade image quality, particularly under conditions of low contrast or insufficient radiation exposure3. Such noise not only obscures fine structural details but also hampers accurate diagnosis and treatment planning, necessitating effective noise removal techniques as a critical preprocessing step. Traditional image denoising methods, including wavelet thresholding4, Wiener filtering, and median filtering have been widely used to suppress noise in medical images5. While these approaches can reduce noise to some extent, they often struggle to balance noise removal with the preservation of important features such as edges and textures which is essential requirement for clinical interpretation. More recent developments in sparse representation and dictionary learning6, as well as deep learning-based models3 have demonstrated improved denoising performance. However, these methods often require high computational complexity, extensive training data, or lack adaptability across different imaging modalities and noise types7,8,9.

In addition to its application in medical imaging, the proposed denoising method offers promising potential for enhancing user experience in consumer-grade photography and videography, where image quality is often compromised by noise under low-light or high-ISO conditions. By leveraging adaptive clustering and non-local means, the method effectively reduces noise while preserving fine textures and edges, making it well-suited for scenarios such as night photography, smartphone imaging, or videography in challenging lighting environments. To contextualize this work, it is important to consider the chronological development of image denoising techniques. Early approaches included spatial domain filtering methods such as mean, median, and Gaussian filters1, followed by transform-domain techniques like wavelet thresholding2 and the original non-local means algorithm by Buades et al.3. Subsequent advances introduced sparse coding4, dictionary learning5, and low-rank matrix factorization6 as powerful denoising strategies. More recently, deep learning-based methods such as DnCNN7, U-Net architectures8, and generative adversarial networks (GANs) for denoising9 have achieved state-of-the-art results. The proposed method builds upon these foundations by integrating adaptive clustering with principal component analysis and non-local similarity exploitation, offering a hybrid approach that balances statistical robustness and structural preservation. This framework situates the method as a competitive solution not only for medical images but also for consumer applications where perceptual quality and detail preservation are critical.

In this study, we propose an efficient and robust image denoising framework that combines adaptive clustering, principal component analysis (PCA) thresholding guided by the Marchenko–Pastur (MP) law10 from random matrix theory (RMT)11,12, and a non-local means (NLM) algorithm to achieve superior noise suppression while preserving fine image details. The key innovation of the proposed approach lies in its integration of statistical properties of eigenvalues under the MP law to estimate global noise levels and guide hard and soft thresholding operations in the PCA transform domain. Furthermore, an adaptive clustering strategy automatically determines optimal clusters of similar image patches, enhancing local feature representation and enabling targeted noise reduction within each cluster. Finally, a non-local means algorithm refines the denoised image by leveraging similarity across non-adjacent pixel neighbourhoods, improving detail preservation without sacrificing denoising effectiveness.

The main contribution of the paper is the development of a robust and efficient image denoising method that integrates adaptive clustering, principal component analysis (PCA)-based thresholding using the Marchenko-Pastur (MP) law, and a non-local means (NLM) algorithm to effectively remove noise while preserving fine image details, particularly in medical images. This study introduces a novel adaptive clustering approach that automatically determines the optimal number of clusters by combining over-clustering and iterative merging strategies to group similar image patches based on texture and edge features. Within each cluster, the method applies a two-stage denoising process: hard thresholding in the singular value decomposition (SVD) domain guided by the MP law to achieve a low-rank approximation, followed by a locally parameterized linear minimum mean square error (LMMSE) soft thresholding to further suppress residual noise. The final integration of an NLM algorithm enhances noise reduction by leveraging non-local similarity across the image. Experimental results demonstrate that the proposed method achieves superior denoising performance and better preserves textures and edges compared to existing techniques, making it highly suitable for precision-critical applications in medical imaging.

The remainder of the paper is as follows: Section "Related work" presents a detailed analysis of recent state-of-the-art image denoising techniques. Section "Proposed methodology" provides a step-by-step explanation of the proposed method for image denoising. The experimental setup and discussion section outlines an experimental setup, and discusses the effectiveness of the proposed approach and compares it with other recent techniques in Section "Experimental setup and results". The Ablation study results showing the effect of outlier exclusion and soft thresholding on denoising metrics is presented in Section "Ablation study". Lastly, Section "Conclusion" presented “Conclusion” summarizes the key findings and contributions of the proposed approach.

Related work

Image processing is a fundamental area in computer vision and medical imaging, encompassing techniques for enhancing, analysing, and interpreting visual information13,14,15,16,17,18. A critical task within image processing is image denoising that aimed at removing noise while preserving essential features like edges and textures19,20,21. Over the years, a variety of denoising methods have been developed, ranging from classical spatial and transform-domain filters (e.g., Gaussian, median, wavelet thresholding) to more sophisticated approaches like sparse coding, dictionary learning, and deep learning models22,23,24. Recently, hybrid techniques combining statistical models, clustering, and non-local similarity measures have emerged to achieve a better balance between noise reduction and detail preservation. Recently, Jifara et al.3 proposes a deep feedforward denoising by leveraging a deep learning framework. Their approach incorporated batch normalisation and residual learning into the deep model. Residual learning was used to effectively extract noise patterns from noisy images, while batch normalisation was integrated with residual learning for enhancing both training accuracy and efficiency. However, this technique did not assess the reduction of compression artifacts in images using a prominent CNN model. Ji et al.25 presents a geometric regularisation system for image denoising. This technique focused on rebuilding surfaces by minimizing approximation and gradient errors to alleviate noise in medical imaging. The surface was created by dividing the coefficients into two groups. The initial group was used to eliminate surface gradient while the subsequent group aimed to enhance estimation accuracy. However, the approach did not enhance the real-time efficiency. Aravindan et al.26 developed a method for effectively denoising a high noise density images with minimum computational expense. This method proposes denoising technique using SSO and DWT. First, salt & pepper, Gaussian and speckle noises are introduced followed by the adaptation of the DWT. Optimization of wavelet coefficients was performed for enhancing coefficient values utilizing SSO. Subsequently, the inverse discrete wavelet transform is employed in the optimization of wavelet parameters. The approach substantially reduced image noise; nevertheless, it was unsuitable to other datasets. Rani et al.27 developed a steepest learning model used in ANN to adaptively minimize image noise. In addition, a soft thresholding function was introduced as an activation function for artificial neural networks. But the strategy neglected to account for higher-order derivatives in addressing mathematical aspects. Laves et al.28 developed a randomly initiated conv network for image denoising. This methodology used a Bayesian scheme with Monte Carlo dropout to evaluate to estimate both epistemic and aleatoric uncertainty. This strategy enhanced predictive uncertainty and connects with predictive error. But the system did not use deformable registration to enhance results. Kumar and Nagaraju29 presents a filter for the restoration of noisy images. This filter integrates a type II fuzzy system with the cuckoo search scheme for image restoration. Noisy pixels in images are found using a circular related search technique, and the detected damaged pixels are removed. This approach is adaptable in numerous noisy circumstances but did not enable adequate exploration. Bai et al.6 presents a denoising technique which combined sparse dictionary learning with the cluster ensemble. This technique used both inherent self-similarity and disparity in image patches. The image feature set are initially created via steering kernel regression. The cluster ensemble is subsequently employed to obtain class labels. The trained dictionary exhibited greater adaptability and facilitated enhancements in recovered image quality; however, the system was hindered by significant mathematical complexity. Miri et al.7 developed a method for removing Gaussian noise using an ant colony optimisation scheme and 2DDCT. This method determines essential frequency coefficients and alleviated noise effects by removing high-frequency components. But the system did not successfully include deep learning techniques to enhance performance. Kollem et al.30 present an image denoising based on a diffusivity function-derived partial differential equation by utilizing the statistical features of observed noisy images. This method employs a QWT which generates the various coefficients of a noisy image. Huang et al.31 present a self-supervised sparse coding technique known as WISTA. This method does not depend on ground truth image pairs, and learns solely from a single noisy image. However, to enhance denoising performance, this method expands the WISTA framework for developing a DNN architecture referred to as WISTA-Net. Atal22 introduces an ODVV filter to denoise Lena and CT images. The input images are processed using a noisy pixel map recognition module, where a DCNN is utilised to identify noisy pixel maps. DCNN training is conducted using an Adam algorithm. Lepcha et al.32 present a low-dose computed tomography (LDCT) denoising technique using sparse 3D transformation and PNLM. The hard thresholding module in a sparse 3D transformation is utilised to alleviate noise. Also, PNLM addresses the shortages of non-local means weighting by providing weights that are more accurately represent patch similarities.

Jang et al.33 presented the Spach Transformer, an efficient encoder decoder transformer which utilizes spatial and channel information through global and local MSAs. Shen et al.34 presents the diffusion probabilistic scheme due to its robust generative capabilities. During the training phase, the model acquires a scoring function network through the optimization of the evidence lower bound (ELBO). Lepcha et al.35 devised an effective LDCT denoising scheme using a constructive non-local means approach using morphological residual processing to eliminate noise. Ma et al.36 employs a meticulously created block as the foundation of the encoder decoder model which merges Swim Transformer modules with residual blocks in a parallel configuration. Swim Transformer modules can adeptly acquire hierarchical representations of noise artifacts using self-attention method through non-overlapping shift windows and cross window connections while residual blocks are beneficial for mitigating the loss of precise information through shortcut connections. Annavarapu et al.37 proposes an adapted DCNN for image denoising and coupled with adaptive watershed segmentation to restore image information, and a hybrid lifting method based on revised bi-histogram equalization. This method is subsequently enhanced with the incorporation of marker-based watershed segmentation. Zeng et al.38 present a TNCDN model for denoising noisy images. The information from two denoising networks which possess ongoing learning capabilities is transmits to the primary denoising model. Sharif et al.39 confronts the constraints of current denoising techniques with an AI-driven two-stage learning approach. The proposed method acquires the capability to estimate the residual noise from the contaminated images. Subsequently, it integrates an innovative noise attention mechanism to associate estimated residual noise with noisy inputs, enabling denoising in a coarse-to-fine approach. Lee et al.40 presents a MSAN, a model designed for an effective denoising of medical images. It consists three primary modules such as (a) a feature extraction layer capable in identifying low and high frequency characteristics, (b) a multiscale self-attention scheme which predicts residual utilising both original and down-sampled feature maps and a decoder which generates a residual image. Huang et al.41 introduce Be-FOI, a weakly supervised model which employs cine images. BeFOI uses Markov chains for denoising and depends only on the fluoroscopic image as a reference, which have been trained using accurate noise measurement and simulation. It outperforms other methods by eliminating noise, improving clarity and diagnostic interpretation.

Proposed methodology

The proposed approach for medical image denoising begins with estimating the noise variance of the input noisy image using the Marchenko–Pastur (MP) law, which provides a robust statistical basis for noise level estimation. Once the noise level is determined, the image is divided into overlapping patches to capture localized structural information. These patches are then grouped into similar clusters through an adaptive clustering process, which initially over-clusters the patches using k-means and subsequently merges similar clusters based on an inter-cluster distance threshold. For each resulting cluster, denoising is performed in two stages: first, hard thresholding is applied in the principal component analysis (PCA)-singular value decomposition (SVD) domain by discarding noise-dominated singular values identified using the MP law; second, a soft thresholding step is performed using a local linear minimum mean square error (LMMSE) estimator to further suppress residual noise while preserving fine details. The denoised patches from all clusters are then aggregated to reconstruct an intermediate denoised image. Finally, a non-local means (NLM) refinement is applied, which computes weighted averages based on pixel neighbourhood similarity, enhancing detail preservation and ensuring effective noise removal. This multi-stage process integrates global statistical estimation, localized structural adaptation, and non-local similarity to achieve robust and detail-preserving denoising of medical images. The workflow of the proposed approach is demonstrated In Fig. 1.

Fig. 1
figure 1

Detailed architecture of the proposed denoising framework.

Noise level estimation

Noise level is the important factor in an image denoising process42. An optimal performance of PCA-based noise level estimate is demonstrated in references43,44,45. In43,44, low-rank patch selection incurs significant computing costs and exhibits instability at elevated noise levels. 45 presents a rapid method for estimating noise levels based on the observation that patches extracted from a noiseless image frequently reside within a low dimensional subspace. The low-dimensional subspace can be obtained through the low-rank approximation of PCA while the noise level can be derived from eigenvalues of the covariance matrix of noisy patches. Motivated by this approach, we present a robust and efficient method for estimating noise levels via combining this estimation scheme with a MP law. The relationship between the noise level \(\sigma\) along with PCA eigenvalue λ of the random matrix N42 allows for the effective use of MP law for estimating the gaussian noise level of the image. The MP law is utilised for locally computing the noise level within a compact 3 D block11,12. Further we utilize the MP law to predict noise levels on a global scale. The global approach is justified as the MP law explains the characteristics of large matrices with greater precision than those of small matrices. Designate the subscript \(\phi\) to signify the whole image. To obtain precise estimations, we partition the entire \(a \times b\) image into overlapping \({d}_{\varnothing }\times {d}_{\varnothing }\) patches and aggregate all the overlapping patches to form a substantial matrix \({X}_{\varnothing } \epsilon {\mathbb{R}}^{{M}_{\varnothing }\times {L}_{\varnothing }}\), where \({M}_{\varnothing }={d}_{\varnothing }^{2}\) and \(L_{\emptyset } = \left( {a - d_{\emptyset } + 1} \right)\left( {b - d_{\emptyset } + 1} \right)\). The large nosy matrix \({X}_{\varnothing }\) is technically characterized as a low-rank matrix \({X}_{0,\varnothing }\) combined with the noisy matrix \({X}_{\varnothing }={X}_{0,\varnothing }+{N}_{\varnothing }\), where each column of \({N}_{\varnothing }\) is a vector which adheres to a multivariate gaussian distribution \({N}_{{M}_{\varnothing }}\left(0, {\sigma }^{2}\text{I}\right)\) with a mean of 0 and a covariance matrix of \({\sigma }^{2}\text{I}\). Estimation of noise level \(\sigma\) is directly correlated with the PCA eigenvalues of \({N}_{\varnothing }\). However, \({N}_{\varnothing }\) is indeterminate whereas \({X}_{\varnothing }\) is known. Low-rank of matrix \({X}_{0,\varnothing }\) allows to examine the correlation the PCA eigenvalues of \({N}_{\varnothing }\) and those of \({X}_{\varnothing }\). Some findings on this correlation are discussed in42. Utilizing the PCA eigenvalue \({\lambda }_{{x}_{\varnothing },i} 1 \le i \le {M}_{\varnothing }\) of \({X}_{\varnothing }\), we compute the noise level \(\sigma\). Statistical mean of the observed eigenvalues between \({\lambda }_{{n}_{\varnothing }-}\) and \({\lambda }_{{n}_{\varnothing }+}\) is almost equivalent to the predicted value of \({\lambda }_{{n}_{\varnothing }}\).

$$\sigma^{2} = E\left( {\lambda_{{n_{\emptyset } }} } \right) \approx \frac{{\mathop \sum \nolimits_{i = 1}^{{M_{\emptyset } }} \lambda_{{n_{\emptyset } ,i}} 1\left( {\lambda_{{n_{\emptyset } - }} \le \lambda_{{n_{\emptyset } ,i}} \le \lambda_{{n_{\emptyset } + }} } \right)}}{{\mathop \sum \nolimits_{i = 1}^{{M_{\emptyset } }} 1\left( {\lambda_{{n_{\emptyset } - }} \le \lambda_{{x_{\emptyset } ,i}} \le \lambda_{{n_{\emptyset } + }} } \right)}}$$
(1)

where \({\lambda }_{{n}_{\varnothing }\pm }= {\sigma }^{2}{\left(1\pm \sqrt{\gamma \varnothing }\right)}^{2}\) with \(\gamma \varnothing =\frac{{M}_{\varnothing }}{{L}_{\varnothing }},\)

$$\sigma^{2} = \frac{{\lambda_{{n_{\emptyset } + }} - \lambda_{{n_{\emptyset } - }} }}{{4 \sqrt {\gamma \emptyset } }}$$
(2)

Above, Eqs. (1) and (2) demonstrate that estimating the noise level involves identifying the range \(\left[{\lambda }_{{n}_{\varnothing }-},{\lambda }_{{n}_{\varnothing }+}\right]\). We can directly set \({\lambda }_{{n}_{\varnothing }-}= {\lambda }_{{x}_{\varnothing }}, {M}_{\varnothing }\) since \({\lambda }_{{n}_{\varnothing }-}\le {\lambda }_{{x}_{\varnothing }}, {M}_{\varnothing } \le {\lambda }_{{x}_{\varnothing }, i}, 1 \le i \le {M}_{\varnothing }.\) Subsequently, we determine a suitable \({\lambda }_{{x}_{\varnothing }+}\) by heuristic greedy search based on Eqs. (1) and (2). In particular for each \({\lambda }_{{x}_{\varnothing }, i},\) if \({\lambda }_{{n}_{\varnothing }+ }= {\lambda }_{{x}_{\varnothing }, i}\), we can compute the variance estimation \({\sigma }_{{i}_{1}}^{2}\) using Eq. (1) and \({\sigma }_{{i}_{2}}^{2}\) using Eq. (2), respectively. Subsequently, we compute the difference \({\Delta }_{i}= {\Vert {\sigma }_{{i}_{1}}^{2}- {\sigma }_{{i}_{2}}^{2}\Vert }_{2}^{2}\). The optimal \({\lambda }_{{n}_{\varnothing }+}\) is ascertained by selecting the suitable \({\lambda }_{{x}_{\varnothing },i}\) that minimizes the difference \({\Delta }_{i}\). The noise level \(\widehat{\sigma }\) is determined as per Eq. (1) after calculating \({\lambda }_{{n}_{\varnothing }\pm }\).

Adaptive clustering

After estimating the global noise level \(\widehat{\sigma }\), the proposed method performs an adaptive clustering of image patches to group similar features together for effective denoising42. The goal of clustering is to ensure that patches with similar local structures, such as edges or textures, are grouped into the same cluster, while patches representing different features are separated. Traditional clustering methods like K-means require specifying the number of clusters in advance, which is impractical because the number of distinct features in an image is generally unknown. Although various cluster validation indices have been proposed to determine the optimal cluster number, they tend to be computationally expensive and complex. To overcome this challenge, an adaptive clustering strategy based on an over-clustering-and-iterative-merging approach46 is developed. In the first stage, over-clustering is deliberately performed to produce a large number of clusters so that different features are initially separated. This is achieved using a K-means-based divide-and-conquer method47, where the initial number of clusters is set as \(\text{max}\left\{\frac{ab}{{256}^{2}}, 4\right\}\), with \(a\) and \(b\) being the image height and width. Each cluster obtained from this stage is further subdivided into \(\lfloor\frac{{L_{{j0}} }}{M} \rfloor\) clusters in a second stage, where \({L}_{j0}\) is the number of patches in cluster \({j}_{0}\), M is the patch dimensionality, and denotes the floor function. After over-clustering, an iterative merging step is applied to combine similar clusters and prevent fragmentation. Two clusters are merged if the squared Euclidean distance between their centroids is below a threshold \(T\), which defines an acceptable similarity level between clusters. To derive this threshold, the distance between cluster centroids is modelled under the assumption that one cluster centre is noise-free and the other is affected by additive Gaussian noise. In this model, the squared distance follows a chi-squared distribution with M degrees of freedom, expressed as

$$D_{B,A}^{2} = y_{B} - y_{A2}^{2} = \mathop \sum \limits_{i = 1}^{M} n_{i.}^{2}$$
(3)

where n represents Gaussian noise. The threshold T is determined to satisfy \(P \left({D}_{B,A} < \sqrt{T}\right)=\epsilon\), where ϵ is a small probability (e.g., 1.3 × \({10}^{-10}\), leading to T being calculated from the chi-squared cumulative distribution function as \(T= {\chi }_{1-\epsilon }^{2}, {M}^{{\sigma }^{2}}\). For practical purposes, when M = 64 and \(\epsilon =1.3 \times {10}^{-10}\), this yields T ≈ 16.0 \({\sigma }^{2}\). This adaptive clustering method does not require explicit specification of the number of clusters, instead automatically adjusting based on the noise level and statistical properties of the image. It ensures that similar image features are grouped effectively while preserving the separation of distinct features, thereby laying a robust foundation for subsequent low-rank denoising of each cluster.

Hard thresholding via MP-SVD with rank estimation

In this step, we denoise each cluster matrix by performing hard thresholding of the singular values based on the Marchenko–Pastur (MP) law, which provides a principled way to separate signal from noise in the singular value spectrum42.

Let \({X}_{j} \in {\mathbb{R}}^{M\times {L}_{j}}\) denote the noisy data matrix for the \(j\)-th cluster, composed of similar patches grouped together. The noisy matrix can be expressed as:

$$X_{j} = X_{0,j} + N_{j}$$
(4)

where \({X}_{0,j}\)​ is the underlying low-rank clean matrix, and \({N}_{j}\)​ represents additive Gaussian noise.

We perform singular value decomposition (SVD) on \({X}_{j}\)​:

$$X_{j} = U_{j} \Sigma_{j} V_{j}^{T }$$
(5)

where \({U}_{j}\) and \({V}_{j}\) are orthonormal matrices, and \({\Sigma }_{j}=diag \left({\sigma }_{j,1},{\sigma }_{j,2},\dots \dots .,{\sigma }_{j,M}\right)\) is a diagonal matrix containing singular values in descending order.

According to the MP law, the noise singular values are asymptotically confined below an upper bound \({\lambda }_{n}^{+},\) given by:

$$\lambda_{n}^{ + } = \sigma^{2} \left( {1 + \sqrt \gamma } \right)^{2}$$
(6)

with \(\gamma = \frac{M}{{L_{j} }} {\text{and}} \sigma^{2}\) the noise variance.

To remove the noise-dominated components, we define a hard threshold:

$$\xi = \sqrt {\mu \lambda_{n}^{ + } }$$
(7)

where μ > 1 is an adjustment parameter to slightly push the threshold above \(\lambda_{n}^{ + }\) for better separation.

We retain only the singular values exceeding \(\xi\), yielding a low-rank approximation \(\tilde{X}_{j}\) as:

$$\tilde{X}_{j} = \mathop \sum \limits_{{i:\sigma_{j,i} > \xi }} \sigma_{j,i} u_{j,i} V_{j,i}^{T}$$
(8)

or equivalently in matrix form:

$$\tilde{X}_{j} = U_{j,r} \Sigma_{j,r} V_{j,r}^{T}$$
(9)

where \({U}_{j,r }{\Sigma }_{j,r }{V}_{j,r}\) correspond to the singular vectors and singular values associated with the \(r\) retained components (i.e., \({\sigma }_{j,i}>\upxi .\)

This hard thresholding effectively nullifies the singular values below ξ, removing the noisiest subspace while preserving the dominant signal structure. The matrix \({\widetilde{X}}_{j}\) serves as the initial denoised output for cluster \(j\), to be further refined (if necessary) in subsequent processing. Hard thresholding is applied during the initial phase of denoising within each adaptively formed cluster. Given a noisy cluster matrix X, we perform singular value decomposition (SVD) and discard singular values below a threshold determined using the Marchenko–Pastur (MP) law from random matrix theory. The threshold is defined as:

$${\upxi } = \sqrt {\mu \lambda_{n + } ,} \lambda_{n + } = \sigma^{2} \left( {1 + \sqrt \gamma } \right)^{2} ,$$
(10)

where \(\mu > 1\) is an empirical constant (we use \(\mu = 1.1),\) \(\sigma^{2}\) is the estimated noise variance, and \(\gamma = \frac{M}{L}\) is the aspect ratio of the patch matrix \(X \in {\mathbb{R}}^{M \times L}\). Only singular values \(\sqrt {\lambda_{x,i} } > {\upxi }\) are retained in the reconstructed low-rank matrix:

$$\overline{X} = \sqrt L \mathop \sum \limits_{i = 1}^{M} \sqrt {\lambda_{x,i} } . 1\left( {\sqrt {\lambda_{x,i} } > {\upxi }} \right) . u_{x,i} v_{x,i}^{T} ,$$
(11)

where the above equation describes the reconstruction of a denoised low-rank matrix \(\widetilde{\text{X}}\) by performing hard thresholding on the singular values obtained from the singular value decomposition (SVD) of a noisy matrix \(X\). In this equation, \(\widetilde{\text{X}}\) represents the denoised approximation, \(L\) is the number of columns (i.e., the number of patch vectors in the cluster), and M is the number of rows (i.e., the dimensionality of each patch vector). The term \({\lambda }_{i,j}\)​ denotes the \(i\)-\(th\) eigenvalue (or squared singular value) of the covariance matrix \(X{X}^{T}\), reflecting the energy captured by the corresponding principal component. The indicator function 1 \(\left(\sqrt{{\lambda }_{x,i}}>\upxi \right)\) takes a value of 1 if the square root of the eigenvalue (i.e., the singular value) exceeds the threshold ξ, and 0 otherwise; this step ensures that only significant (signal-dominated) components are retained, while noise-dominated components are discarded. The threshold ξ is calculated using the Marchenko–Pastur law, adapting to the noise level in the data. The vectors \({u}_{x,i}\) and \({v}_{x,i}\) are the \(i\)-\(th\) left and right singular vectors of matrix X, respectively, which describe the principal component directions in the row and column spaces. The summation runs over the M singular vectors and their corresponding singular values, multiplying each retained component to reconstruct the low-rank approximation. The scaling factor \(\sqrt{L}\) normalizes the reconstruction based on the number of columns. Together, these terms define a process that selectively reconstructs the data by retaining only those components that exceed the noise threshold, effectively suppressing noise while preserving meaningful signal structures.

Soft thresholding via local LMMSE estimation in PCA domain

After the removal of the dominant noise components through hard thresholding in the SVD domain, we obtain the signal-dominated low-rank matrix \({\widetilde{X}}_{j}\) for each cluster \(j\). While the hard thresholding discards the noisiest singular values, residual noise remains in the retained components. To further suppress this noise while preserving fine details, we apply a soft-thresholding process in the PCA transform domain, leveraging the linear minimum mean-square error (LMMSE) estimator with locally estimated parameters42.

Let \({\widetilde{X}}_{j}\) be decomposed as:

$$\tilde{X}_{j} = {\text{ U}}_{{\tilde{X}_{j} }} {\Sigma }_{{\tilde{X}_{j} }} V_{{\tilde{X}_{j} }}$$
(12)

where \({\text{U}}_{{\tilde{X}_{j} }} {\text{and }}V_{{\tilde{X}_{j} }}\) are the left and right singular vectors, and \({\Sigma }_{{\tilde{X}_{j} }}\)​​ is a diagonal matrix containing the retained singular values. Projecting \(\tilde{X}_{j}\) into the PCA domain yields the transform coefficient matrix:

$$S_{j} = U_{{\tilde{X}_{j} }}^{T} X_{j}$$
(13)

Each element \(s_{{j,\left( {i,k} \right)}}\) in \(S_{j}\) represents the \(k\)-th coefficient of the \(i\)-th principal component in cluster \(j\). We perform soft-thresholding on each coefficient \(s_{{j,\left( {i,k} \right)}}\) using an adaptive weight \(w_{{j,\left( {i,k} \right)}}\), computed from the local variance estimate \(\tilde{S}_{{j,\left( {i,k} \right)}}\) as:

$$\hat{s}_{{j,\left( {i,k} \right)}} = w_{{j,\left( {i,k} \right) }} s_{{j,\left( {i,k} \right)}}$$
(14)

With

$$w_{{j,\left( {i,k} \right) }} = \max \left\{ {\frac{{\tilde{S}_{{j,\left( {i,k} \right)}} - \sigma^{2} }}{{\tilde{S}_{{j,\left( {i,k} \right)}} }}, 0} \right\}$$
(15)

where \({\sigma }^{2}\) is the estimated noise variance, and \({\widetilde{S}}_{j,\left(i,k\right)}\) is the local variance estimated around position \(k\) in the \(i\)-th band:

$$\tilde{S}_{{j,\left( {i,k} \right)}} = \frac{{\mathop \sum \nolimits_{{k_{s} = 1}}^{{L_{j} }} s_{{j,\left( {i,k_{s} } \right)}}^{2} . 1 \left( {\left| {k_{s} - k} \right| \le \zeta } \right)}}{{\mathop \sum \nolimits_{{k_{s} = 1}}^{{L_{j} }} 1 \left( {\left| {k_{s} - k} \right| \le \zeta } \right)}}$$
(16)

where ζ controls the size of the local window and 1() is the indicator function.

Finally, the denoised patch matrix \(\hat{X}_{j}\)​ for cluster \(j\) is reconstructed by inverse PCA transformation as

$$\hat{X}_{j} = {\text{U}}_{{\tilde{X}_{j} }} \tilde{S}_{j}$$
(17)

where \({\widetilde{S}}_{j}\) is the matrix of soft-thresholded coefficients \({\widehat{s}}_{j,\left(i,k\right)}\). This locally adaptive soft-thresholding enables effective suppression of residual noise while preserving signal variations intrinsic to the principal components, leading to superior detail preservation compared to global thresholding strategies.

Non-local means algorithm

The Non-Local Means (NL-means)48 algorithm is an image denoising method that restores each pixel by computing a weighted average of all pixels in the image, rather than limiting the averaging to a local neighbourhood. Given an image \(v \approx {\widehat{X}}_{j}\) (i.e. Eq. 17), the denoised value at a pixel \(i\) and denoted as \(NL\left[v\right]\left(i\right)\) is calculated by summing the intensity values of all pixels \(j\) in the image, each multiplied by a weight \(w\left(i,j\right)\) that reflects the similarity between the neighbourhoods around pixels \(i\) and \(j\). Mathematically, this is expressed as

$$NL\left[ v \right]\left( i \right) = \mathop \sum \limits_{j \in 1} w\left( {i,j} \right) v \left( j \right)$$
(18)

The weight \(w\left( {i,j} \right)\) is defined as

$$w\left( {i,j} \right) = \frac{1}{Z\left( i \right)}\exp \left( { - \frac{{v \left( {N_{i} } \right) - v \left( {N_{j} } \right) _{2,a}^{2} }}{{h^{2} }}} \right)$$
(19)

where \(Z\left(i\right)\) is a normalization factor ensuring that the sum of the weights equals one, and \({\Vert v \left({N}_{i}\right)-v \left({N}_{j}\right) \Vert }_{2,a}^{2}\)​ represents the weighted squared Euclidean distance between the neighbourhoods \({N}_{i}\)​ and \({N}_{j}\) of pixels \(i\) and \(j\) computed using a Gaussian kernel with standard deviation \(a\). The parameter \(h\) controls the decay of the exponential function, effectively determining how quickly the weight decreases as the neighborhood similarity diminishes. This approach allows pixels that have similar surrounding patterns to contribute more significantly to the denoised value, enabling the algorithm to preserve textures and fine details by leveraging repetitive structures throughout the image. Unlike traditional local filters, the NL-means algorithm utilizes information from the entire image, making it particularly effective in maintaining important features while reducing noise.

Algorithm
figure a

Adaptive clustering and non-local means denoising

Experimental setup and results

In the experiments, the MRI and CT image datasets were obtained from https://www.med.harvard.edu/aanlib/home.html49, a publicly available repository of medical images for research purposes. The datasets were pre-processed by dividing them into training and testing sets to ensure robust evaluation of the denoising performance. Specifically, 70% of the images were randomly selected for training, allowing the model to learn from a diverse set of anatomical structures and noise patterns, while the remaining 30% were reserved for testing to evaluate the generalization capability of the proposed method. Both sets included images with varying noise levels (σ ≈ 10, 20, 30, 40) to simulate realistic noise conditions. Each image was divided into overlapping patches of size 8 × 8 for noise reduction and 10 × 10 for noise level estimation as described in the experimental methodology. The split ensured no overlap of image instances between training and testing to avoid data leakage. Performance was evaluated using standard metrics (PSNR, SSIM, FSIM, Entropy, BRISQUE, PIQE, NIQE) to quantify denoising quality across both sets. The quality of the image is better when PSNR50,51, SSIM, FSIM50,51 and Entropy have higher values, whereas BRISQU52, PIQE, and NIQE53 should have lower values. The quantitative study confirms that the proposed noise level estimation achieves superior denoising performance on average compared to other approaches. For calculating the efficiency of the image denoising method, we compare the proposed approach with several other popular algorithms, namely OD-CNN54, NDiff30, SDPM34, UDLF55, NGN56, CNCL57,BP-MEC25, FDCT58, NSCT-BT59, IDM-PM60, ONLM61, DCT-ACO7, ANLM62, Z-NLM63. Three test images with \(\sigma\) values of 10, 20, 30, and 40 are utilized for representation: two MRI images (MRI1 and MRI2) and one CT image from https://www.med.harvard.edu/aanlib/home.html49. All these methods are configured to their default parameters for optimal performance. All experiments were conducted on a desktop computer with an Intel Core i5-4460 Quad-Core CPU running at 3.2 GHz, 16 GB of RAM, and using MATLAB R2020b as the software platform. The comparative methods were evaluated using publicly available implementations where accessible, or re-implemented according to the algorithms described in the respective publications when source code was not provided.

The ability to preserve image details is demonstrated using visual analysis in Figs. (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13) for three images at different noise levels. We begin by presenting the visual results and then present the quantitative analysis using several performance metrics. Figures (2, 3, 4) illustrates the denoising results. We introduce several degrees of Gaussian noise (\(\sigma =\) 10, 20, 30,40) to the ground truth image and implement different techniques for denoising. Figures 2, 3, 4b–c illustrates that OD-CNN and NDiff preserve the external details while the inside details are pixelated. In Figs. 2, 3, 4d–e, the SDPM and NGN approaches preserve the outer margins while the inside details remain pixelated. Figures 2, 3, 4f, g, h demonstrates that UDLF, CNCL, and BP-MEC effectively maintain the edges to some degree; nevertheless, the inside details remain indistinct. Figures 2, 3, 4k, l, m illustrates that the IDM-PM, DCT-ACO, and ANLM approaches are also ineffective in preserving the edges of the images. In Figs. 2, 3, 4b–c, it is obvious that OD-CNN and NDiff preserve the outer edges although the inner features of the image appear blurred. Figures 2, 3, 4i, j illustrates that FDCT and NSCT-BT fail to preserve edge sharpness which results in distortion. Figures 2, 3, 4n, o illustrates that Z-NLM and ONLM distort the image which results in an increased edge width. Figures 2, 3, 4p illustrates that the image generated using proposed method closely resembles the ground truth image while maintaining both internal and external details. Table 1 displays the quantitative evaluation which demonstrated computed values of PSNR, SSIM, FSIM, Entropy, BRISQUE, PIQE, and NIQE for an estimated denoised image derived from the use of several techniques across diverse noise levels in imaging. The Entropy, PSNR, FSIM, and SSIM of the denoising image produced by the proposed method are superior to those of existing techniques and PIQE, NIQE, and BRISQUE values are comparatively lower which is optimal. Figures (5, 6, 7) presents the qualitative results of the MRI2 image. Figures 5, 6, 7b, c demonstrate that the results of OD-CNN and NDiff exhibit pixelation. Figures 5, 6, 7d, e illustrate the results of the SDPM and NGN methodologies wherein the edges remain unclear and pixelated. Figures 5, 6, 7f, g, h illustrates the outputs of UDLF, CNCL, and BP-MEC, while the complex details are absent. In Figs. 5, 6, 7k, l, m, IDM-PM, DCT-ACO, and ANLM are used for image denoising which results in less clear edges that reduce the visual quality of the image. Figures 5, 6, 7b, c presents the outcomes achieved by the OD-CNN and NDiff methodologies which reveals that the outer edges are clear but the inside details are blurred. Figures 57i, j illustrates the results achieved by FDCT and NSCT-BT which reveals that the edges lack clarity and noise persists in the image, hence reduces its visual quality. Figures 5, 6, 7n, o illustrates that the images produced by the Z-NLM and ONLM methods exhibit blurriness and average visual quality. Figures 5, 6, 7p illustrates the findings derived from the proposed methodology which demonstrates that our model effectively minimizes noise to a greater degree than comparative methods. Table 2 lists average error scores achieved by several methodologies on the MRI2 image. The table indicates that the proposed method has superior values in Entropy, PSNR, FSIM, and SSIM compared to other methods. The proposed strategy exhibits lower PIQE, NIQE, and BRISQUE values in comparison to alternative methods which is optimal. Figures (8, 9, 10) illustrates the qualitative results of the denoised image reconstructed using several denoising techniques used in CT image affected by Gaussian noise. In Figs. 8, 9, 10b, c, OD-CNN and NDiff preserve the image structure; however, the emergence of.

Fig. 2
figure 2

Denoising results of the MRI1 image (\(\sigma \approx 10)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) Ours.

Fig. 3
figure 3

Denoising results of the MRI2 image (\(\sigma \approx 10)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 4
figure 4

Denoising results of the CT image (\(\sigma \approx 10)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Table 1 The performance of different denoising algorithms is evaluated based on Entropy, PSNR, SSIM, FSIM, BRISQUE, PIQE and NIQE metrics for MRI1 image.
Fig. 5
figure 5

Denoising results of the MRI1 image (\(\sigma \approx 20)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 6
figure 6

Denoising results of the MRI2 image (\(\sigma \approx 20)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 7
figure 7

Denoising results of the CT image (\(\sigma \approx 20)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) Ours.

Table 2 The performance of different denoising algorithms is evaluated based on Entropy, PSNR, SSIM, FSIM, BRISQUE, PIQE and NIQE metrics for MRI2 image.
Fig. 8
figure 8

Denoising results of the MRI1 image (\(\sigma \approx 30)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 9
figure 9

Denoising results of the MRI2 image (\(\sigma \approx 30)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 10
figure 10

Denoising results of the CT image (\(\sigma \approx 30)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

grainy formations result in unclear edges. Figures 8, 9, 10d, e demonstrates that SDPM and NGN alleviate noise to a certain extent; however, the structure is slightly shifted from its original position. Figures 8, 9, 10f, g, h demonstrates that UDLF, CNCL, and BP-MEC fail to smooth the homogenous zones which results in a fragmented appearance. In Figs. 810k, l, m, we note that the outputs of IDM-PM, DCT-ACO, and ANLM are unclear which makes it difficult to distinguish the borders. Figures 810b, d demonstrates that overall, noise reduction is achieved by the OD-CNN and NDiff approaches; nevertheless, finer details could be enhanced. Figures 8, 9, 10i, j displays the denoised outputs from FDCT and NSCT-BT whereby noise remains visible. Figures 8, 9, 10n, o illustrates that Z-NLM and ONLM unclear the image, hence reduces its visual quality. Figures 8, 9, 10p demonstrates the result obtained from the proposed methodology that reveals that the reconstructed image closely resembles the ground truth image. The result obtained from the proposed approach for image denoising reduces.

the noise level to some degree. Table 3 presents the quantitative findings of all methods across various noise levels, indicating that the Entropy, PSNR, FSIM, and SSIM scores of the proposed approach surpass those of other methods while the PIQE, NIQE, and BRISQUE values of the proposed approach are comparatively low. We note that IDM-PM, DCT-ACO, and ANLM fails to preserve the edges, where interpreting boundary differentiation impossible. SDPM and NGN preserve the edges; nonetheless, noise remains in the image. The result indicates that the structure is preserved by UDLF, CNCL, and BP-MEC; nevertheless, the details in the image are compromised. The image produced by utilizing IDM-PM, DCT-ACO, and ANLM algorithms on the input image fails to maintain structural integrity. SDPM and NGN effectively retain edges; nonetheless, noise persists in the homogeneous areas of the image. The image produced by FDCT and NSCT-BT remains significantly noisy. It indicates that Z-NLM and ONLM cause the image to become blurred, where interpreting any portion of the image unclear. Our method outperforms other methods while preserving edges and maintaining visual quality. The data presented in Figs. 11, 12, 13 and Tables (1, 2, 3) indicates that the proposed approach outperforms other methods. Our noise level estimation approach offers more precise and consistent assessment of noise levels in medical images compared to current leading algorithms. The proposed denoising method generates superior restoration of essential image information as compared to state-of-the-art denoising techniques. It is noticed that the hard thresholding of singular values for the non-local groups is also illustrated in64. But our method differs from64 by using adaptive clustering of non-local patches and a more advanced PCA thresholding using to MP law. In addition to the PCA thresholding in the initial phase, the local approximation is used in the coefficient-wise LMMSE filtering during a subsequent phase which further aids in maintaining image detail when compared with the global LMMSE. The incorporation of the non-local means (NLM) algorithm provides substantial advantages by maintaining complex details and textures while efficiently reducing noise. By leveraging the redundancy of similar pixels across the image and it ensures denoising is adaptive and data-driven while making it mostly effective for handling complex and textured regions. This leads to significantly improved visual quality compared to state-of-the-art methods particularly for images with complex details or challenging low signal-to-noise ratios.

Table 3 The performance of different denoising algorithms is evaluated based on Entropy, PSNR, SSIM, FSIM, BRISQUE, PIQE and NIQE metrics for CT image.
Fig. 11
figure 11

Denoising results of the MRI1 image (\(\sigma \approx 40)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 12
figure 12

Denoising results of the MRI2 image (\(\sigma \approx 40)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) ours.

Fig. 13
figure 13

Denoising results of the MRI2 image (\(\sigma \approx 40)\); (a) Reference image, (b) ONLM, (c) Z-NLM, (d) ANLM, (e) DCT-ACO, (f) IDM-PM, (g) NSCT-BT, (h) FDCT, (i) BP-MEC, (j) CNCL, (k) UDLF, (l) NGN, (m) SDPM, (n) NDiff, (o) OD-CNN, and (p) Ours.

Ablation study

To validate the effectiveness of outlier exclusion and smoothing within the adaptive clustering and denoising process, we performed an ablation study to quantify their individual contributions to overall denoising performance. Specifically, we compared the full proposed method against two reduced variants: (1) a version without the iterative merging phase of adaptive clustering (thereby reducing outlier exclusion), and (2) a version omitting the soft thresholding (smoothing) step based on LMMSE. The evaluation was performed using MRI and CT datasets from the Cancer Imaging Archive under Gaussian noise levels σ = 10, 20, 30, and 40.

We used PSNR, SSIM, FSIM, Entropy, BRISQUE, NIQE, and PIQE as performance metrics. The quantitative results are presented in Table 4. The results show that excluding the merging phase reduced PSNR by an average of 1.2 dB and SSIM by 0.03 across noise levels, indicating the importance of outlier exclusion in preserving structural similarity. Similarly, omitting the LMMSE-based soft thresholding reduced PSNR by 0.8 dB and SSIM by 0.02, while slightly increasing BRISQUE, NIQE, and PIQE scores. These findings demonstrate that both strategies synergistically enhance the method’s denoising effectiveness by balancing noise suppression and detail preservation. Figure 14 presents a comprehensive comparison of quantitative evaluation metrics such as PSNR, SSIM, FSIM, Entropy, BRISQUE, and PIQE across different noise levels (\(\sigma\) = 10, 20, 30, 40) for the proposed method, a variant without the merging phase (no merging), and a variant without the soft thresholding step (no soft thresholding). The results clearly demonstrate that the proposed method consistently outperforms both ablated variants across all metrics and noise levels. Specifically, the proposed method achieves the highest PSNR, SSIM, and FSIM values, indicating superior noise suppression, structural similarity preservation, and feature similarity retention, respectively. In contrast, the no merging and no soft thresholding variants exhibit a noticeable degradation in performance, particularly at higher noise levels. Similarly, the proposed method yields lower BRISQUE, PIQE and NIQE scores across noise levels, signifying improved perceptual quality compared to the ablated variants. The entropy values also suggest that the proposed method better maintains image detail while suppressing noise. These findings collectively validate the individual and combined contributions of the merging phase (outlier exclusion) and the soft thresholding step in enhancing the overall denoising effectiveness of the proposed approach.

Table 4 Ablation study results showing the effect of outlier exclusion and soft thresholding on denoising metrics at varying noise levels.
Fig. 14
figure 14

Quantitative metric comparisons (PSNR, SSIM, FSIM, Entropy, BRISQUE, PIQE, NIQE) for the ablation study across noise levels (σ = 10, 20, 30, 40), comparing the proposed method, no merging variant, and no soft thresholding variant.

In order to evaluate the computational efficiency of the proposed method under varying noise conditions, we measured the average running time at different noise standard deviations (\(\sigma\) = 10, 20, 30, 40) across MRI1, MRI2, and CT images. The results, previously presented in Table 3, were further visualized in a running time versus noise level curve (see Fig. 15) to facilitate a clearer comparative interpretation. This figure demonstrates that while the running time slightly increases with noise density, the increase is relatively modest and remains computationally feasible within clinical and research settings. Additionally, the proposed method maintains a favourable balance between running time and denoising performance, especially at higher noise levels, outperforming several benchmark methods in robustness despite a moderate increase in execution time. This visualization confirms that the scalability of the method with respect to noise level is practically acceptable.

Fig. 15
figure 15

Running time (in seconds) vs. noise level (\(\sigma\) = 10, 20, 30, 40) for MRI1, MRI2, and CT images.

Conclusion

This study proposes a robust noise level estimation method along with a cluster-wise denoising approach. Our noise level estimation approach is more precise and reliable for medical images compared to other methods. The proposed denoising method uses an adaptive clustering method to employ effectively grouping nonlocal patches based on their special features. Each cluster is then accurately denoised using a PCA thresholding strategy. The denoising process begins with the removal of the most of noise through an enhanced singular value hard thresholding system based on MP law. This is followed by further refinement of noise reduction using coefficient-wise PCA domain LMMSE filtering which is a specific method of soft thresholding with locally computed parameters. Our method delivers superior detail preservation and outperforms other comparative methods in image denoising by combining denoising frameworks with an adaptive clustering method. Besides, the NL-Means algorithm reduces noise by estimating weighted averages of the pixels using the similarity of patches. It evaluates patch similarity using a Gaussian-weighted Euclidean distance which ensures further effective noise suppression while preserving textures and structures. Our future work will focus on enhancing the proposed approach and exploring its applications in other medical imaging and computer vision. In particular, we aim to improve image restoration techniques to address noise and alleviate any degradation in medical imagery by ensuring higher accuracy and reliability in clinical applications.