Introduction

Ghost imaging (GI), known as coincidence imaging, two-photon imaging, or correlated-photon imaging, is an unconventional imaging technique in which the image of an object is reconstructed based on the intensity correlation between two light beams. The beam that interacts with the object is referred to as the signal beam, while the other beam, which does not interact with the object but reaches the detector, is called the reference beam.

Pittman et al. were the first to demonstrate GI in 1995, using quantum-entangled photons1. For several years, GI was believed to be feasible only with quantum sources. However, later studies showed that similar correlations could also be achieved using classical light sources - indicating that ghost imaging is not exclusive to quantum mechanics but can be explained through classical physics as well. In particular, Bennink et al. were among the first who reconstructed GI with classical light in 2002, which validated the classical interpretation2,3,4,5,6.

The concept of Computational Ghost Imaging (CGI) was first proposed by Shapiro7, introducing a new method that eliminates the need for a reference beam by using patterns calculated offline. This innovation simplified the optical setup and made the practical implementation of ghost imaging more feasible. In 2009, Bromberg et al. demonstrated CGI experimentally using a single-pixel detector8. Over time, ghost imaging has evolved into an effective platform for developing novel imaging technologies and cost-effective imaging systems. Its applications span a wide range, including imaging through turbulence9,10,11,12, lensless imaging13,14, x-ray imaging15,16,17, 3D imaging with single-pixel detectors18,19, compressive radar20, terahertz single pixel imaging21,22. These advances highlight the growing importance of CGI in real-world applications.

In response to the limitations of traditional computational ghost imaging (CGI) methods, this study explores a novel approach called Orthogonal Ghost Imaging (OGI), which employs two-dimensional Discrete Cosine Transform (2D-DCT) patterns. Unlike random or sinusoidal illuminations, the use of orthogonal DCT basis functions allows for more efficient sampling and sharper reconstructions. A new image reconstruction formula is derived based on the inner product properties of these orthogonal patterns. Furthermore, to reduce environmental noise, a phase-shift modulation technique is incorporated. The efficacy of the proposed method is validated through both simulation and experimental results, showing significant improvements in image quality and reconstruction performance.

Over the years, researchers have developed various reconstruction methods to overcome the limitations of traditional CGI. For example, correlation of intensity powers23, Differential Ghost Imaging (DGI)24, Normalized GI25, and Pseudo-inverse GI26 were among the first efforts to improve the signal-to-noise ratio (SNR) of the reconstructed images. More recently, Singular Value Decomposition (SVD)-based techniques27,28,29 have been applied to extract meaningful components from measurement data more effectively. In parallel, compressive sensing methods have gained attention for their ability to reconstruct images from significantly fewer measurements by exploiting signal sparsity30,31,32,33. These methods rely on the assumption that most natural images are sparse in some transform domain. Another effective method has involved the use of structured and orthogonal illumination patterns. For example, sinusoidal ghost imaging (SGI)34and orthogonal basis approaches35,36,37,38 aim to improve overlap between the patterns and the target object’s spatial features, which enhances reconstruction performance. These studies demonstrate that exploiting spatial frequency characteristics of natural images can significantly enhance image quality and efficiency.

More recent research has explored diverse approaches to further enhance the efficiency and versatility of ghost imaging. For example, hybrid transform-based methods that integrate DCT, Hadamard, and Haar matrices have demonstrated improved image reconstruction performance with fewer measurements39. Other approaches focus on data augmentation to enhance the quality of ghost imaging40, or leverage compressive sensing in plenoptic imaging setups to recover spatial and angular information simultaneously41. These recent findings show how ghost imaging continues to evolve toward more flexible and efficient designs, supporting the value of simple, low-shot reconstruction approaches like the one presented here. Our method builds upon these ideas by proposing an orthogonal pattern-based framework with improved computational efficiency and noise robustness.

Model and principle

The schematic experimental setup of CGI is shown in Fig. 1. The target object is illuminated by a projector with random intensity patterns \(\{I_{r}(x,y)\}\), and a bucket detector measures the corresponding transmitted intensity\(\{B_{r}\}\). The image of the target scene is usually reconstructed through Bromberg Equation as8:

$$\begin{aligned} G_{Bro}(x,y)&=\frac{1}{N}\sum _{r=1}^{N}(B_r-<B>)I_r(x,y)\nonumber \\= & lt;BI(x,y)>-<B><I(x,y)>, \end{aligned}$$
(1)

where \(\langle \cdot \rangle = \frac{1}{N} \sum _{r=1}^{N} \cdot\) denotes an ensemble average over N phase realizations. A higher value of B means more overlap of random intensity patterns with the target, and these patterns have the greatest effect on image reconstruction. Another algorithm that increases the quality of ghost imaging is differential ghost imaging (DGI)24:

$$\begin{aligned} G_{D}(x,y)=<BI_r(x,y)>-\frac{<B>}{<I_r>}<I_rI_r(x,y)>, \end{aligned}$$
(2)

where \(I_r\) is the sum of all pixels of the illumination pattern. Random patterns are used for both reconstruction methods mentioned above. In this research, orthogonal functions of DCT as illumination patterns are used instead of random intensity patterns, one formulate as:

$$\begin{aligned} s(x,y;u,v)=\sum _{x=0}^{m_{1}-1}\sum _{y=0}^{m_{2}-1} \cos \left( \frac{2x+1}{m_{1}}u\pi \right) \cos \left( \frac{2y+1}{m_{2}}v\pi \right) , \end{aligned}$$
(3)
Fig. 1
figure 1

Schematic of the optical setup used for ghost imaging experiments, including the laser source, beam splitter, DMD, object plane, and bucket detector.

where xy are the 2D cartesian coordinates, \(\{u = 0, 1, ..., m_{1}-1\}\) and \(\{v = 0, 1, ..., m_{2}-1\}\) are the horizontal and vertical spatial frequencies, respectively, and \(\{m_{1}\times m_{2}\}\) is pixel resolution of image. The following matrix is obtained for each u, v by sweeping x, y from 0 to \({m_{1}-1}\) and 0 to \({m_{2}-1}\), respectively:

$$\begin{aligned} S_{uv}= \begin{bmatrix} s(0,0;u,v) & s(0,1;u,v) & \cdots & s(0,m_{2}-1;u,v) \\ s(1,0;u,v) & \vdots & \cdots & \vdots \\ \vdots & \vdots & \cdots & \vdots \\ \vdots & \vdots & \cdots & \\ \vdots & & & \\ s(m_{1}-1,0;u,v)& s(m_{1}-1,1;u,v) & \cdots & s(m_{1}-1,m_{2}-1;u,v)\\ \end{bmatrix}. \end{aligned}$$
(4)

In image processing42, the above matrices are called the mapping matrices from the cartesian space to the DCT space, which are known as the basis of DCT space. The image is described using these basis as follows:

$$\begin{aligned} F(x,y)=\sum _{u=0}^{m_{1}-1}\sum _{v=0}^{m_{2}-1}T_{uv}S_{uv}. \end{aligned}$$
(5)

where \(T_{uv}\) represents the weighting coefficients of each of the basis of the DCT space in image reconstruction. Each of the matrices illuminates the object as illuminating patterns, and the Bucket detector records the intensity of the reflected pattern. According to the property of orthogonal basis, we define \(<S_{uv},S_{u^{\prime } v^{\prime }}>=||S_{uv}||^2 \delta _{uu^{\prime }}\delta _{vv^{\prime }}\) where \(\delta\) is Kronecker delta. Therefor, \(T_{uv}={<S_{uv},F>}\). Where \({<S_{uv},F>}\) is the value measured by the bucket detector in GI setup. The formula of ghost imaging reconstruction is as follows:

$$\begin{aligned} G_{orth}={<B_{uv},S_{uv}>}. \end{aligned}$$
(6)

where \(B_{uv}\) is intensity of signal measured by detector.

The presence of negative intensity is necessary to prevent intensity saturation during image reconstruction. However, a challenge is the way the Bucket detector works so that the detector is not able to distinguish between the environmental noise and the signals received from the object. To solve this problem, the phase-shift method can be used, which leads to the elimination of noises that are statistically identical35,36,37,38. This phenomenon is especially evident when the sampling rate significantly differs from the flashing rate of environmental light sources, resulting in substantial noise reduction from environmental illumination. For this purpose each weighting factor of these patterns requires two measurements with patterns that have a phase difference of \(\pi\). Therefore, we use the following two groups of patterns:

$$\begin{aligned} {\left\{ \begin{array}{ll} S(x,y;u,v)=\sum _{x=0}^{m_{1}-1}\sum _{y=0}^{m_{2}-1} \cos \left( \frac{2x+1}{m_{1}}u\pi \right) \cos \left( \frac{2y+1}{m_{2}}v\pi \right) ,\\ S^{\prime }(x,y;u,v)=-\sum _{x=0}^{m_{1}-1}\sum _{y=0}^{m_{2}-1} \cos \left( \frac{2x+1}{m_{1}}u\pi \right) \cos \left( \frac{2y+1}{m_{2}}v\pi \right) \end{array}\right. } \end{aligned}$$
(7)

Figures 2(a)-(b) show part of two sets of 2D-DCT \((m_{1}=m_{2}=64)\) patterns generated from Eq. (7), respectively. The cycle frequency increases by 1/2 with each step from left to right or from top to bottom. As it is clear from the Fig. 2, the frequency patterns will become more intense with each step moving to the right and down of Figs. 2(a)-(b).

Fig. 2
figure 2

Part of the two sets of illumination patterns for \(m_{1}=m_{2}=8\) used in this paper. (a) Patterns generated by S(xyuv) in Eq. (7), (b) Phase-shifted counterparts generated by \(S^{\prime }(x,y;u,v)\) in Eq. (7). These two groups form the basis of the differential acquisition strategy used to eliminate background noise.

The detector can collect each type of light, whether direct or indirect, or a combination of both for imaging. Thus, the total response of the detector received for each series of orthogonal basis can be expressed as follows:

$$\begin{aligned} {\left\{ \begin{array}{ll} D_{uv} = \langle T(x,y) S(x,y;u,v) \rangle + b_{DC}, \\ D^{\prime }_{uv} = \langle T(x,y) S^{\prime }(x,y;u,v) \rangle + b_{DC}, \end{array}\right. } \end{aligned}$$
(8)

where T(xy) is the transmission function of an object and \(b_{DC}\) is the environmental background light. The following differential equation can annihilate the \(b_{DC}\) term and obtain the spectrum of DCT coefficients without background noise:

$$\begin{aligned} B_{uv}=\frac{1}{2} (D_{uv}-D^{\prime }_{uv}). \end{aligned}$$
(9)

Figure 3(a) shows the order of displaying patterns on the object. By changing this order, as shown in Fig. 3(b), and adopting a zig-zag scanning strategy that prioritizes low-frequency DCT components, the image can be reconstructed with fewer measurements-demonstrating that the projection order significantly influences reconstruction efficiency.

Fig. 3
figure 3

Two different display orders of illumination patterns projected onto the scene for \(m_{1}=m_{2}=4\). (a) Default scanning order; (b) Zig-zag style ordering inspired by compressive sampling, which prioritizes low-frequency components earlier in the reconstruction processs.

Another challenge is using cosine functions, because projector can not display negative values. To solve this issue, we normalize the bucket data to the range of (0,1). Accordingly, the intensity corresponding to a black pattern illumination is set to 0, and for a white pattern illumination, it is set to 1. Subsequently, applying \(B_{uv}=2 B_{uv}-1\) leads to the generation of the corresponding negative intensities. The subtraction in Eq. (9) implements a differential phase-shift strategy designed to remove static environmental interference and system offsets.

Our method differs from previous orthogonal ghost imaging approaches in several important ways.These differences are mainly visible in the convergence speed, robustness to noise and implementation complexity. For instance, the technique described in35 uses a block-style ordering of DCT patterns. In contrast, we apply a zig-zag scan that starts with low-frequency components. This choice helps capture the image’s overall structure early on, which speeds up convergence and tends to yield better reconstructions, especially when measurements are limited.

In36, the authors use sinusoidal illumination patterns based on Fourier components, which are inherently complex-valued and require four separate projections per component to recover both real and imaginary parts. Our approach, by contrast, uses real-valued symmetric DCT patterns, needing only two complementary measurements. This significantly simplifies the setup and improves acquisition efficiency. Also, the method in38 performs DCT projections along each axis separately - first horizontal, then vertical - which effectively splits the image into two one-dimensional components. This method lacks full 2D coverage, which can lead to slower convergence and reduced accuracy in preserving important spatial details such as edges and textures. In contrast, we use full 2D-DCT patterns that better match the structure of natural images. This allows for faster convergence and improved image detail preservation43,44,45,46,47.

Experimental setup

A digital light projector (Casio XJ-A146) was used to illuminate the scene using two groups of 2D-DCT patterns with spatial frequencies (uv). These patterns, with varying spatial frequencies, were generated at a resolution of \(64 \times 64\) pixels and 256 grayscale levels using a computer. The projector was placed at a fixed distance of 72 cm from the target and illuminated a \(2 \times 2~cm^2\) area of the scene for 0.15 seconds per pattern.

The transmitted light corresponding to each pattern was simultaneously measured by a Thorlabs DCC1545M camera, which acted as a bucket detector. To fully reconstruct a \(64 \times 64\) pixel image, a total of 8192 measurements were acquired. The complete acquisition process took approximately 68 minutes. Using this setup, we carried out both simulations and experiments to evaluate how well the proposed method performs in practice. The results are presented and discussed in the following section.

Discussion and results

We initiate our evaluation process with numerical simulation to compare the image quality produced by the novel approach with the other existing methods. Here, the target object for simulation is an image of \(200\times 200\) pixels with a square aperture of \(100\times 100\) pixels shown in Fig. 4a(I). DCT matrices are generated by the computer, and after interrogating the object matrix, the set of new matrix elements is reported as the intensity of the detector. The reconstructed image using Eq. (6) for 2000 shots can be seen in Fig. 4b(I). To quantitatively compare image quality, the parameter of \(SNR=\bar{s}^2/\sigma _n\) is used34,42, where \(\bar{s}\) is the average of the signal intensity, and \(\sigma _n\) is the variance of the background intensity.

Fig. 4
figure 4

Numerical simulation of a \(200\times 200\)-pixel image containing a square aperture of size \(100\times 100\). (a) Target image; (b) Reconstructed image using DCT-based illumination and Eq. (6) with 2000 shots. Subfigures ((I) and (II) show the reconstructed image and the intensity distribution along its central horizontal line, respectively, highlighting the spatial fidelity of the reconstruction.

Table 1 SNR comparison of DGI34, SGI34, and the proposed OGI method under equal sampling conditions.

In this paper, the difference in average intensity between the bright and dark areas of images conceptualizes as the signal, while the variations in the dark background conceptualizes as the noise. We compare the SNR result of the reconstructed image in Fig. 4 with the SNR values obtained from two methods of the DGI24 and the SGI34, presenting the results in Table 1.

In the SGI approach, the illuminant matrices are generated by the sinusoidal patterns34, and the image is reconstructed by \(G_{SGI}=\sum _{i=1}^{N}{B_{i}I_i(x,y)}\), which N is number of shots, \(I_i(x,y)\) is the sinusoidal patterns and \(B_{i}\) is the intensity corresponding to each pattern that is recorded by the detector after interacting with the object. As indicated in Table 1, the SNR value of the DGI method is 10.89 dB for \(10^5\) number of shots. While in the SGI method, despite a remarkable reduction in the number of shots, we have a significant improvement in the quality of the reconstructed image, and SNR is \(33.28\) \(dB\). As Table 1 shows, the SGI method improves image quality compared to DGI, achieving a higher SNR despite using fewer measurements. However, our proposed OGI method outperforms both, delivering the highest SNR (36.95 dB) while requiring even fewer shots than SGI. This demonstrates that modifications to the reconstruction equation can remarkably improve the quality of the reconstructed image.

Increasing the number of shots leads to sharper edges with higher visibility. But the question is, “How many shots are satisfactory as a lower bound to reconstruct an image?” The answer to this question depends on two factors. The first is related to the illuminant matrices, which according to the results of Table 1, DCT matrices are more satisfactory in the number of fewer shots, and the second is related to the level of image sparse. A sparse image is one in which most of the pixels are zero or close to zero, and the more sparse the image is, the fewer shots are needed for reconstruction43,44.

To study the image sparse effect, we choose a group of images (\(200\times 200\) ) pixels with different square aperture dimensions (non-zero image components) (\(40\times 40\)), (\(60\times 60\)), (\(80\times 80\)), (\(100\times 100\)) pixels which can be seen in Fig. 5a(I-IV) respectively.

We estimate the imaging quality by representing the peak signal-to-noise ratio (PSNR), which can be described by42:

$$\begin{aligned} PSNR=10\log _{10}\left( \frac{max({U_{0}(x,y)})^{2}}{\frac{1}{m_{1}m_{2}}\sum _{x=1}^{m_{1}}\sum _{y=1}^{m_{2}}[U_{0}(x,y)-U(x,y)]^2}\right) , \end{aligned}$$
(10)

where \(U_{0}\) is the target image matrix (\(m_{1}\times m_{2}\)) pixels, and U is the corresponding reconstructed image. The PSNR for the different apertures at corresponding sampling ratios (\(\beta\)) of 1, 2, 5, 10 and \(20\%\) are shown in Table (2). The sparsity of the image evidently has a strong influence on reconstruction quality. As the non-zero region of the image increases-from \(40\times 40\) to \(100\times 100\)-the PSNR values consistently decrease under a fixed sampling rate. This behavior aligns well with the principles of compressive sensing, which state that sparser images require fewer measurements to achieve reliable reconstruction. In accordance with Shannon’s entropy, for an image with n pixels that only k pixels have non-zero, if \(k\ll n\), the minimum number of measurements required for successful reconstruction is approximately given by47:

$$\begin{aligned} -\frac{k}{n}\hspace{.2cm}log\frac{k}{n}-\left( 1-\frac{k}{n}\right) \hspace{.2cm}log\left( 1-\frac{k}{n}\right) \approx \frac{k}{n}\hspace{.2cm}log\frac{k}{n} \approx k\hspace{.2cm}log\frac{n}{k}. \end{aligned}$$
(11)

This expression is derived from the binary Shannon entropy and is commonly used in compressive sensing theory to estimate the minimum number of measurements required for accurate recovery of sparse signals. It assumes that the support of the sparse signal follows a Bernoulli distribution, where each component is non-zero with a fixed probability. Although real-world images often exhibit structured (rather than random) sparsity, the Bernoulli model is widely adopted in theoretical analysis as it provides a general information-theoretic lower bound. This approach helps characterize the fundamental limits of compressive acquisition independently of specific image content44,45,46.

Fig. 5
figure 5

(a) (I)-(IV) Target binary images with square apertures of sizes (\(40\times 40\)), (\(60\times 60\)), (\(80\times 80\)), (\(100\times 100\)), respectively. (b) (I)-(IV) Reconstructed results using the proposed OGI method with 5150 measurements. The comparison illustrates how increasing sparsity (i.e., smaller apertures) leads to better reconstruction performance under fixed measurement conditions.

Table 2 PSNR (dB) values for reconstructed images using different aperture sizes and sampling rates.

In our simulation, the full image contains \(200 \times 200\) pixels , but only a \(40 \times 40\) central region contains non-zero values-meaning just 1600 pixels carry information and the rest remain zero. This makes the image highly sparse, with \(k=1600\) out \(n=40000\). Under this condition (\(k\ll n\)), using 5150 measurements meets the lower bound predicted by Shannon’s entropy, as described in Eq. (11). Figure 5b (I-IV). shows the reconstructed images for four different aperture sizes, all obtained using the same number of measurements (5150 shots). While the overall structure is visually preserved across all cases, the sparsest one \(40 \times 40\) achieves a PSNR above 30 dB. This results suggests that the method remains reliable even with limited sampling, and performs remarkably close to the theoretical lower bound predicted by Shannon entropy.

Fig. 6
figure 6

Reconstructed grayscale images of four standard test images (“house”, “boat”, “lake”, and “livingroom”) using the proposed OGI method at sampling rates of \(10\%\), \(30\%\), \(50\%\), and \(60\%\). The results show that visual quality improves with higher sampling, but becomes nearly saturated at around \(30\%\), confirming the method’s efficiency under compressed conditions.

To further evaluate the robustness of the proposed method on more realistic content, standard grayscale test images commonly used in image processing-namely, “house”, “boat”, “lake”, and “livingroom”-were also included in the simulations. Figure 6 (a-d) shows how the proposed method performs on these images at different sampling rates. As expected, the most noticeable jump in reconstruction quality happens around \(30\%\). Adding more measurements beyond that point yields only minimal visual enhancement, confirming a point of diminishing returns. This isn’t just something we observed; it’s supported by theory. As discussed in45,47, natural images tend to have a sparse representation in the DCT domain, where a small number of low-frequency components capture most of the structural information. That’s exactly what the DCT basis is good at-it captures essential features efficiently.

This alignment between the structure of natural images and the design of our method likely explains why the system performs so well even with relatively low sampling rates. In our method, the \(30\%\) sampling rate yields a PSNR above 30 dB, which confirms that the method remains robust and effective under sub-Nyquist conditions, and operates near the theoretical efficiency limit predicted by information theory. This can be explained by the fact that most of the coefficients in DCT-based representations are either zero or close to zero, while approximately one-third carry most of the meaningful information. These larger coefficients, due to strong alignment with image structures, play a key role in reconstructing the fine details with fewer measurements. To experimentally validate the performance of our proposed method, we implemented the setup shown in Fig. 1 collected experimental measurements. The results are shown in Figure 7a(I-V), which shows the reconstructed images at different sampling rates. The experimental reconstructions clearly improve as the sampling rate increases, particularly in terms of edge sharpness and structural clarity. The \(30\%\) sampling level appears to mark a practical threshold in the experimental setup-preserving essential image structure and fine edges without requiring significantly more measurements.

Fig. 7
figure 7

(a) Experimental reconstruction results using the proposed OGI method at sampling rates of 2, 10, 20, 30 and \(40\%\). (b) Corresponding 3D intensity surface plots for each case, where x and y represent spatial pixel positions and z indicates intensity. The increasing sharpness and edge definition in the surfaces highlight how the method preserves structural detail even under limited measurements.

To get a more intuitive sense of how measurement count influences reconstruction-especially near object boundaries-we also generated 3D intensity surface plots from the same experimental data, shown in Figure 7b(I-V). These visualizations help emphasize subtle differences that aren’t always easy to spot in the grayscale images alone. In the 3D plots, the surfaces become noticeably sharper and more defined as the sampling rate increases. The steepness and clarity of the edges, in particular, improve significantly, making it easier to assess the system’s ability to recover fine spatial details. This form of visualization not only complements the 2D results visually, but also gave us deeper insight into how edge contrast evolves across different sampling conditions. These experimental findings remain consistent with the entropy-based expectations introduced earlier, reinforcing the theoretical foundation observed in the PSNR analysis. Even with a limited number of measurements, the reconstructed images maintain both visual clarity and structural detail-suggesting that the method operates close to the information-theoretic limit described by Shannon.

Up to this point, we’ve looked at the quality of the reconstructed images-but understanding how much computational effort each method requires is just as important. An algorithm isn’t truly efficient if it only works well on specific hardware. For this reason, we also assess computational complexity, quantified by the number of additions and multiplications, to provide a clearer, hardware-independent view of how computationally efficient each method really is42. To better understand how the two methods compare in terms of computational cost, we examined the number of arithmetic operations required to implement GIT (Eq. (1)) and OGI (Eq. (6)), where N is the number of measurements and \(m^2\) is the image resolution. The results of this comparison are summarized in Table 3.

We observed that the computational advantage becomes more apparent as the number of measurements increases. Specifically, the complexity ratio is given by \(C(N)={(3Nm^2+4m^2+N+1)}/{(2Nm^2)}\) ,which simplifies to approximately 1.5. In other words, when sampling is high, OGI cuts the number of operations by about one third compared to GIT. This reduction translates into a noticeably more efficient reconstruction process. Table 4. illustrates this trend in practice for a \(64\times 64\) image at various measurement levels. The data confirms what we expected: the computational load of OGI stays consistently lower, while reconstruction quality remains high. All in all, the results show that OGI doesn’t just perform better-it does so without relying on special hardware, and even with fewer measurements.

Table 3 Total number of additions and multiplications required to implement the reconstruction formulas in Eq. (1) and Eq. (6), representing GIT and the proposed OGI method, respectively.
Table 4 Calculated values of the computational complexity ratio C(N) for various numbers of measurements (N), comparing the efficiency of OGI versus GIT. The results show that as N increases, the advantage of OGI approaches a consistent factor of 1.5.

Conclusion

In this paper, we proposed an orthogonal ghost imaging (OGI) framework based on two-dimensional discrete cosine transform (2D-DCT) patterns, integrated with a two-phase shift acquisition strategy to effectively address the issue of negative intensity values. This method benefits from the inherent energy compaction and sparsity properties of the DCT basis, which allow for efficient sampling and accurate image reconstruction with fewer measurements.

Through extensive simulations and physical experiments, we demonstrated that the proposed method consistently outperforms traditional ghost imaging techniques. It achieves peak signal-to-noise ratios (PSNR) exceeding 30 dB at sampling rates as low as \(30\%\), while also reducing computational complexity by approximately one-third compared to benchmark methods. These improvements highlight the method’s practicality, especially in scenarios where minimizing acquisition time and computational resources is critical.

The strong agreement between simulation and experimental results further validates the robustness of the proposed approach and its applicability in real-world imaging setups. Even under limited sampling conditions, the method maintains high visual fidelity and structural accuracy, operating near the theoretical efficiency bounds predicted by compressive sensing theory.

Looking ahead, the flexibility of the OGI framework makes it well-suited for further extensions, including color image reconstruction, real-time processing, and imaging under dynamic lighting or noisy environments. These directions open up promising pathways toward developing scalable, low-cost, and hardware-friendly ghost imaging systems for practical applications.