Abstract
To address the burgeoning demand for computing capacity in artificial intelligence, researchers have explored optical neural networks that show advantages of ultrafast speed, low power consumption, ultra-high bandwidth, and high parallelism. However, most existing optical networks are reciprocal, where forward and backward propagation are intrinsically coupled. This results in the backward pathway remaining largely unexplored, hindering the realization of integrated perception-response systems. Here, we present a nonreciprocal neural network leveraging enhanced magneto-optical effect in spoof surface plasmon polaritons transmission line to decouple forward and backward paths. Moreover, the computing function of the network can be flexibly modulated by the magnetization orientation in ferrites and variations in operating frequency. We demonstrate broadband bidirectional decoupled image processing across various operators, where the operator configuration can be precisely designed by encoding the input signals. This decoupling achieves independent control and signal isolation within the same structure, effectively emulating the unidirectional transmission of biological networks. Furthermore, matrix-solving operations can be facilitated by incorporating feedback waveguides for desired recursion paths. Our findings open pathways to nonreciprocal architectures for independent bidirectional algorithms in analogue computing.
Similar content being viewed by others
Introduction
As machine learning revolutionizes the fields of science and technology1,2, there has been an exponential surge in demand for computing power in artificial intelligence3,4,5. Optical analog computing offers a promising solution to meet this demand6,7,8,9, providing ultra-fast speed, low power consumption, ultra-high bandwidth, and high parallelism. It serves as an effective platform for realizing the next generation of ultra-fast, compact, and efficient hardware neural networks10,11,12,13,14,15,16. Previous studies have demonstrated the feasibility of implementing deep neural network algorithms using all-optical integrated photonic circuits17,18,19,20,21. Subsequently, researchers have successfully mapped optical systems to parameterized neural networks through all-optical matrix multiplication22,23,24. The Lu group proposed an optical processor that enables deep learning through symmetric photon cores7. This optical processing technique ensures that both data and error propagation share the same direction, thereby facilitating purely forward-mode training of optical neural networks. However, most existing networks are reciprocal, i.e., the time-reversal symmetry of the basic module is not broken, resulting in the constraints of the forward and backward scattering matrices (\({{{{\bf{S}}}}}_{12}={{{{\bf{S}}}}}_{21}^{{{{\rm{T}}}}}\)). Therefore, the forward propagation (FP) and backward propagation (BP) functions of the reciprocal network are always coupled to each other, making it impossible to achieve independent control and signal isolation within the same structure. In a typical perception-response integrated system, an additional directional module is often required to activate a particular network (FP or BP), to achieve the decoupling of perception and response functions, avoiding signal interference and the disorder of functions.
Biological neural systems offer a compelling model to address this limitation. They are inherently nonreciprocal, embedding directional selection into their networks25. As illustrated in Fig. 1a, the inter-neuronal communication within the human body occurs exclusively in one direction, offering two distinct and independent paths for information transmission between organs26,27,28. In other words, perception and response in the nervous system are decoupled through direction in the networks. For the same stimulus event (Fig. S1), the information transmission between the brain and the organs can be compared to a circuit network: the brain (output port) collects inputs from all sensing organs (input ports), which is perception (FP). Next, the brain (output port) selectively responds (BP) to specific target organs (input ports), without interfering with the original organs (input ports). This intrinsic nonreciprocity avoids the conflict among different functions, guaranteeing the orderliness and efficiency of nervous system functionality29,30. Based on this mechanism, bidirectional recurrent neural networks employed in computers also encompass two separate information transmission pathways that are independent of each other31,32. This allows bidirectional recurrent neural networks to simultaneously utilize input information in the past, current, and future33,34. Therefore, learning from biological neural systems, nonreciprocal networks represent a promising yet largely unrealized pathway toward achieving highly integrated and resource-efficient optical computing systems.
a The unidirectional transmission properties of synapses establish two independent paths between human organs. b Leveraging magneto-optical effects, the microwave computing network achieves effective decoupling of forward and backward processing paths. c Demonstration of three distinct functionalities. Nonreciprocal propagation: FP and BP paths perform different computational tasks. Frequency division multiplexing: signals at different frequencies in the same direction exhibit distinct computational functions. Matrix solving: by incorporating a feedback mechanism, the network is capable of executing matrix operations.
Recently, there has been growing interest in nonreciprocal signal process. Various approaches have been explored, including magneto-optical effects35,36, optical nonlinearity37, spatiotemporal modulation38, and transistor loading39. While time-modulated systems and active transistors offer high integration and low cost, their performance is often limited by noise and low power handling. In contrast, magneto-optical materials are attractive for their stable and directional response35,36. Nonetheless, most nonreciprocal research (such as phase shifters40, isolators41, circulators42) remains focused on device-level functions for improving communication stability and suppressing reflections. The intrinsic directional decoupling of nonreciprocal networks has not been widely explored.
In this paper, we present a nonreciprocal spoof plasmonic neural network that can perform microwave data procession with independent functionalities in two directions. The nonreciprocity is introduced by covering the spoof surface plasmon polaritons (SSPP) waveguides with magnetic-biased ferrite coverages, as illustrated in Fig. 1b. The interaction between confined circular polarized fields and static magnetic fields breaks the time-reversal symmetry in FP and BP paths, enabling independent control over the weight matrices for each unidirectional path. To maximize the magneto-optical effect and transmission efficiency, a curved SSPP line that exhibits local field enhancement and enlarged mode impedance is designed and integrated into a basic 2-input 2-output cell. By staggered interconnecting reciprocal and nonreciprocal cells, a four-layer deep neural network with diverse computational functionalities is constructed. Bidirectional decoupled image processing using various operators is demonstrated to validate the nonreciprocal propagation capability of the network (Fig. 1c). The loaded operators can be flexibly adjusted through the encoding of the input signals. The FP performs regional integration to emulate global perception, while the BP executes edge detection to simulate localized response, forming a bidirectional decoupled perception-response integrated neural network. The bidirectional functionality within a single structure eliminates the need for external directional modules and reduces the complexity of the system. As a result, this simplified design can improve resource efficiency and integration density, especially for large-scale, densely connected networks. Furthermore, recursion can be introduced by incorporating feedback waveguides to enable matrix-solving operations. The solvable matrices can be tuned by adjusting the propagation direction, operating frequency, and magnetization orientation. Our findings open up a new route for building multifunctional, highly integrated, and resource-efficient neural networks.
Results
Design of nonreciprocal cells and network strategy
The 8-input 8-output nonreciprocal network consists of 4 reciprocal cells and 3 nonreciprocal cells, as shown in Fig. 2a. With well-matched mode impedances, the ports of these computational cells can serve as the basic low-reflection nodes in the multi-port network. Through layer-by-layer cascading of nodes, we construct an integrated computing network capable of reconfigurable multi-channel transmission. It is worth noting that we use staggered interconnected cells to cascade four layers of nodes. This approach enhances coupling between cells, allowing flexible adjustment of nonreciprocal transmission of the overall network. In this interleaved, interconnected structure, there are not only cross-connections and direct connections between nodes across layers, but also interconnections among nodes within each layer. Notably, the nodes exhibit four distinct connection modes: reciprocal (R), nonreciprocal (NR), FP, and BP connections. R and NR connections work simultaneously in bidirectional pathways. In contrast, FP or BP connection operates through a single pathway. In other words, FP and BP connections can only be “activated” by signals reflected from the (i + 1)th layer back to the ith layer. Therefore, the interleaved interconnections of cells achieve high energy utilization of the network. Furthermore, all connections exhibit notable differences in weight magnitudes. The weights of connections are 0.5, 1, or \(\sqrt{2}/2\).
a The 8-input, 8-output deep neural network composed of low-reflection nodes and four types of connections. The network features R and NR connections for bidirectional paths, as well as FP and BP connections for unidirectional paths. b Design strategy of the S-matrix. The S-matrix of the multi-port network is derived from the S-matrices of its constituent units. c Physical implementation of the nonreciprocal computing network using SSPP waveguides. The 8-input, 8-output network is constructed by interleaving cells, each comprising a power divider and connections. NR connections are realized by NRPS. d Simulated dispersion curves. Comparisons of the dispersion curves for light, microstrip (MS), and SSPP waveguides, which highlight the near-field properties of SSPP waveguides. Under a magnetic bias, NRPS exhibits asymmetric forward and backward dispersion. e DPS spectrum under magnetic modulation. The DPS of NRPS varies with the magnitude and direction of the bias magnetic field. f Amplitude spectra of three typical S-parameters. The bidirectional amplitude transmission remains symmetric. g Phase spectra characteristics. The first three processed groups (passing through NRPS) exhibit significant forward-backward splitting, while the fourth group (without passing through NRPS) remains overlapped.
Interestingly, the introduction of nonreciprocity expands the transformation space of the T-matrix of a multi-port network. The T-matrix of the network is mainly obtained by multiplying the sub-matrices of each cell. In reciprocal networks, all connections in cells are subject to the constraint of det (TR) = 1, limiting the overall transformation space. In contrast, for the nonreciprocal networks, this determinant constraint on the connections is lifted, i.e., det (TNR) ≠ 1 (see Supplementary Text 1 for details). Therefore, the nonreciprocity introduced by the connections can add degrees of freedom (DOF) for the design of the T-matrix of the cascaded network, thereby expanding the scope of the transformation space. The enlarged transformation space improves the bidirectional computing performance. In this situation, the S-matrix of the network becomes asymmetric (Fig. 2b), thereby satisfying the prerequisite for independent bidirectional transmission matrices depicted in Fig. 1b. This provides a theoretical basis for bidirectional decoupled propagation in multi-port networks.
To achieve the asymmetric matrix, the NRPS controlled by a static bias magnetic field is utilized. The NRPS in Fig. 2c comprises ferrite, SSPP transmission lines, and a trapezoidal transition structure. Time-reversal symmetry can be broken by the magneto-optical effect, when the spin direction of a circularly polarized field aligns with the magnetic bias in gyromagnetic media. Upon application of an external magnetic field, the dispersion curves for forward and backward waves split as shown in Fig. 2d, with the difference between them defined as the differential phase shift (DPS). To enhance this effect, a pure circularly polarized field is needed40. However, the generation of circularly polarized modes is challenging in ordinary transmission lines. Here, we employ a curved SSPP transmission line to synthesize circularly polarized fields at equidistant points of adjacent wires in Fig. S4a. By periodic arrangement of quasi-TEM cells to introduce a specific phase shift, circular polarization can be formed (in Fig. S4b). Therefore, the magneto-optical effect is enhanced in a minimal area. To demonstrate the effect of curvature on the magneto-optical effect in SSPP waveguides, three different types of SSPP waveguides are compared in Fig. S6. Details are shown in Supplementary Text 2. Unlike surface plasmon polaritons at optical frequencies, SSPPs are artificially structured to operate at microwave frequencies, offering greater design flexibility and frequency scalability43,44 (for more details, see Supplementary Text 3). As shown in Fig. S6, the DPS per unit length along the propagation direction for the large-curved SSPP, weak-curved SSPP, and microstrip line are 11.54°/mm, 4.04°/mm, and 0.22°/mm, respectively. The large-curved SSPP has more obvious nonreciprocity in unit length; thus, the volume of ferrite required is effectively reduced for minimization.
The simulation results of the NRPS are shown in Fig. 2d. The propagation constant of the SSPP waveguide exceeds those in microstrip and optical lines (e.g., 0.88, 0.26, and 0.18, respectively). It thus offers slow-wave and strong confinement characteristics that are beneficial in low-crosstalk miniaturization45. Notably, one design challenge is the mode matching. Due to the high dielectric constant of the substrate, a larger DPS requires an exceedingly narrow SSPP line, posing challenges for interconnection. To address the issue, a gradient-progressive transition structure is thus designed to achieve optimal impedance matching. Further details of this structure can be found in Supplementary Text 4. The DPS of the NRPS can be continuously tuned from 80° to 280° as the static magnetic field increases, as shown in Fig. 2e.
The designed NRPS can be cascaded to a 50:50 power divider (i.e., the R cell), forming the 2-input 2-output NR cell in Fig. 2c. Subsequently, the 8-input 8-output network is constructed using four R and three NR cells, as demonstrated in Fig. 2c. The primarily transmission coefficients of the network are plotted in Fig. 2f, confirming the symmetrical amplitude transmission in FP and BP. Additionally, the transmission coefficients’ amplitudes from port 11 to ports 4, 5, and 7 vary in a step of 6 dB. It originates from the required weights of connections between nodes for balanced and high-efficiency energy transmission among ports. Figure 2g shows the corresponding phase curves of the S-parameters, with their slopes determined by the number of activated BP connections. The bidirectional phase curves in NR cell-integrated channels exhibit a clear split, whereas those in R cell-integrated channels show complete overlap. The fabricated sample of the 8-input 8-output computing network is illustrated in Fig. 3a. As shown in Fig. 3b, the phase curves of S1,11 and S11,1 clearly separate, demonstrating the bidirectional path splitting capability of the sample. As a result, the FP and BP matrices are decoupled, enabling the splitting of bidirectional pathways on a single circuit.
a Image of the fabricated sample. b Phase spectrum of the S-parameters. The phase response in the forward path (S11,1) deviates from the backward one (S1,11) over a broad band. c PD of DCH. The PD can be flexibly tuned from −π to π across a wide frequency band, such as f1 (10 GHz), f2 (10.6 GHz), and f3 (12 GHz). d PD calculation for unidirectional transmission. The PDs for the three DCHs are calculated by subtracting the adjacent rows of the weight matrix. e, f Comparison of PDs in the BP path. The PDs of the three DCHs to ports 1–8 at f1 and f2. The simulated (red) and experimental (blue) data are consistent. g S-parameter’s phase of bidirectional propagation. The S-parameter’s phase from ports 9–16 to ports 1–8 at f2. Processing associated with the NRPS (ports 11, 13, and 15) exhibits a clear difference between forward (blue) and backward (red).
Next, we introduce the performance of the frequency division multiplexing. For the BP path, we define two adjacent ports (e.g., ports 10 and 11) as a differential channel (DCH). The phase difference (PD) from the DCH to the same port (e.g., port 1) varies with frequency, as shown in Fig. 3c. It arises from the differing transmission line lengths in the DCH and is influenced by the magneto-optical effect of the ferrite. This PD reflects the computational function of the DCH, such as addition (0°) or subtraction (180°) operators. Therefore, the function of the DCH changes in a wide frequency range. Three DCHs (ports 10 and 11, ports 12 and 13, and ports 14 and 15) are chosen to be discussed here. Figure 3e, f illustrates the PDs at two frequencies. At 10.6 GHz, the PDs of all three DCHs are ~180°, demonstrating their potential for differential operations. At 10 GHz, the PDs are ~70°, highlighting their potential for orthogonal and integral operations. Moreover, to show the total transmission coefficients between each port, all elements of the forward and backward transmission matrices are presented in Fig. 3g. Among these 64 data points, only the transmission coefficients (highlighted in yellow) related to ports 11, 13, and 15 (passing through the NRPS) differ between the bidirectional directions, reflecting the nonreciprocal effect of the NR cells.
Bidirectional decoupled image processing
To apply the decoupling capability between the FP and BP, we implement structure-based bidirectional decoupled image processing. As shown in Fig. 4, the two unidirectional propagations of the multi-port network have different computational functions. The FP generates a region-integrated image, mimicking global perception in biological neural systems, while the BP performs edge detection, simulating localized response. This confirms that the brain-inspired nonreciprocal neural network effectively realizes integrated perception and response, without the external directional modules.
a Image processing results of grayscale images with basic, Robert, and Prewitt operators. The network effectively enhances image contrast and extracts the edge features in the FP and the BP paths, respectively. b Image processing results of binary images. Only the BP path implements edge detection. c Frequency division multiplexing characteristics of the network. In the forward path, the network performs edge detection and image contrast enhanced at f1 and f2, respectively. The backward path is exactly the opposite.
Image processing in the multi-port network is performed by matrix-vector multiplication. We first encode and divide the two-dimensional input image into several 1 × 8 pixel vectors, denoted as Iini = [I1 I2 I3 I4 I5 I6 I7 I8], where I1–I8 is the pixel value of the image. The number of pixel vectors depends on the size of the original image and the encoding method. The 8 elements of a single pixel vector excite the 8 input ports of the network simultaneously in a single operation. Then the measured S-parameters are used to construct the 8 × 8 weight matrix W of the network. Thus, the signal of the output port is O = Iini × W = [O1 O2 O3 O4 O5 O6 O7 O8]. To maximize energy utilization, the output signals of 8 ports are summed as the result of image processing. This is similar to a virtual series connection of 8-in 1-out coupler at the output of the multi-port network, and the output of the virtual coupler is taken as the final output port. To show the process more clearly, the overall transmission matrix K is defined, which describes the transmission coefficient from each input port to the final output port. Each element of K is represented by \({{{{\rm{k}}}}}_{j}={{{{\rm{a}}}}}_{j}\times {e}^{i{\varphi }_{j}}\), which is numerically equal to the sum of the transmission coefficients from port 1 to ports 9–16. As a result, our final output result is R \(={\sum }_{j=1}^{8}{{{{\rm{O}}}}}_{j}={{{{\rm{I}}}}}_{{ini}}\times {{{\bf{K}}}}={\sum }_{j=1}^{8}{{{{\rm{I}}}}}_{j}{{{\boldsymbol{\times }}}}{{{{\rm{k}}}}}_{j}\). Here, j represents the sequential number of the input port. Note that the K varies with frequency within a wide band (see Supplementary Text 5 for details). Leveraging the interleaved cascading topology of the circuit and the phase modulation of NRPS, we can achieve: For the FP path, uniform \({{{{\rm{a}}}}}_{j}\) and \({\varphi }_{j}\) across all ports generate region-integrated images with enhanced contrast. For the BP path, phase modulation of NR cells (e.g., \({{\varphi }_{2}-\varphi }_{3}\approx 180^\circ\)) enables weighted first-order differentiation, producing an edge detection.
Based on the above principles, the image processing for grayscale and binary images is demonstrated through three different operators. For the grayscale image in Fig. 4a, the original image of size 1024 × 1024 in FP has low contrast, with the ratio of the brightest to the darkest pixels being 4.51. After the signal processing by the network, the pixel ratios of the three operators are 7.98, 7.94, and 9.27, respectively. This indicates that forward processing achieves enhancing image contrast, effectively simulating the global perception of the biological systems. For the image of size 256 × 256 in BP of Fig. 4a, the three gradient operators all effectively extract the edges of the original image, indicating backward processing accomplishes edge detection, imitating the backward localized specific response. Interestingly, the images of arbitrary resolution are decoupled and processed in both the forward and backward paths. For a binary image of size 876 × 876 in Fig. 4b, the forward allows image passthrough, while the backward performs edge detection. Moreover, operators can be flexibly adjusted by varying the encoding of the input data (for more details, see Methods). Interestingly, the implementation effects of three edge detection operators vary. Taking the grayscale image as an example, we define edge contrast as the ratio of pixel values between bright and dark areas of edges. The absolute edge contrast for the three operators is as follows: Prewitt (4.37), Robert (1.76), and basic gradient (1.41). It indicates the clarity of image edges: Prewitt operator performs best, followed by the Robert, while the last is basic gradient. This difference comes from the operator itself. Additionally, compared to the basic operator, the edges extracted by the Robert operator are thicker. The Prewitt operator demonstrates the most significant edge extraction effects in both horizontal and vertical directions. However, when the image edges approach ±45°, the Robert algorithm performs better. The difference is due to the operator itself.
Furthermore, to improve image generation quality, the image processing results here are the direct output of a linear digital layer. This linear layer is the weight matrix of the network, projecting the input image onto the final desired output in the digital platform. Compared to direct optical readout, the high-precision computations of digital computers enhance the performance of optical computing hardware. Therefore, by combining the analog and digital domains, the computational power is increased and the burden on digital hardware is alleviated. The development of similar strategies has already become quite mature46.
Bidirectional frequency division multiplexing
The network not only supports nonreciprocal propagation but also features frequency division multiplexing capabilities. Different computational functions can be achieved by varying the frequency. It is shown by the PDs in the DCH transmission coefficients in Fig. 3c. Specifically, at 10.2 GHz, the PD is 86.3°, allowing for orthogonal modulation; at 10.6 GHz, it is −185.5°, enabling differential operations; and at 12 GHz, it is −2.8°, facilitating integral operations. Figure 3d shows that the S-matrix of the network varies with frequency. Multiple asymmetric S-matrices are constructed over a broad frequency range, thus enabling diverse computational functions.
Figure 4c illustrates the bidirectional processing for the original image of size 256 × 256 at two frequencies. Both frequencies use the same encoding method based on the Prewitt gradient operator. At f1, the FP path performs edge detection, while the BP outputs a regional integral image; at f2, the FP outputs an integral image, and the BP conducts edge detection. And the edge extraction capabilities of f2 are different from f1. The edge contrast of the output image at the two frequencies is f1 (2.31) and f2 (3.14), indicating that edges extracted at f2 are clearer. The distinctive is because the PD at f1 is 120°, while at f2 is −185.5°. This makes the operation at f2 closer to the ideal gradient (180°). Overall, both frequencies accomplish integral and gradient operations, but they exhibit completely opposite functionalities in the same propagation direction. Thus, this structure achieves a wide range of nonreciprocal computational functions through frequency division multiplexing, effectively enhancing the versatility of the functions in the perception-response system.
Inverse matrix operation capability through a feedback network
The network can also perform matrix operations by incorporating recursive feedback waveguides, such as matrix solving. As shown in Fig. 5a, the system consists of a computational network, feedback structures, and matched loads. Three waveguides are used to sample input and output signals, while three feedback waveguides introduce recursion. Note that the feedback line lengths are precisely designed to be integer multiple of the wavelength to ensure phase-matching conditions for multi-port inputs. In this case, steady-state recursion is utilized to achieve the Neumann series, which is mathematically equivalent to matrix inversion. This demonstrates that the computing capabilities of the network are scalable with a high level of integration.
a The closed-loop network for inverse matrix computation. The network consists of three feedback waveguides that connect ports 3 and 11, 5 and 13, 7 and 15, respectively. The feedback waveguides produce recursive wave propagation within the network. Arrows indicate the direction of wave flow through the structure. b Inverse matrix computation results under two spin encoding conditions. The results (nine elements of the 3 × 3 matrix) are extracted from the output port, and the experimental (red) and theoretical (blue) data [real part (boxed) and imaginary part (unboxed)] demonstrate good consistency.
First, to positively design and analyze the feedback system model, the formula representing the transmission of the system is derived. The signal is carried in the form of a complex-valued electromagnetic field, propagating in the recursive path of the system. Both forward and backward recursive paths exist due to the presence of both internal and external connections. Therefore, the output signal is the superposition of the inner and outer loops. We consider typical conditions (eigenvalues of the target matrix being less than one), and the field in the system converges to a steady state. The solution is obtained in this steady state. For instance, for the BP path, the signal is input into the system as shown in Fig. 5a, and the output is:
Here, \({{{{\bf{I}}}}}_{{{{\rm{in}}}}}\) is any input signal, and SN, gF and gB represent ground noise, forward recursion, and backward recursion, respectively. Si represents 3 × 3-sized sub-matrices of the S-matrix of the original 8-input 8-output computational network, and the detailed information is shown in the Supplementary Text 6. The series summation corresponds to the matrix inversion, so Eq. (1.3) becomes:
Therefore, once the scattering matrix is known, the output signal is the inverse matrix modulation operation of any input signal. The theoretical inverse matrix is obtained through noise reduction and de-embedding operations. Note that the inverse matrix operation (\({{{{\bf{g}}}}}_{{{{\rm{B}}}}}\) or \({{{{\bf{g}}}}}_{{{{\rm{F}}}}}\)) represented by the inner loop (\({{{{\bf{S}}}}}_{1}\)) or the outer loop (\({{{{\bf{S}}}}}_{2}\)) can be regarded as the target, and the remaining part (SN + gF or SN + gB) only needs to be removed as noise. It effectively reflects the diverse computational capabilities of the feedback network, as two different inverse matrix operations can be calculated at once. And the steady-state solution for the FP path is similarly presented in the Supplementary Text 6.
Furthermore, unlike traditional feedback-based matrix solvers47, the solvable matrices can be tuned flexibly. As shown in Eq. (1.4), the matrix to be solved is related to the scattering matrix of the initial system. Thus, by modulating the transmission of the nonreciprocal network, the matrix changes accordingly. As discussed in the previous sections, this system features two decoupled unidirectional paths, and the paths are modulated by the ferrite bias magnetic field. Therefore, the matrix-solving capabilities of the feedback network are flexibly controlled by the propagation direction, operating frequency, and ferrite bias orientation. This demonstrates the reconfigurable computational functionality of the network.
To verify the feasibility of matrix inversion, simulations based on Fig. 5a are conducted in CST for the feedback network. To demonstrate the reconfigurable computational capabilities of the network, we encode the spin direction of the ferrite to distinguish different situations. 1 and −1 are used to represent the bias magnetic field direction of the ferrite, where 1 is the y-direction bias and −1 is the −y-direction bias. When the spin codes are set to [1 1 1] and [−1 1 1], the simulations are both basically consistent with the theoretical results, as shown in Fig. 5b. Here, the theoretical solution is based on Eq. (1.4). The specific structure diagram and field map distribution are in the Supplementary Text 7. The Fig. 5b not only verifies the accuracy of the formula, but also verifies the multi-matrix solution capability of the feedback system by the spin code of ferrites.
Discussion
Our work is inspired by biological neural systems (Fig. S1), where the unidirectional transmission of synapses establishes distinct FP and BP. This enables perception and response to be decoupled by direction within a single network. For example, in the visual system, external stimuli are first processed globally to construct a holistic scene understanding. Subsequently, specific backward pathways are selectively activated to identify salient features and trigger targeted responses. This integrated perception-response ensures the efficiency and stability of neural function. Such nonreciprocal, brain-like behavior (achieving the decoupling between FP and BP) offers important insights for developing the next generation of high-efficiency computing networks.
Most current optical neural networks are reciprocal. While multiple reciprocal networks can each perform distinct computing functions, they fundamentally require external selection modules to activate the target functionality, as shown in Fig. S1. These function-selection modules inherently disrupt system reciprocity. The external modules introduce additional resource overhead and design complexity. Moreover, deploying multiple networks and ensuring their efficient interconnection becomes increasingly challenging as system scale grows. Therefore, such architectures severely limit system integration density and energy efficiency. In contrast, nonreciprocal neural networks offer a compelling solution through inherent architectural reuse and directional path independence. Nonreciprocity embeds the function-selection module within the network itself via propagation direction, eliminating the need for external modules. As a result, perception and response can be integrated within the same structure and decoupled directionally. This avoids redundant deployment and simplifies system-level design. Such nonreciprocal optical neural networks lay the foundation for future high-density, multifunctional, and co-optimized intelligent computing systems.
We have proposed and experimentally demonstrated a nonreciprocal (bidirectionally decoupled) multi-port network that emulates the integrated perception (FP) and response (BP) of the brain. By tuning the phase of nonreciprocal cells, the network allows independent control of weight matrices and computing functions along FP and BP. Our results confirm that such FP-BP separation can be achieved over a broad frequency range via magneto-optical effects. At specific frequencies and magnetic bias, FP performs area integration to mimic global perception, while BP executes edge detection to emulate selective localized response. By decoupling FP and BP within a single structure, the design eliminates the need for external function-selection modules, improving both resource efficiency and system integrability. Additionally, recursive feedback waveguides enable matrix inversion operations. Furthermore, since the FP and BP are decoupled, the number of independently tunable weight coefficients in the network is significantly increased. This effectively enhances the computational capacity of a single network while improving spatial utilization and communication throughput. In future designs, fully nonreciprocal interconnections can be implemented to further expand the DOF for weight control (in Fig. S3), equivalent to a doubling of computing power. Notably, the increase in DOF scales nonlinearly with the number of ports, leading to dramatic performance gains in large-scale, high-capacity networks. This makes the proposed approach highly suitable for high-density and high-capacity computing systems. Our nonreciprocal design methodology is also applicable to other optical frequency bands and platforms, including integrated photonic waveguides, programmable photonic circuits, and terahertz communication systems. Future work may further increase the number of units per chip and expand vertical stacking to support simultaneous multi-task processing. As integration density and stacking technologies advance, this platform is expected to support massively parallel, multimodal, and feedback-rich intelligent systems. It unlocks new possibilities for efficient, large-scale physical AI in the post-Moore era.
Methods
Input signal encoding method for loading operators
Operators can be flexibly adjusted by varying the encoding of the input data. We utilize three edge detection operators, each of them consists of two filtering operators: the basic gradient (dx = [1 −1; 0 0]; dy = [1 0; −1 0]); Robert (dx = [1 0; 0 −1]; dy = [0 −1; 1 0]); Prewitt (dx = [1 1 1; 0 0 0; −1 −1 −1]; dy = [1 0 −1; 1 0 −1; 1 0 -1]). The physical implementation process of the operators in the BP path is described here. Assume that the input detected image matrix is set as I = [A B C; G H I; M N O]. Note that we only discuss a single calculation here. The input signal is calculated as follows: when the input pixel vector is Iin1 = [−G A B A G A B A], the final output is R1 = 2a × φ1 × (A−B + A−G). It simultaneously achieves both horizontal and vertical local gradient operations, providing the basic gradient operator. When the input is Iin2 = [−G A H B G A H B], the output is R2 = 2a × φ1 × (A−H + B − G), achieving the cross-gradient operation of the Robert operator. When the input vector is Iin3 = [0 A M B N C O 0], the output is R3 = a×φ1 × (A + B + C−M−N−O), corresponding to the horizontal gradient operator of Prewitt. The vertical gradient operator is implemented by rotating the input image 90° (transposing the input pixel matrix), yielding an output of R4 = a × φ1 × (A + G + M−B−H−N). The final output of the Prewitt operator is R5 = R3 + R4, satisfying the requirement of the Prewitt operator.
Next, the network works as a basic platform to implement different operators. Considering an input image of size m × n, the encoding process of the loading operator in image processing is analyzed here. The “encoding” here refers to a process of reducing the dimensionality, mapping 2D pixel data to the 8 input ports of the network. Leveraging an 8-input 8-output single-chip, full-dimensional image traversal is achieved through multiple computational steps. For simplicity, the analysis assumes an ideal gradient computation scenario for the image \({{{{\bf{I}}}}}_{{{\mathrm{input}}}}={\left[\begin{array}{cc}\begin{array}{ccc}{A} & {B} & {C}\\ {G} & {H} & {I}\\ {M} & {N} & {O}\end{array} & \begin{array}{ccc}{D} & {E} & {F}\\ {J} & {K} & {L}\\ {P} & {Q} & {R}\end{array}\end{array}\right]}_{3\times 6}\), where letters represent numerical values.
The computation process on the single-chip involves the following steps: the m × n input image is first divided into several 2 × 2 pixel blocks for the need of basic gradient operators, such as \({{{{\bf{I}}}}}_{{{\mathrm{pix}}}1}={\left[\begin{array}{cc}{A} & {B}\\ {G} & {H}\end{array}\right]}_{2\times 2}\), \({{{{\bf{I}}}}}_{{{\mathrm{pix}}}2}={\left[\begin{array}{cc}{B} & {C}\\ {H} & {I}\end{array}\right]}_{2\times 2}\) and so on. Each 2 × 2 pixel block Ipixj is then encoded (dimension-transformed), based on the logic of the loading operator, into a 1 × 8 vector Iinj. For instance, Ipix1 is encoded as Iin1 = [−G A B A G A B A]. An input signal of the same frequency carrying the eight elements of this vector is used to excite the eight ports of the network simultaneously. The output signals from these eight ports are collected and summed to produce the final result Ri, representing the gradient value of the corresponding pixel block Ipixj. By repeating this process for all 2 × 2 pixel blocks and arranging the results Ri in their respective positions, the processed image is reconstructed. This approach enables full-dimensional traversal of the image. Furthermore, in theory, full-dimensional image computation can be achieved by either vertically stacking multiple boards or increasing the number of ports on a single board.
This physical process above can be mathematically described as matrix-vector multiplication. As previously discussed, we utilize the direct output of the linear digital layer for post-processing based on structural measurement results, thereby reducing the computational burden on digital hardware. In this setup, the signals carrying the input image, fed into multiple ports simultaneously, can be treated as a one-dimensional vector Iinj. The process of repeated inputs can be simplified by expanding the 1 × 8 vector Iinj into an m × 8n matrix Ibasic. In other words, the physical excitation processes in the network are mathematically represented as the multiplication of a multi-dimensional matrix, directly related to the image, and a weight matrix. Similarly, the repeated propagation (or computation) processes in the network are equivalent to the operations of a high-dimensional 8m × n weight matrix. The details of this process are in the Supplementary Text 8.
Data availability
The minimum dataset supporting the findings and the released code has been deposited in (https://doi.org/10.5281/zenodo.15875691). Additional data are available under restricted access due to privacy regulations. Access requests can be directed to the corresponding author, with an expected response time of up to 60 days. Requesters should sign a data use agreement and submit a research proposal specifying the intended use. Access approval is contingent upon an ethical review to ensure compliance with institutional policies and legal requirements. Approved data use will be restricted to noncommercial research purposes. Source data are provided in this paper.
References
Jordan, M. I. & Mitchell, T. M. Machine learning: trends, perspectives, and prospects. Science 349, 255–260 (2015).
Dai, L. et al. A deep learning system for predicting time to progression of diabetic retinopathy. Nat. Med. 30, 584–594 (2024).
Chen, H. et al. Vanillanet: the power of minimalism in deep learning. Adv. Neural Inf. Process. Syst. 36, 7050–7064 (2024).
Yang, H. et al. Off-resonance spin-locked metasurfaces empowered by quasi-bound states in the continuum for optical analog computing. Adv. Funct. Mater. 33, 2305110 (2023).
Gill, S. S. et al. Modern computing: vision and challenges. Telemat. Inform. Rep. 13, 100116 (2024).
Gao, X. et al. Terahertz spoof plasmonic neural network for diffractive information recognition and processing. Nat. Commun. 15, 6686 (2024).
Xue, Z. et al. Fully forward mode training for optical neural networks. Nature 632, 280–286 (2024).
Lin, X. et al. All-optical machine learning using diffractive deep neural networks. Science 361, 1004–1008 (2018).
Fan, Z. et al. Holographic multiplexing metasurface with twisted diffractive neural network. Nat. Commun. 15, 9416 (2024).
Lin, P. et al. Assembling reconfigurable intelligent metasurfaces with synthetic neural network. IEEE Trans. Antennas Propag. 72, 5252–5260 (2024).
Hamerly, R. et al. Large-scale optical neural networks based on photoelectric multiplication. Phys. Rev. X 9, 021032 (2019).
Fang, M. Y.-S. et al. Design of optical neural networks with component imprecisions. Opt. Express 27, 14009–14029 (2019).
Wang, Z. et al. Single-layer spatial analog meta-processor for imaging processing. Nat. Commun. 13, 2188 (2022).
Bente, I. et al. The potential of multidimensional photonic computing. Nat. Rev. Phys. 7, 439–450 (2025).
Lugnan, A. et al. Emergent self-adaptation in an integrated photonic neural network for backpropagation-free learning. Adv. Sci. 12, 2300011 (2025).
Zhang, H. et al. Microwave speech recognizer empowered by a programmable metasurface. Adv. Sci. 11, 2309826 (2024).
Yang, Z. et al. A vision chip with complementary pathways for open-world sensing. Nature 629, 1027–1033 (2024).
Nikkhah, V. et al. Inverse-designed low-index-contrast structures on a silicon photonics platform for vector–matrix multiplication. Nat. Photonics 18, 501–508 (2024).
Meng, X. et al. Compact optical convolution processing unit based on multimode interference. Nat. Commun. 14, 3000 (2023).
Zhang, H. et al. An optical neural chip for implementing complex-valued neural network. Nat. Commun. 12, 457 (2021).
Ashtiani, F., Geers, A. J. & Aflatouni, F. An on-chip photonic deep neural network for image classification. Nature 606, 501–506 (2022).
Shen, Y. et al. Deep learning with coherent nanophotonic circuits. Nat. Photonics 11, 441–446 (2017).
Clements, W. R. et al. Optimal design for universal multiport interferometers. Optica 3, 1460–1465 (2016).
Gao, X. et al. Programmable surface plasmonic neural networks for microwave detection and processing. Nat. Electron. 6, 319–328 (2023).
Südhof, T. C. The cell biology of synapse formation. J. Cell Biol. 220, e202103052 (2021).
Südhof, T. C. Towards an understanding of synapse formation. Neuron 100, 276–293 (2018).
Yamamoto, H. et al. Unidirectional signal propagation in primary neurons micropatterned at a single-cell resolution. Appl. Phys. Lett. 109, 123701 (2016).
Dejanovic, B., Sheng, M. & Hanson, J. E. Targeting synapse function and loss for treatment of neurodegenerative diseases. Nat. Rev. Drug Discov. 23, 23–42 (2024).
Rabinowitch, I., Colón-Ramos, D. A. & Krieg, M. Understanding neural circuit function through synaptic engineering. Nat. Rev. Neurosci. 25, 131–139 (2024).
Yuste, R., Cossart, R. & Yaksi, E. Neuronal ensembles: building blocks of neural circuits. Neuron 112, 875–892 (2024).
Schuster, M. & Paliwal, K. K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45, 2673–2682 (1997).
Barjasteh, A., Ghafouri, S. H. & Hashemi, M. A hybrid model based on discrete wavelet transform (DWT) and bidirectional recurrent neural networks for wind speed prediction. Eng. Appl. Artif. Intell. 127, 107340 (2024).
Laghari, A. A. et al. Deep residual-dense network based on bidirectional recurrent neural network for atrial fibrillation detection. Sci. Rep. 13, 15109 (2023).
Alemu, A. A., Melese, M. D. & Salau, A. O. Towards audio-based identification of Ethio-Semitic languages using recurrent neural network. Sci. Rep. 13, 19346 (2023).
Sounas, D. L., Kodera, T. & Caloz, C. Electromagnetic modeling of a magnetless nonreciprocal gyrotropic metasurface. IEEE Trans. Antennas Propag. 61, 221–231 (2013).
Pintus, P. et al. Integrated nonreciprocal magneto-optics with ultra-high endurance for photonic in-memory computing. Nat. Photonics 19, 54–62 (2024).
Jin, B. Y. & Argyropoulos, C. Self-induced passive nonreciprocal transmission by nonlinear bifacial dielectric metasurfaces. Phys. Rev. Appl. 13, 054056 (2020).
Wang, X. et al. Amplification and manipulation of nonlinear electromagnetic waves and enhanced nonreciprocity using transmissive space-time-coding metasurface. Adv. Sci. 9, 2105960 (2022).
Gao, X. et al. Nonmagnetic spoof plasmonic isolator based on parametric amplification. Laser Photonics Rev. 16, 2100578 (2022).
Yan, H. et al. Broadband nonreciprocal spoof plasmonic phase shifter based on transverse Faraday effects. Opt. Express 30, 24000–24008 (2022).
Shen, Z. & Dong, C. H. Femtosecond laser writes a broadband miniaturized isolator. Nat. Photonics 19, 224–225 (2025).
Hu, Y. et al. Giant elastic-wave asymmetry in a linear passive circulator. Nat. Commun. 16, 3991 (2025).
Khurgin, J., Bykov, A. Y. & Zayats, A. V. Hot-electron dynamics in plasmonic nanostructures: fundamentals, applications and overlooked aspects. eLight 4, 15 (2024).
Li, L. et al. Intelligent metasurfaces: control, communication and computing. eLight 2, 7 (2022).
Yao, X. et al. Kirigami-triggered spoof plasmonic interconnects for radiofrequency elastronics. Research 7, 0367 (2024).
Zhan, Z. et al. Photonic diffractive generators through sampling noises from scattering media. Nat. Commun. 15, 10643 (2024).
Mohammadi Estakhri, N., Edwards, B. & Engheta, N. Inverse-designed metastructures that solve equations. Science 363, 1333–1338 (2019).
Acknowledgements
The work at Zhejiang University was sponsored by the National Natural Science Foundation of China (NNSFC) (62222115), the Key Research and Development Program of Zhejiang Province under Grant No.2024C01241(SD2).
Author information
Authors and Affiliations
Contributions
Z. Wang and H. Chen initiated the plan and supervised the entire study. Z. Wang and X. Li conceived the idea of this work and designed the simulations and experiments. X. Li, H. Yang, E. Wu, X. Yao, F. Gao, and Y. Li carried out the measurements and data analyses. X. Li and Z. Wang prepared the manuscript with input from all authors. All authors discussed the research.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Communications thanks Cheng-Wei Qiu and the other, anonymous, reviewer for their contribution to the peer review of this work. A peer review file is available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Li, X., Yang, H., Wu, E. et al. Nonreciprocal surface plasmonic neural network for decoupled bidirectional analogue computing. Nat Commun 16, 7703 (2025). https://doi.org/10.1038/s41467-025-63103-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-025-63103-z