Abstract
A silicon photonic deep optical neural network integrating convolutional and fully connected layers with on-chip optoelectronic nonlinear activations operates with partially coherent light to achieve high-speed, energy-efficient, end-to-end inference. This demonstration establishes a functional and scalable platform for evaluating complete optical neural processing, representing another step toward specialised, ultrafast photonic architectures beyond electronics.
Artificial Neural Networks (ANNs) are trainable algorithms inspired by the structure and functioning of the brain, designed for solving classification or regression tasks by training on extended datasets. These algorithms lie at the core of today’s progress in the AI field, being an essential part of large language models and transformer architecture as one of the most prominent examples. Although introduced in the 1940s by McCulloch and Pitts1, only in the last three decades has it gained increased attention because of the error back-propagation algorithm proposed by Rumelhart, Hinton, and Williams2. This enabled the training of multi-layer networks and effectively led to the demonstration of today’s large and deep models with billions of parameters. But this technological step could not have been done without the parallel development of digital electronic devices capable of more and more efficient computations, including CPUs, GPUs, and, more recently, processing units solely devoted to the acceleration of ANN computations. The most recent progress also sparked interest in other fields, motivated by the collapse of electronic scaling “laws”, as further development of AI inevitably relies on and demands high computational efficiency3.
Particularly, this led to the resurrection of the analog computing paradigm - abandoned in the past because of the unmatched universality of digital computing - based on the assumption that analog computing enables asynchronous, highly parallel processing and in-memory computing, hardware-emulating the inherent properties of ANN algorithms. A similar belief is at the root of the reborn of the optical analog computing field, although with the additional motivation of capitalizing on the unmatched high throughput of optical links inherited from telecom technology4.
This comeback of optical computing over the past decade was also strongly motivated by the maturation of silicon integrated photonics technology, which coincided with the rapid rise of deep learning algorithms relying heavily on matrix multiplication operations repeated across many layers, often billions of times during training. Because matrix multiplication dominates both computation time and energy use, optimizing this operation became central to accelerating AI computations (as shown in a simplified scheme, Fig. 1a). As a result, optical approaches have been viewed as a promising route to handle the large-scale linear algebra, seeing rapid development from proof of principle demonstration in 2017 by Shen et al. towards recently demonstrated complete hybrid accelerators with photonic tensor cores5,6,7. This successful route stemmed from the hybrid approach, where photonics were used for the acceleration of specific operations, while a configurable electronic interface provided universality and usefulness of the device.
a Acceleration of linear algebra operations using a photonic tensor core, where network layer outputs are sequentially computed and the CPU offloads matrix operations to the photonic hardware. b An interfaced end-to-end optical neural network that accelerates inference tasks in sensory applications, enabling high-speed, real-time processing. c Simplified schematic of an end-to-end optical neural network architecture from ref. 24, comprising matrix multiplication layers, optically driven nonlinear activators (NAF), and operating with partially coherent light
Meanwhile, the number of publications describing theoretically or demonstrating functional components and architectures of end-to-end Optical Neural Networks (ONN) has rapidly increased8. As for now, the burden of analog technology still prevents demonstrations of optical deep networks, which can compete with the scalability of electronics, with the main reason being footprint constraints and cascadability issues9,10. Although large-scale deep networks are still ahead, the high-speed processing capabilities of such hardware can be harnessed for practical solutions based on shallower networks in specific domains, especially those requiring real-time sensing and processing at GHz rates11. In such, as schematically shown in Fig. 1b, optical solutions have the potential to outperform electronics already now, taking advantage of non-von-Neumann architectures12 and lightspeed transfers13,14,15. However, shifting from basic research towards application-specific designs requires demonstrations of complete end-to-end networks that can be treated as a test-bed for verification of future use. Here, the hindering is related to the lack of configurable energy-efficient components, realizing rapidly switching non-linear activation function (NAF) of optical neurons14,16,17 - one of the critical neural network functionalities – to be combined with already demonstrated optical tensor core technologies18. Another challenges are related to material incompatibility, negative net gain, or lack of configurability in the case of optical NAFs, or relatively high complexity and lagging in optoelectronic NAFs, with both solutions often decreasing energy efficiency.
Despite these difficulties, a number of end-to-end photonic networks have already been demonstrated, ranging from free-space planar architectures to on-chip integrated systems employing various types of NAFs19,20,21,22,23. Building on these advances, a recent paper published in Light: Science & Applications takes a significant step forward by addressing key issues in achieving improved end-to-end on-chip photonic demonstrations.
Wu et al. report the demonstration of an end-to-end partially coherent deep optical neural network (PDONN) implemented on a silicon photonic chip, as shown in a simplified scheme in Fig. 1c24. Their architecture directly addresses two challenges in optical neural networks: limited network depth, due to insufficiently cascadable nonlinear activation functions, and restricted input dimensions, imposed by the requirements of coherent light sources and detection schemes. The PDONN introduces a monolithically integrated opto-electro-opto (O-E-O) nonlinear activation function based on micro-ring modulators driven by differential photocurrents. This configuration achieves positive net optical gain, enabling multiple nonlinear layers to be cascaded on a single chip without the need for external electrical amplifiers—a crucial step for realizing deeper photonic neural architectures. The network also integrates two convolutional layers and two fully connected layers in a compact 17 mm² silicon photonic chip with a 64-channel optical input, reporting the largest demonstrated input size and layer count for an end-to-end on-chip photonic neural network to date.
By operating with partially coherent optical sources such as amplified spontaneous emission (ASE) or LED-based emitters, the system avoids reliance on narrow-linewidth lasers and complex coherent detection, thereby enhancing manufacturability and robustness while maintaining high classification accuracy. The PDONN achieved 96% accuracy for binary fashion image classification and 94% for four-class handwritten digit recognition, with performance close to that obtained under fully coherent illumination. The design further benefits from real-valued optical computation, reducing the hardware complexity associated with phase control and improving tolerance to fabrication variations. The total latency per inference is approximately 4.1 ns, dominated by the RC delay of the O-E-O nonlinearities, and the estimated energy efficiency is ~122 pJ per operation. Together, these results demonstrate an optical computing platform that combines high speed, energy efficiency, and architectural scalability.
The work by Wu et al. exemplifies a broader shift in optical neural network research — from demonstrating individual components to realizing fully integrated computational pipelines. Their system showcases a self-contained optical neural processor capable of performing end-to-end inference at high speed, demonstrating both completeness and functional viability. Although such systems are still far from matching electronic accelerators in terms of scale or configurability, their architectural integrity makes them valuable platforms for testing and refining optical computing concepts under realistic conditions25,26. Continued co-optimization and experimental validation could pave the way for high-speed, application-specific systems, highlighting the potential of optical approaches to outperform electronic hardware in tasks requiring ultrafast inference.
References
McCulloch, W. S. & Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 115–133 (1943).
Rumelhart, D. E., Hinton, G. E. & Williams, R. J. Learning representations by back-propagating errors. Nature 323, 533–536 (1986).
Indiveri, G. Introducing `neuromorphic computing and engineering’. Neuromorp. Comput. Eng. 1, 010401 (2021).
Stroev, N. & Berloff, N. G. Analog photonics computing for information processing, inference, and optimization. Adv. Quantum Technol. 6, 2300055 (2023).
Rizzo, A. Photonic chips provide a processing boost for AI. Nature 640, 323–325 (2025).
Hua, S. Y. et al. An integrated large-scale photonic accelerator with ultralow latency. Nature 640, 361–367 (2025).
Ahmed, S. R. et al. Universal photonic artificial intelligence acceleration. Nature 640, 368–374 (2025).
Fu, T. Z. et al. Optical neural networks: progress and challenges. Light Sci. Appl. 13, 263 (2024).
Miller, D. A. B. Why optics needs thickness. Science 379, 41–45 (2023).
Miller, D. A. B. Are optical transistors the logical next step? Nat. Photonics 4, 3–5 (2010).
Wetzstein, G. et al. Inference in artificial intelligence with deep optics and photonics. Nature 588, 39–47 (2020).
Feldmann, J. et al. Parallel convolutional processing using an integrated photonic tensor core. Nature 589, 52–58 (2021).
Wang, J. et al. Terabit free-space data transmission employing orbital angular momentum multiplexing. Nat. Photonics 6, 488–496 (2012).
Matuszewski, M., Prystupiuk, A. & Opala, A. Role of all-optical neural networks. Phys. Rev. Appl. 21, 014028 (2024).
Bernstein, L. et al. Single-shot optical neural network. Sci. Adv. 9, eadg7904 (2023).
Tyszka, K., Opala, A. & Piętka, B. Advancing optical spiking neural networks with exciton-polaritons. J. Phys.: Photonics 7, 041002 (2025).
Xu, Z. F. et al. Reconfigurable nonlinear photonic activation function for photonic neural network based on non-volatile opto-resistive RAM switch. Light Sci. Appl. 11, 288 (2022).
Zhou, H. L. et al. Photonic matrix multiplication lights up photonic accelerator and beyond. Light Sci. Appl. 11, 30 (2022).
Ashtiani, F., Geers, A. J. & Aflatouni, F. An on-chip photonic deep neural network for image classification. Nature 606, 501–506 (2022).
Bandyopadhyay, S. et al. Single-chip photonic deep neural network with forward-only training. Nat. Photonics 18, 1335–1343 (2024).
Skalli, A. et al. Photonic neuromorphic computing using vertical cavity semiconductor lasers. Opt. Mater. Express 12, 2395–2414 (2022).
Chen, Z. J. et al. Deep learning with coherent VCSEL neural networks. Nat. Photonics 17, 723–730 (2023).
Zuo, Y. et al. All-optical neural network with nonlinear activation functions. Optica 6, 1132–1137 (2019).
Wu, B. et al. Scaling up for end-to-end on-chip photonic neural network inference. Light Sci. Appl. 14, 328 (2025).
Huang, C. R. et al. A silicon photonic–electronic neural network for fibre nonlinearity compensation. Nat. Electron. 4, 837–844 (2021).
Zhang, W. P. et al. A system-on-chip microwave photonic processor solves dynamic RF interference in real time with picosecond latency. Light Sci. Appl. 13, 14 (2024).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The author declares no competing interests.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Tyszka, K. Demonstrating completeness in optical neural computing. Light Sci Appl 15, 39 (2026). https://doi.org/10.1038/s41377-025-02123-2
Published:
Version of record:
DOI: https://doi.org/10.1038/s41377-025-02123-2
