• Photonics Research
  • Vol. 13, Issue 4, 915 (2025)
Zheng Huang1,2, Conghe Wang1,2, Caihua Zhang1,2, Wanxin Shi3..., Shukai Wu1,2, Sigang Yang1,2 and Hongwei Chen1,2,*|Show fewer author(s)
Author Affiliations
  • 1Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
  • 2Beijing National Research Center for Information Science and Technology, Beijing 100084, China
  • 3China Mobile Research Institute, Beijing 100053, China
  • show less
    DOI: 10.1364/PRJ.539630 Cite this Article Set citation alerts
    Zheng Huang, Conghe Wang, Caihua Zhang, Wanxin Shi, Shukai Wu, Sigang Yang, Hongwei Chen, "Brain-like training of a pre-sensor optical neural network with a backpropagation-free algorithm," Photonics Res. 13, 915 (2025) Copy Citation Text show less

    Abstract

    Deep learning has rapidly advanced amidst the proliferation of large models, leading to challenges in computational resources and power consumption. Optical neural networks (ONNs) offer a solution by shifting computation to optics, thereby leveraging the benefits of low power consumption, low latency, and high parallelism. The current training paradigm for ONNs primarily relies on backpropagation (BP). However, the reliance is incompatible with potential unknown processes within the system, which necessitates detailed knowledge and precise mathematical modeling of the optical process. In this paper, we present a pre-sensor multilayer ONN with nonlinear activation, utilizing a forward-forward algorithm to directly train both optical and digital parameters, which replaces the traditional backward pass with an additional forward pass. Our proposed nonlinear optical system demonstrates significant improvements in image classification accuracy, achieving a maximum enhancement of 9.0%. It also validates the efficacy of training parameters in the presence of unknown nonlinear components in the optical system. The proposed training method addresses the limitations of BP, paving the way for applications with a broader range of physical transformations in ONNs.

    1. INTRODUCTION

    Deep learning models, like convolutional neural networks (CNNs) [1], recurrent neural networks (RNNs) [2], and transformers [3], have emerged as pervasive instruments within the realms of science and engineering, facilitating the accomplishment of diverse and intricate tasks. Nevertheless, traditional electronic computing hardware grapples with challenges like the breakdown of Moore’s Law and heightened power consumption, presenting obstacles in meeting the escalating demand for computational power. Deep physical neural networks capitalize on specific physical carriers, such as light, as computing media. This paradigm relocates the mathematical computation process, originally confined to the digital domain, to physical hardware, showcasing notable advantages in speed, power efficiency, and throughput.

    Optical neural networks (ONNs) [46] are being rigorously investigated for their potential to accelerate computations in deep learning efficiently, attributed to the characteristics of high speed, broad bandwidth, and low power consumption. Various ONN implementations modulate the amplitude, phase, and polarization of light by optical elements such as metasurfaces and spatial light modulators (SLMs) [718], executing linear computations in free space. Nevertheless, the advent of next-generation applications such as autonomous driving and augmented reality (AR) necessitates a high response speed and superior energy efficiency to facilitate real-time decision-making in edge devices. To this end, pre-sensor optical computing has been proposed to offload substantial computational tasks from electronics into optics [19], thereby harnessing the inherent advantages of optical systems, where three keys must be met to make it feasible. Firstly, the pre-sensor processor must be capable of handling incoherent light illumination in real-world scenarios, processing the incoming information, and performing specific computations during light propagation. This processor should be integrated seamlessly with the mature machine vision pipeline, effectively replacing traditional imaging procedures. Secondly, the imperative for nonlinear activation between layers in ONNs remains, and optical-to-optical nonlinear activation could extend the layer depth and computational capabilities of ONNs. Thirdly, the design of optical elements, such as the patterns on an SLM or the specifications of silicon cylinders in the metasurface, must be optimized to ensure best performance [20,21].

    To date, the predominant methodology for training ONNs has been the backpropagation (BP) algorithm [22,23], which has led to the emergence of three distinct paradigms: in-silico training, in-situ training, and combined training, as depicted in Fig. 1(b). The in-silico training approach involves creating a mathematical dual to the physical system. Model parameters are optimized through gradient descent, derived using the chain rule, thus enabling updates to the corresponding parameters of the optical system. A precise and differentiable mathematical representation is indispensable for the backward pass within the complete end-to-end training framework. In-situ training [2426] entails direct physical computation and updates of the physical structure parameters on the device itself, leveraging the product approximation of the forward propagation and backpropagation results as a partial derivative of the applied loss function with respect to the weight values. This approach mitigates specific errors, such as those arising from fabrication discrepancies and the divergence between the mathematical digital model and the physical model. Nonetheless, the in-situ method entails performing both training and inference on the devices, which escalates hardware costs. The combined training [2734] approach aims to balance training accuracy against hardware expenditures. For example, the physics-aware training method is proposed as a hybrid in-situ-in-silico method [27]. Here, the forward pass is conducted physically, while the backward pass is executed in-silico using a digital model that simulates the input-output transformations. This model is refined through extensive data training. Such dual training methodologies are also employed to design hardware parameters or rectify hardware-induced errors.

    Paradigms for training ONNs: in-silico training, in-situ training, combined training.

    Figure 1.Paradigms for training ONNs: in-silico training, in-situ training, combined training.

    However, the training algorithm based on BP possesses inherent limitations, especially the requirement that the physical process or its digital counterpart in forward propagation be differentiable. If a black box, whose operational principles and mathematical representation are undisclosed, is integrated into the forward propagation pathway, the BP-based training is predisposed to failure. Concurrently, to more closely mimic the training mechanisms of the human brain, ONNs can adapt to processing and learning from transmission data across different stages in real-time rather than relying solely on executing the BP process after data propagation. The forward-forward algorithm (FFA) is a local learning method that defines a local loss function within a segment of a model. Similar to reservoir computing, where the hidden layer is randomly initialized and remains fixed, the output layer serves as the sole trainable component, optimized through localized adjustments [3540]. FFA is utilized in training physical neural networks; however, its application is primarily targeted towards digital linear layers rather than the physical parameters themselves, limiting the broader applicability of the FFA in diverse training contexts where adjustments to physical parameters are necessary.

    In this paper, we demonstrate a pre-sensor optical computing system employing FFA to train both optical and digital parameters directly, analogous to the learning processes in the human brain. The optical component comprises a passive amplitude mask, an image intensifier, and an SLM, which facilitate nonlinear activation and linear modulation. Utilizing the training method of FFA, it becomes possible to train the optical linear segments and the corresponding digital linear components independently of understanding the nonlinear modulation processes executed by the image intensifier, enhancing the overall efficiency and adaptability of the system. The experimental results reveal an accuracy of 93% in the classification of the MNIST dataset using a single-layer ONN and an accuracy of 87.5% in the classification of the Fashion-MNIST dataset using a double-layer ONN. Broadly, the training method offers a promising approach to automatically design parameters for functional optical devices, particularly when certain components of the system cannot be explicitly modeled, empowering scenarios for pre-sensor optical computing with high speed, high efficiency, and low power consumption.

    2. FORWARD-FORWARD ALGORITHM

    The FFA was initially devised to address the lack of precise details in forward computation [41]. In this method, the conventional backward pass of backpropagation is replaced with a layer-specific local loss function, referred to as the “goodness function.” By presenting both positive and negative samples to each layer of the network, the objective of the goodness function is to ensure that the goodness for positive samples significantly exceeds that for negative samples, thereby classifying the input data as positive or negative. During the positive pass, data with the correct label are processed through forward propagation, with the weights adjusted to increase the goodness of the trainable layer. Conversely, during the negative pass, data with incorrect labels undergo the same forward propagation. Specifically, the parameters are optimized for each trainable layer by minimizing the following loss function: L=1Ni=1Nlog(1+exp(ypos,i+τ))+log(1+exp(yneg,iτ)).

    The local loss function is designed to maximize the margin between positive and negative samples in the feature space, effectively promoting numerical stability for robust optimization. ypos,i is the output for the positive sample. yneg,i is the output for the negative sample. τ represents a manually set threshold that controls the separation degree between positive and negative samples. In supervised learning tasks, such as M-class classification, any given sample can generate one positive sample and M–1 negative samples, thus resulting in M–1 positive and negative data pairs. Additionally, layer normalization is applied to remove value information, preserving only orientation information to facilitate feature learning in subsequent layers. The local loss function is computed following a series of operations, including nonlinear activation and squaring. The FFA-driven training approach performs local optimization within each trainable layer using a customizable local loss function, updating parameters based on gradient descent or other optimization techniques, thereby eliminating the need for inter-layer backpropagation through the chain rule. The schematic for the FFA is illustrated in Fig. 1(b). During the inference stage, for each sample to be inferred, M samples with different labels are generated and passed through the entire network separately. The label corresponding to the maximum output is taken as the predicted label. FFA can learn by continuously pipelining sequential data through a neural network without the need to store neural activities and access to parameters of each layer, enabling the use of low-power analog hardware. In local training, each layer independently endeavors to extract its own distinctive features. It is also effective for unsupervised and contrastive supervised learning paradigms.

    3. NONLINEAR OPTICAL ACTIVATION FUNCTION OF IMAGE INTENSIFIER

    Similar to the transmission rules of human neurons, which transmit signals only under specific stimuli, the activation function in neural networks acts as a control valve that regulates neuron activity. Nonlinear activation is crucial for neural networks to represent more complex functions. Without nonlinear activation functions between layers, all hidden layers lose their original purpose, as multiple layers of linear operations are equivalent to a single layer of linear operations. For ONNs, there are numerous methods to construct linear operations that lack nonlinear activation. Researchers have been exploring various approaches to achieve all-optical interlayer nonlinearity, such as using phase change materials [42], saturable absorbers [43,44], graphene materials [45], and quantum dots [19,46,47]. Two-dimensional (2D) spatial nonlinearity is particularly suitable for integration between two optical layers, providing stable and consistent nonlinearity. Recently, Wang et al. proposed a multilayer neural network using a commercial image intensifier to compress information [18]. The image intensifier converts incident photons into electrons, where the microchannel plate (MCP) is applied voltage, causing internal collisions and multiplication of electrons to form an avalanche reaction for signal amplification. The intensifier’s fluorescent screen then converts the electrons back into light signals. The local saturation of the MCP in the image intensifier leads to a saturating nonlinear response that qualitatively resembles the positive half of the sigmoid function. Additionally, the image intensifier can connect two layers and amplify signals, thereby increasing the system’s light flux.

    4. ONNS WITH TRAINING OF FFA FOR APPLICATIONS

    A. Optimization of the Trainable Weights in ONNs

    The training schematic is illustrated in Fig. 2(a). The entire process consists of two components: the black boxes, representing the optical process without a mathematical dual, and the trainable parts, which can be either optical or digital processes. The weights in the trainable parts are updated based on the local loss. Each input sample is embedded with a correct or incorrect label, corresponding to positive or negative data, respectively. The method for constructing data is depicted in Fig. 2(b). White squares are added to the four corners of a sample, representing binary values from 0001 to 0101 (decimal values from 1 to 10), which correspond to the digits from 0 to 9. The 4-bit binary range, spanning from 0000 to 1111, accommodates up to 16 distinct classes. This approach enhances the labeling capacity within a constrained marking area.

    Optimization of the ONNs. (a) The schematic of the training process with FFA. (b) The method for constructing positive and negative samples, using handwritten numbers as an example.

    Figure 2.Optimization of the ONNs. (a) The schematic of the training process with FFA. (b) The method for constructing positive and negative samples, using handwritten numbers as an example.

    B. Experiments for Machine Vision Tasks

    We assessed the effectiveness of the proposed approach by developing a pre-sensor ONN for the task of classifying handwritten digits from the MNIST [48,49] dataset. The experimental setup is shown in Fig. 3(a). The samples are displayed on an LCD screen positioned approximately 15 cm away from the ONN encoder. In the pre-sensor ONN, an amplitude mask with a fixed pattern is placed in front of the image intensifier, which performs a convolution operation and extracts features from the input scene. The process of convolution in the optical domain is based on the principles of geometric optics, performing parts of the computing function [16]. The image intensifier then conducts an optical-to-optical nonlinear activation. An SLM (HOLOEYE LC2012) is positioned at the image plane of the image intensifier (S20 photocathode, IIM-C225 MCP, P46 phosphor screen), serving as a linear amplitude modulation via the Hadamard product. The sensor (Sony IMX 264) is positioned snugly against the SLM. In this configuration, the entire system comprising the mask and the image intensifier is treated as a black box. Due to the fixed pattern of the mask and the difficulty in accurately modeling the nonlinear activation function, these components are considered non-trainable and collectively regarded as a black box. The pattern in the SLM and the weights of the linear layers in the digital neural network are optimized using the FFA.

    (a) Prototype of the system. (b) Corresponding network structure diagram of the system. (c) Confusion matrix in the classification of MNIST dataset, for the condition that only black box exists in the pre-sensor processing. (d) Confusion matrix in the classification of MNIST dataset, for the condition that black box and SLM coexist in the pre-sensor processing. (e) The optimized pattern of SLM. (f) The visualization of the output after black box. The samples represent the result of the convolutional layer formed by the mask and the nonlinear layer provided by the image intensifier. (g) Accuracy and parameters comparison across different models. The output neurons of the digital linear layer vary from 16 to 24. (h) The performances for the classification of CIFAR-4 dataset. The processing approaches contain only digital computing and the combination of optical and digital computing. (i) The optical nonlinear response of the image intensifier used in the experiments.

    Figure 3.(a) Prototype of the system. (b) Corresponding network structure diagram of the system. (c) Confusion matrix in the classification of MNIST dataset, for the condition that only black box exists in the pre-sensor processing. (d) Confusion matrix in the classification of MNIST dataset, for the condition that black box and SLM coexist in the pre-sensor processing. (e) The optimized pattern of SLM. (f) The visualization of the output after black box. The samples represent the result of the convolutional layer formed by the mask and the nonlinear layer provided by the image intensifier. (g) Accuracy and parameters comparison across different models. The output neurons of the digital linear layer vary from 16 to 24. (h) The performances for the classification of CIFAR-4 dataset. The processing approaches contain only digital computing and the combination of optical and digital computing. (i) The optical nonlinear response of the image intensifier used in the experiments.

    When measuring the nonlinear response of the image intensifier, the output light intensity is recorded as the input intensity varies. Figure 3(i) illustrates the relationship between the normalized input and normalized output of one output region in the image intensifier. Referring to Ref. [18], a four-parameter curve can be fit to the measured data pairs. However, the equation for the nonlinear activation curve is derived based solely on empirical observations, assuming its mathematical representation in advance. The nonlinear response can vary and be uneven among different neurons in the MCP, which leads to errors from inaccurate fitting and inconsistent nonlinearity. Moreover, the nonlinear curve is measured using power values, creating a gap between the actual physical and digital quantities when the nonlinear representation is used between layers in the end-to-end optimization of ONNs. Consequently, we can optimize the trainable weights in ONNs using the FFA without knowing the detailed properties of the nonlinear layer or training a digital twin model of it.

    The input images from the MNIST dataset, originally at a resolution of 28×28 pixels, are padded to 32×32 pixels to facilitate marking the positive and negative samples at the edges. The mask features a kernel size of 3×3 with a feature size of 50 μm. The sensor captures an output at a resolution of 300×300 pixels, with each pixel measuring 3.45 μm. The SLM has a pixel size of 36 μm, resulting in a matched pixel arrangement of 30×30. Consequently, the 300×300 pixels are downsampled to 30×30. The prototype of the system and the corresponding hybrid network structure are shown in Figs. 3(a) and 3(b). Several simulated experiments have been conducted to demonstrate the effectiveness of this approach.

    The performances for the classification of the MNIST dataset are presented in Table 1. The fully digital network employing BP and FFA achieved accuracies of 89.0% and 88.5% as the baseline. The integration of the black box enhanced the performance by 4%, reaching an accuracy of 92.5%, thereby underscoring the importance of linear and nonlinear modulation in optical processing. Following the capture post-black box, we optimized the weights in the element-wise Hadamard product layer, as well as the amplitude pattern of the SLM, as illustrated in Fig. 3(e). The incorporation of a linear layer, implemented through the SLM, further improved performance by 1.5% in the simulation and by 0.5% in the experiment, with the corresponding confusion matrices shown in Figs. 3(c) and 3(d). The performances of nonlinear optical processing and linear digital processing in classifying the MNIST dataset with varying output neuron counts are presented in Fig. 3(g). As the number of output neurons increases from 16 to 20, the classification accuracy improves steadily. However, beyond 20 output neurons, the rate of accuracy improvement diminishes. Meanwhile, the number of parameters and computational load continue to grow linearly with the increase in output neurons. Considering the trade-off between performance and computational efficiency, the number of output neurons is set to 20.

    Comparison of Accuracies between Different Processing Approaches in the Classification of the MNIST Dataset

    ProcessingTest AccuracyNetwork (FC)FLOPs
    FC-BP89.0%(784,22,10)3.5 × 104
    FC-FFA88.5%(1024,20)4.1 × 104
    Black Box (optics) + FC-FFA92.5%(900,20)3.6 × 104
    Black Box (optics) + Linear (optics) + FC-FFA-sim94.0%(900,20)3.6 × 104
    Black Box (optics) + Linear (optics) + FC-FFA-exp93.0%(900,20)3.6 × 104

    To further verify the role of linear and nonlinear modulation in the optical black box, we conducted experiments on the CIFAR-4 dataset [50,51], which contains four classes of “dog”, “frog”, “horse”, and “ship”. Initially, the RGB images with three channels were converted to grayscale. These grayscale images were directly connected to a fully connected linear layer (1024-20), whose weights were optimized using the FFA loss function. The results are shown in Fig. 3(h). The baseline classification accuracy for CIFAR-4 was 60.0%. After processing by an optical encoder, it yielded a classification accuracy of 67.5% for CIFAR-4, demonstrating that the optical black box improved task accuracy by 7.5%. It proves the function of pre-sensor optical computing.

    We then performed the classification of the Fashion MNIST [52,53] dataset to evaluate the performance of the multilayer ONN encoder. The experimental setup closely mirrors that used for the classification of the MNIST dataset. The schematic diagram and corresponding network structure are shown in Figs. 4(a) and 4(b). The output of the optical processors, encompassing both nonlinear and linear components, is displayed on the LCD screen, resulting in an additional pre-sensor processing step in the optical domain. Consequently, the input images of fashion categories are processed sequentially through the black box and SLM in two cycles. It can also be realized by sequential placement of image intensifiers, SLM, and even other modulator components. Continuous linear and nonlinear transformations are performed with comprehensive feature extraction.

    Experiments for Fashion MNIST dataset. (a) Schematic diagram of the system. The input includes positive and negative samples. (b) Corresponding network structure diagram of the system. “LM” means “linear modulation.” (c) The optimized pattern of SLM in the first optical linear layer of the Hadamard product. (d) The optimized pattern of SLM in the second optical linear layer of Hadamard product. (e) Confusion matrix in the classification of Fashion MNIST dataset, for the condition that the input images are processed sequentially through the black box and SLM in two cycles.

    Figure 4.Experiments for Fashion MNIST dataset. (a) Schematic diagram of the system. The input includes positive and negative samples. (b) Corresponding network structure diagram of the system. “LM” means “linear modulation.” (c) The optimized pattern of SLM in the first optical linear layer of the Hadamard product. (d) The optimized pattern of SLM in the second optical linear layer of Hadamard product. (e) Confusion matrix in the classification of Fashion MNIST dataset, for the condition that the input images are processed sequentially through the black box and SLM in two cycles.

    The data preparation method for constructing positive and negative samples for the Fashion MNIST dataset is the same as that for the MNIST dataset. The comparison of classification accuracies with different processing approaches is presented in Table 2. As a benchmark, we trained the digital fully connected layer using BP and FFA, achieving accuracies of 81.0% and 78.5%, respectively. When the amplitude mask and image intensifier were added to the optical path for pre-sensor computing, the accuracy improved by 1.5%, reaching 80.0%. We then co-trained the weights of the element-wise Hadamard product layer and the fully connected layer with FFA, achieving an accuracy of 86.0%. With the addition of further nonlinear and linear processing in the optical path, accuracies improved to 87.0% and 87.5%. Optical pre-sensor computing, which includes both linear and nonlinear components, improved accuracy by 9% while maintaining similar floating point operations (FLOPs) in the electronics.

    Comparison of Accuracies between Different Processing Approaches in the Classification of Fashion MNIST Dataset

    ProcessingTest AccuracyNetwork (FC)FLOPs
    FC-BP81.0%(784,22,10)3.5 × 104
    FC-FFA78.5%(1024,20)4.1 × 104
    Black Box (optics) + FC-FFA80.0%(900,20)3.6 × 104
    Black Box (optics) + Linear (optics) + FC-FFA86.0%(900,20)3.6 × 104
    Black Box (optics) + Linear (optics) + Black Box (optics) + FC-FFA87.0%(900,20)3.6 × 104
    Black Box (optics) + Linear (optics) + Black Box (optics) + Linear (optics) + FC-FFA87.5%(900,20)3.6 × 104

    In the classification of the Fashion MNIST dataset, we utilized a loop structure on the hardware to represent the effect of multiple optical computations. Due to the alternation of non-modellable linear layers with optimizable linear layers, the parameters of the linear layers can only be optimized using FFA. This method can also accommodate other nonlinear physical layers without requiring detailed knowledge of light propagation, enhancing the robustness of the optoelectronic system. Compared to the method of first fitting the parts lacking mathematical dual expressions with deep neural networks and then optimizing with BP, FFA enhances the efficiency of the training and reduces memory consumption by isolating the weight updates for individual layers.

    5. DISCUSSION AND CONCLUSION

    In this paper, we propose a pre-sensor multilayer ONN with nonlinear activation, trained using a brain-inspired learning algorithm, FFA. This approach is employed for training ONNs in situations where complete knowledge of the physical processes in the forward pass is unavailable, such as unknown or immeasurable nonlinearities during propagation. In the hardware system, a multilayer ONN is constructed with optimizable linear components and immeasurable nonlinear components. Compared to other ONN training methods, our approach eliminates the need to measure and fit nonlinear curves or to model input-output relationships with deep learning models, resulting in increased training accuracy and reduced computational costs. In the classification of the MNIST and Fashion MNIST datasets, the accuracies were improved by 4.5% and 9.0%, respectively, due to the addition of optical processing.

    The current processing framework integrates pre-sensor computing within the optical domain and post-sensor processing in the electrical domain. Notably, the construction of all-optical neural networks using optical components is both feasible and promising. Convolution leverages the optical principles of spatial geometry, using passive masks or SLM, while the fully connected layer can be implemented through a combination of lenses and LCD modulators. These ONNs can facilitate linear and nonlinear modulation while allowing for training through backpropagation-free methodologies. For nonlinear activation between layers, the nonlinear methods employed in this study enhance the performance of classification tasks. However, the experimental results indicate that the improvement achieved through optical nonlinear activation is indeed limited. Unlike digital deep neural networks, where nonlinear activation functions can be designed arbitrarily, optical nonlinearity is primarily constrained by the inherent properties of the nonlinear materials employed. As a result, the performance gains from nonlinear components in optical systems may differ significantly from those observed in digital systems with mature nonlinear functions such as the “ReLU”. Furthermore, additional approaches to optical nonlinearity, such as quantum dots and photonic crystals, are being explored, which are expected to achieve customized optical nonlinear activation and reduce both power consumption and delay in decision-making with passive materials [19,46,47]. Thus, pre-sensor optical computing can offload part of the computation from electronics into optics by leveraging passive optical elements, further highlighting the advantages such as low power consumption, high speed, and high practicability, easily coupled with mature machine vision systems in edge computing.

    6. MATERIALS AND METHODS

    A. Mask Fabrication

    The optical elements employed for pre-sensor computing in the experiment include an amplitude mask, an image intensifier, and an SLM. The mask utilized was created using photolithography methods on glass substrates coated with chromium. This fabrication process comprised multiple stages, including photolithography, development, etching, and demolding. The specific dimensions of the mask were chosen based on factors such as diffraction and geometric blurring. For our experiment, the feature size of the mask was established at 50 μm.

    B. Dataset Processing and Neural Network Training

    In our experiment, we utilized two standard classification datasets: the MNIST dataset and the Fashion MNIST dataset. We randomly selected a subset of 800 training images and 200 testing images for each dataset, with each class containing 80 training images and 20 testing images. These images were expanded into positive and negative samples. For training, each class had 80 positive samples and 720 negative samples. For testing, each class had 200 samples to evaluate.

    The network training was conducted using PyTorch 1.13.1. During training, the Adam optimizer was employed with a learning rate set at 0.05 for the first 500 epochs and then adjusted to 0.008 for the last 500 epochs. The parameters were optimized by minimizing the local loss, as described in Section 2. All training operations were executed on a workstation equipped with an NVIDIA GeForce RTX 3090 GPU.

    References

    [1] K. O’shea, R. Nash. An introduction to convolutional neural networks. arXiv(2015).

    [2] A. Sherstinsky. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Physica D, 404, 132306(2020).

    [3] N. Parmar, A. Vaswani, J. Uszkoreit. Image transformer. International Conference on Machine Learning(2018).

    [4] X. Sui, Q. Wu, J. Liu. A review of optical neural networks. IEEE Access, 8, 70773-70783(2020).

    [5] J. Liu, Q. Wu, X. Sui. Research progress in optical neural networks: theory, applications and developments. PhotoniX, 2(2021).

    [6] E. Paolini, L. De Marinis, M. Cococcioni. Photonic-aware neural networks. Neural Comput. Appl., 34, 15589-15601(2022).

    [7] X. Lin, Y. Rivenson, N. T. Yardimci. All-optical machine learning using diffractive deep neural networks. Science, 361, 1004-1008(2018).

    [8] J. Li, D. Mengu, Y. Luo. Class-specific differential detection in diffractive optical neural networks improves inference accuracy. Adv. Photonics, 1, 046001(2019).

    [9] Y. Chen, M. Nazhamaiti, H. Xu. All-analog photoelectronic chip for high-speed vision tasks. Nature, 623, 48-57(2023).

    [10] O. Kulce, D. Mengu, Y. Rivenson. All-optical information-processing capacity of diffractive surfaces. Light Sci. Appl., 10, 25(2021).

    [11] M. S. S. Rahman, J. Li, D. Mengu. Ensemble learning of diffractive optical networks. Light Sci. Appl., 10, 14(2021).

    [12] J. Li, D. Mengu, N. T. Yardimci. Spectrally encoded single-pixel machine vision using diffractive networks. Sci. Adv., 7, eabd7690(2021).

    [13] T. Zhou, X. Lin, J. Wu. Large-scale neuromorphic optoelectronic computing with a reconfigurable diffractive processing unit. Nat. Photonics, 15, 367-373(2021).

    [14] M. Miscuglio, Z. Hu, S. Li. Massively parallel amplitude-only Fourier neural network. Optica, 7, 1812-1819(2020).

    [15] T. Yan, J. Wu, T. Zhou. Fourier-space diffractive deep neural network. Phys. Rev. Lett., 123, 023901(2019).

    [16] W. Shi, Z. Huang, H. Huang. LOEN: lensless opto-electronic neural network empowered machine vision. Light Sci. Appl., 11, 121(2022).

    [17] H. Zheng, Q. Liu, I. I. Kravchenko. Intelligent multi-channel meta-imagers for accelerating machine vision. arXiv(2023).

    [18] T. Wang, M. M. Sohoni, L. G. Wright. Image sensing with multilayer nonlinear optical neural networks. Nat. Photonics, 17, 408-415(2023).

    [19] Z. Huang, W. Shi, S. Wu. Pre-sensor computing with compact multilayer optical neural network. Sci. Adv., 10, eado8516(2024).

    [20] M. H. Tahersima, K. Kojima, T. Koike-Akino. Deep neural network inverse design of integrated photonic power splitters. Sci. Rep., 9, 1368(2019).

    [21] Y. Qu, H. Zhu, Y. Shen. Inverse design of an integrated-nanophotonics optical neural network. Sci. Bull., 65, 1177-1183(2020).

    [22] R. Rojas, R. Rojas. The backpropagation algorithm. Neural Networks: A Systematic Introduction, 149-182(1996).

    [23] J. Leonard, M. A. Kramer. Improvement of the backpropagation algorithm for training neural networks. Comput. Chem. Eng., 14, 337-341(1990).

    [24] Z. Wang, C. Li, P. Lin. In situ training of feed-forward and recurrent convolutional memristor networks. Nat. Mach. Intell., 1, 434-442(2019).

    [25] T. Zhou, L. Fang, T. Yan. In situ optical backpropagation training of diffractive optical neural networks. Photonics Res., 8, 940-953(2020).

    [26] X. Guo, T. D. Barrett, Z. M. Wang. Backpropagation through nonlinear units for the all-optical training of neural networks. Photonics Res., 9, B71-B80(2021).

    [27] L. G. Wright, T. Onodera, M. M. Stein. Deep physical neural networks trained with backpropagation. Nature, 601, 549-555(2022).

    [28] Y. Huo, H. Bao, Y. Peng. Optical neural network via loose neuron array and functional learning. Nat. Commun., 14, 2535(2023).

    [29] Z. Zheng, Z. Duan, H. Chen. Dual adaptive training of photonic neural networks. Nat. Mach. Intell., 5, 1119-1129(2023).

    [30] I. Oguz, J. Ke, Q. Weng. Forward–forward training of an optical neural network. Opt. Lett., 48, 5249-5252(2023).

    [31] X. Yuan, Y. Wang, Z. Xu. Training large-scale optoelectronic neural networks with dual-neuron optical-artificial learning. Nat. Commun., 14, 7110(2023).

    [32] A. Momeni, B. Rahmani, M. Malléjac. Backpropagation-free training of deep physical neural networks. Science, 382, 1297-1303(2023).

    [33] M. Nakajima, K. Inoue, K. Tanaka. Physical deep learning with biologically inspired training method: gradient-free approach for physical hardware. Nat. Commun., 13, 7847(2022).

    [34] Z. Xue, T. Zhou, Z. Xu. Fully forward mode training for optical neural networks. Nature, 632, 280-286(2024).

    [35] D. J. Gauthier, E. Bollt, A. Griffith. Next generation reservoir computing. Nat. Commun., 12(2021).

    [36] M. Yan, C. Huang, P. Bienstman. Emerging opportunities and challenges for the future of reservoir computing. Nat. Commun., 15, 2056(2024).

    [37] D. Brunner, M. C. Soriano, G. Van der Sande. Photonic reservoir computing. Gruyter, 8, 19(2019).

    [38] S. Sunada, A. Uchida. Photonic reservoir computing based on nonlinear wave dynamics at microscale. Sci. Rep., 9, 19078(2019).

    [39] S. Phang. Photonic reservoir computing enabled by stimulated Brillouin scattering. Opt. Express, 31, 22061-22074(2023).

    [40] G. Tanaka, T. Yamane, J. B. Héroux. Recent advances in physical reservoir computing: A review. Neural Netw., 115, 100-123(2019).

    [41] G. Hinton. The forward-forward algorithm: some preliminary investigations. arXiv(2022).

    [42] J. Feldmann, N. Youngblood, C. D. Wright. All-optical spiking neurosynaptic networks with self-learning capabilities. Nature, 569, 208-214(2019).

    [43] M. A. Nahmias, B. J. Shastri, A. N. Tait. A leaky integrate-and-fire laser neuron for ultrafast cognitive computing. IEEE J. Sel. Top. Quantum Electron., 19(2013).

    [44] C. Teng, J. Zou, X. Tang. MoS2 as nonlinear optical material for optical neural networks. IEEE J. Sel. Top. Quantum Electron., 29(2023).

    [45] G. K. Lim, Z. L. Chen, J. Clark. Giant broadband nonlinear optical absorption response in dispersed graphene single sheets. Nat. Photonics, 5, 554-560(2011).

    [46] W. Shi, X. Jiang, Z. Huang. Lensless opto-electronic neural network with quantum dot nonlinear activation. Photonics Res., 12, 682-690(2024).

    [47] M. Miscuglio, A. Mehrabian, Z. Hu. All-optical nonlinear activation function for photonic neural networks. Opt. Mater. Express, 8, 3851-3863(2018).

    [48] Y. LeCun, B. Boser, J. S. Denker. Backpropagation applied to handwritten zip code recognition. Neural Comput., 1, 541-551(1989).

    [49] https://yann.lecun.com/exdb/mnist/. https://yann.lecun.com/exdb/mnist/

    [50] A. Krizhevsky, G. Hinton. Learning multiple layers of features from tiny images, 7(2009).

    [51] https://www.cs.toronto.edu/~kriz/cifar.html. https://www.cs.toronto.edu/~kriz/cifar.html

    [52] H. Xiao, K. Rasul, R. Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv(2017).

    [53] https://tensorflow.google.cn/datasets/catalog/fashion_mnist. https://tensorflow.google.cn/datasets/catalog/fashion_mnist

    Zheng Huang, Conghe Wang, Caihua Zhang, Wanxin Shi, Shukai Wu, Sigang Yang, Hongwei Chen, "Brain-like training of a pre-sensor optical neural network with a backpropagation-free algorithm," Photonics Res. 13, 915 (2025)
    Download Citation