Human emotion recognition with a microcomb-enabled integrated optical neural network
-
Junwei Cheng
, Yu Liu
, Ke Zhou
und Xinliang Zhang
Abstract
State-of-the-art deep learning models can converse and interact with humans by understanding their emotions, but the exponential increase in model parameters has triggered an unprecedented demand for fast and low-power computing. Here, we propose a microcomb-enabled integrated optical neural network (MIONN) to perform the intelligent task of human emotion recognition at the speed of light and with low power consumption. Large-scale tensor data can be independently encoded in dozens of frequency channels generated by the on-chip microcomb and computed in parallel when flowing through the microring weight bank. To validate the proposed MIONN, we fabricated proof-of-concept chips and a prototype photonic-electronic artificial intelligence (AI) computing engine with a potential throughput up to 51.2 TOPS (tera-operations per second). We developed automatic feedback control procedures to ensure the stability and 8 bits weighting precision of the MIONN. The MIONN has successfully recognized six basic human emotions, and achieved 78.5 % accuracy on the blind test set. The proposed MIONN provides a high-speed and energy-efficient neuromorphic computing hardware for deep learning models with emotional interaction capabilities.
1 Introduction
Along with the surge in researches related to artificial intelligence (AI) and massive success in the past decade, deep learning–owing to their ability to mimic human activities such as thinking and decision-making as well as understand human emotion–has become a powerful tool for advanced scientific and industrial applications, including image recognition [1–3], autonomous driving [4–6], and chatbots [7–9]. To achieve smarter applications, the size of parameters in deep learning models is starting to explosively increase, for example, the initial classical models represented by handwritten digit recognition generally have less than a thousand parameters, while the state-of-the-art ChatGPT large language model released by OpenAI already has 175 billion parameters [8], and its next-generation model based on GPT-4 may have up to one trillion parameters. Current deep learning models are typically trained and inferred using digital-clock-based platforms, such as central processing units (CPUs) or graphics processing units (GPUs). They are both highly reconfigurable microelectronic processors that can perform various types of computational tasks, but their clock frequency is only a few GHz [10]. The mismatch between the processor’s computing speed and memory access speed will cause bandwidth bottlenecks that cannot meet the demand for high-speed and low-latency data processing.
Compared with microelectronic devices, photonic devices have the unique advantages of low loss and large bandwidth. Massive amounts of data can be encoded in multiple dimensions of light such as wavelength, amplitude, phase, mode, and polarization by advanced multiplexing techniques to achieve parallel processing [11–21]. Optical neural networks (ONNs) use light as information carrier to implement compute-intense operations in deep neural networks (DNNs), which can significantly improve computing speed and reduce energy consumption. Light propagation can be considered as computation in ONNs. Thus, a specific computational function can be implemented while light propagating through a specially designed dielectric structure or free space. Motivated by these, a myriad of novel hardware structures for optical computing have recently been reported, including coherent architectures based on an on-chip Mach–Zehnder interferometer (MZI) mesh [22–26], incoherent architectures based on wavelength division multiplexing (WDM) techniques [27–33], and diffractive optical frameworks [34–39]. Among these optical computing architectures, the WDM-based scheme has excellent parallelism. Each wavelength can be regarded as an independent channel, and the computational capacity of the system can be expanded by increasing the number of wavelengths. Therefore, this architecture is particularly suitable for tensor computations in practical deep learning models.
However, WDM-based ONNs face significant challenges in complex deep learning applications. Limited by the integration density and computational precision, the ONN-related studies reported so far are mainly demonstrated on simple datasets (e.g., MNIST handwritten digit recognition). The mainstream time-wavelength modulation ONN architectures are bulky because they rely on fiber components, leading to difficulties in integrating them into system-on-a-chip. On the other hand, WDM-based ONN architectures based on on-chip resonant devices (microring [27, 31, 40], [41], [42], nanobeam [43], etc.) are vulnerable to the external environment, and the state of the resonant devices is difficult to maintain accurately, which might introduce additional errors. Moreover, the structure of neural network models is relatively simple, and the problem of error accumulation for optical analog computation in running deep multilayer models has not yet emerged under such model scale.
In this paper, we propose and demonstrate a microcomb-enabled integrated optical neural network (MIONN) architecture with key photonic integrated components, including an integrated optical frequency comb for providing multi-wavelength light source and an on-chip microring weight bank for tensor convolution. The dissipative Kerr solitons (DKS) microcomb generated in a silicon nitride microring cavity provides equally spaced frequency lines with 100 GHz spacing. The microring weight bank consists of a series of configurable microring synapses with self-calibrating capability. Thanks to the feedback control, the weighting precision of microrings can achieve 8 bits. To evaluate the proposed MIONN, we perform two proof-of-concept experiments, including parallel image convolution and human emotion recognition. Large-scale pixels of images can potentially be processed at the speed of light with the throughput of 51.2 TOPS and high energy efficiency of 4.18 TOPS/W. The results of image convolution are comparable to that of a 64-bit digital computer, and the human emotion recognition task has achieved 78.5 % accuracy on the blind test set. Six basic human emotions can be successfully recognized in single image test. Our MIONN provides a scalable neuromorphic architecture to implement tensor computations in practical deep learning models, and presents a promising avenue towards deploying photonic-electronic AI hardware in modern scientific and industrial applications.
2 Principle
Physically implementing compute-intense operations in neural network models is a novel computational paradigm. Our proposed MIONN architecture uses light as an information carrier and can perform tensor convolution operations. Figure 1 shows the basic principle of the MIONN and a convolutional neural network (CNN) model of human emotion recognition. This CNN model consists of convolutional layers, nonlinear layers (including batch normalization and ReLU nonlinear activation), and fully-connected layers. A DKS microcomb generated in a silicon nitride microring serves as the multi-channel light source, and these equally spaced frequency lines are shaped in intensity using a WaveShaper, and then divided by a wavelength division demultiplexer to independently carry normalized data from input images. An external arbitrary waveform generator is used to encode the image information in the electro-optic (EO) modulators. Normalized grayscale values from input images are intensity-encoded in the waveform in radio frequency (RF) domain via an arbitrary waveform generator, and then amplified by an electronic driver. After that, the input data encoded in the RF waveform is fed in the intensity modulator array with different frequency lines. Discrete and modulated frequency lines can be regarded as a series of independent signal channels, and they are multiplexed in the bus fiber and then coupled into the on-chip microring weight bank.

Conceptual drawing of the MIONN for human emotion recognition. Convolution operations in CNN models for human emotion recognition can be implemented at light speed in the MIONN. The DKS microcomb containing a series of equally spaced frequency lines is generated in the silicon nitride microring cavity as a multi-channel light source for the MIONN. The data of input images is encoded on the intensity of each frequency line through electro-optic modulation as the input of the on-chip microring weight bank. The weight parameters of the convolutional kernels are configured on the microrings of the on-chip weight bank, and the weight values can be changed on demand by adjusting the state of microrings. Photodetectors receive the output of on-chip microring weight bank and perform weighted summation operation. DeMUX: wavelength division demultiplexer, MUX: wavelength multiplexer.
The weight parameters of convolutional kernels are configured on the on-chip microring weight bank by applying programmable voltages to the microheaters of each microring. The on-chip microring weight bank is composed of N rows of microring synapses that correspond to N convolutional kernels to implement tensor convolution, and the size of synapses can be flexibly scaled up or down to match different kernel size. Since microrings are sensitive to thermal variations and inter-channel cross talk, we specially developed a self-calibration approach to maintain the weight stability. The convolution computation is completed when the modulated light propagates through the on-chip microring weight bank, and the weighted summation operation is achieved in photodetectors (PDs). Because the on-chip microring weight bank is configured with multiple different convolutional kernels at one time, multiple different feature maps can be obtained by only one optical tensor convolution operation.
Benefiting from the high refractive index, high nonlinear coefficient, and high damage threshold, the silicon nitride micro-cavity is suitable for generating nonlinear optical phenomena at high power laser input. Figure 2(a) shows a microring cavity fabricated on the silicon nitride platform, which is available at commercial photonic foundries. The silicon nitride microring cavity used for on-chip comb generation has a width-height cross section size of 1.65 × 0.8 μm2 and a free spectral range (FSR) of 100 GHz. The auxiliary laser heating method [44] is adopted to quickly generate and precisely tune the DKS microcomb, and an automatic feedback algorithm is developed to compensate for external variations such as frequency noise, frequency drift and power fluctuations to ensure that the comb state could be maintained for up to several hours. Figure 2(b) shows the photograph of the packaged silicon nitride microring cavity chip with optical input/output (I/O) and the thermoelectric cooler (TEC). Polarization-maintaining fiber is used in optical I/O for stable external driving [45], and the TEC is mounted below the chip to precisely control the temperature. The optical spectral snapshot of the generated Kerr frequency comb with the single DKS state is depicted in Figure 2(c), and four frequency lines near 1550 nm are selected for the proof-of-concept experiments.

Design of the DKS microcomb and the on-chip weight bank in MIONN. (a) Microscope image of the silicon nitride microring cavity. (b) Photograph of the packaged comb chip. (c) Measured optical spectrum of microcomb. Four frequency lines near 1550 nm are selected for the proof-of-concept experiments. (d) Microscope image of the on-chip microring weight bank. (e) Photograph of the packaged microring weight bank chip. (f) The transmission spectra of the microring synapse at the through and drop ports and the optical power of the shaped frequency lines in the optical spectral snapshot.
Different from electronic processors, our MIONN performs convolution operations on the integrated photonics platform. The 2D convolution operation (stride = 1) of the input image X and k h × k w convolutional kernel W can be mathematically expressed as
where (m, n) is the position coordinate value of the convolutional result Y m,n in the output feature map, w i,j is the weight parameter of the convolutional kernel, k h is the height of the convolutional kernel, and k w is the width of the convolutional kernel. To visualize how the original image is encoded in different wavelength channels and the mapping of the convolutional kernel and the microring weight bank, we provide a schematic of the working data flow in Supplementary Note 1.
Figure 2(d) shows the micrograph of the microring weight bank chip, which is fabricated by a commercial photonic foundry with a standard silicon-on-insulator (SOI) process. Details of device fabrication can be found in Supplementary Note 2. The microring weight bank is comprised of N microring synapses, and each synapse corresponds an individual k h × k w convolutional kernel, thus the equivalent dimensions of this microring weight bank are [N = 4, k h = 2, k w = 2]. The height and width of convolutional kernel can be set as desired, as long as the product of k h × k w does not exceed the number of microrings in a synapse. Importantly, the equivalent dimensions [N, k h , k w ] can also be flexibly extended by increasing the number and size of microring synapses in the weight bank to enable massive tensor convolution. Thermo-optic phase shifters made of TiN heaters are used to tune the microrings, and the weight parameters of each microring can be precisely configured by setting corresponding voltages on its heater. Thermal isolation trenches are employed between each microring of the weight bank to reduce the thermal cross talk. The packaged weight bank chip (see in Figure 2(e)) has optical and electronic I/O for light propagation and weight parameter control. Optical signals that encoded the image data enter and leave the weight bank chip via vertical grating couplers, and programmed voltage control is applied to the on-chip heaters through wire bonding. The TEC module is mounted below the weight bank chip to compensate temperature variations. Figure 2(f) displays the transmission spectra of the microring synapse at the through and the drop ports and the optical power of the shaped frequency lines. Detailed characterization of the microring in the on-chip weight bank can be found in Supplementary Note 3.
3 Results
In this section, we first experimentally validate the effectiveness of high-precision photonic tensor convolution using the proposed MIONN, including the systematic feedback control and the tensor convolution results. Next, we demonstrate the successful recognition of six human emotions using the MIONN.
3.1 High-precision photonic tensor convolution using MIONN
Precision is an important metric to evaluate the analog computing hardware, especially when facing large human-computer emotional interaction models and complex computing tasks. As an analog computing architecture, the optical comb generation and the on-chip weight bank of the MIONN are both realized by microring cavity chips. Because microring cavity is a compact resonant device, its state is sensitive to fabrication and environmental variations, including fabrication errors, frequency drift, optical power fluctuations, thermal crosstalk, etc. Therefore, MIONN requires customized automatic feedback control to ensure high computational precision. Here, we specifically designed a systematic feedback control procedure for MIONN to adjust the photonic devices of each functional module in real time by monitoring related optical parameters in the whole optical link. There are two procedures to maintain the system stable. One is to maintain the stability of the generated DKS microcomb, and the other is to calibrate the weight of each microring in the on-chip weight bank.
For the DKS microcomb, the auxiliary laser heating method [44] is adopted to quickly generate the microcomb. Figure 3(a) shows 25 central frequency lines of the DKS microcomb at single soliton state. To maintain the single soliton state for a long time, our feedback procedure uses the total power of the microcomb to compensate for the effect of environmental variations. Particularly, the total power of the single soliton state is regarded as the threshold value, and we define a figure of metric (FOM) as the difference between the real-time power of the microcomb and the threshold value. When the FOM is less than zero, the feedback control is activated and all resonances are thermally red-shifted while the auxiliary laser is tuned closer to the resonance peak, then the pump laser can re-enter the cavity from the red detuning region and trigger the DKS burst. The power deviations of the 25 central frequency lines of the microcomb in Figure 3(a) within 60 min are measured, and the results of statistical analysis are depicted in Figure 3(b). Results show that the power fluctuation of the central frequency lines are all less than 2 dB, and most relative power deviations are less than 1 dB, which proves the validity of the feedback control and ensures high computational precision of the MIONN.

Systematic feedback control for the MIONN. (a) Optical spectra of the DKS microcomb at single soliton state with a FSR of 100 GHz. (b) Power deviations of the 25 central frequency lines of the microcomb. (c) Comparation of the weighting precision in a two-microring synapse before and after calibration. Each blue/red dot represents a measured weight before/after calibration, and each point on the grid is tested three times.
Efficient calibration of the on-chip weight bank is the other key to ensure the precision of the MIONN. For every microring in the on-chip weight bank, a one-to-one mapping between the configured weights and the applied voltages is established before the calibration. We specially developed a self-calibration procedure that works in concert with the circuit hardware to monitor and calibrate the microring weight bank (see the Supplementary Note 4 for details). To achieve high-precision voltage scanning, customed power supply (see the Supplementary Note 5 for details) with 16-bit resolution to ensure accurate voltages applied to the microring, and the technical information of the balanced photodetector is given in the Supplementary Note 6. The weight of each microring can be precisely adjusted and corrected once the current measured weight does not match the set weight. Figure 3(c) shows the test result on a mesh-plot. This mesh-estimation is a common method to visually describe the weighting precision of microrings [40, 42, 46], in which two microrings are configured to equidistant grid points of (w 1, w 2), w 1,2 ∈ [−1, 1], and the weight precision is evaluated at each grid point. Grid points correspond to combinations of weight, and each combination is tested three times. From the test results, we observe that the calibrated weights are closer to the set weights, and the weighting precision can reach 8 bits (the estimation method can be found in Ref. [42]), verifying the validity of the self-calibration procedure.
Image convolution is a fundamental operation in CNN models, which is used to learn spatial patterns and features from input images. The image convolution operation can be expressed as the weighted summation of element-wise products of a small matrix (called the convolutional kernel) with corresponding image patch in the input image, i.e., the kernel weights are assigned to pixel values of the input image. In each image convolution operation, the convolutional kernel slides over the input image with preset step length and the weighted summation results are reshaped to a feature map. By decomposing the original matrix into the positive matrix and the negative matrix [30], the computational domain can be extended from the domain of positive numbers to the domain of real numbers. Details on the photonic matrix multiplication including negative number are given in Supplementary Note 7. To validate the convolution performance of the MIONN, we use a 256 × 256 grayscale ‘Cameraman’ image with 8-bit color depths as the standard test image, and nine commonly used convolutional kernels are employed to extract its different features. In the MIONN, each frequency line of the microcomb represents a signal channel used for loading multi-channel input images. The grayscale values of the pixels of input images are normalized and then arranged as the serial data flow fed in 10 GHz-bandwidth electro-optic modulators. After temporal modulation, these pixel values are encoded in the intensities of frequency lines and then weighted by the on-chip microring weight bank during light propagation. Optically-weighted results are received by the high-speed PDs, in which the power of overlapped frequency lines is added as the optical convolution results and then reshaped to final feature maps. Figure 4 presents the optical convolution results of the standard test image convolved with nine different kernels, including classical image processing and edge extraction. The Sobel operator is used as the kernels for image edge extraction to compute the gradient approximation of the image intensity at each pixel. We observe that the feature maps yielded by optical convolution successfully implement different image processing (blur, motion blur, emboss, outline, sharpen) and highlight the sharp edges in the origin image in bottom/top/left/right direction. Experimental results closely resemble the theoretical feature maps generated by a 64-bit digital computer, which verifies the capability of the MIONN to perform tensor convolution.

Tensor convolution results implemented by the MIONN. The standard test image is convolved with nine different kernels, including classical image processing (blur, motion blur, emboss, outline, sharpen) and edge extraction with bottom/top/left/right Sobel operator.
3.2 MIONN for human emotion recognition
Beyond the multi-channel image convolution, we further demonstrate a CNN model to recognize human emotion in the real-world affective faces (RAF-DB) database [47]. Importantly, our MIONN is fully reconfigurable and scalable, thus in addition to convolutional layers, average pooling and fully connected layers can also be implemented. The matrices involved in pooling and convolution layers are usually large and difficult to be computed at once on the optical hardware, so hardware-friendly algorithms are needed to decompose the large matrices into smaller ones. A matrix decomposition algorithm dedicated to microring weight bank is proposed in our previous work [30] and can be used in optical CNN models. Since the goal of our task is to recognize human facial expressions with rich details, versus handwritten digit images or two-digit recognition, both the size of images in database and the complexity of the CNN model in our experiments are challenging. Figure 5(a) shows the processes and structure of the CNN model used in our experiment, including five convolutional layers, five nonlinear layers (including batch normalization and ReLU nonlinear activation), and two fully-connected layers. The nonlinear activation function in this work is performed by the electronic hardware. We randomly choose 12,000 human emotion images from the RAF-DB dataset and split them into the training set (10,000 images) and testing set (2000 images). The input image is a matrix with a size of 100 × 100, which is much larger than handwritten digit images in the MNIST dataset. Details on the CNN training can be found in Supplementary Note 8. The parameters of CNN are firstly trained with a standard back-propagation algorithm on a digital computer and the MIONN is used to implement convolutional layers and fully-connected layers in the inference phase of single human emotion image.

Human emotion recognition of the RAF-DB dataset using the MIONN. (a) The processes and structure of the CNN model for human emotion recognition. (b) The class activation mappings (CAMs) and the top-3 probability of human emotion images. The MIONN and a 64-bit digital computer are used for the recognition of six basic human emotions, including surprise, fear, disgust, happiness, sadness, and anger. The red bars in the figure represent the probabilities obtained by optical inference, and the blue bars represent the probabilities obtained by electrical inference. (c) Scatter plot for convolution accuracy measurement with a fixed kernel matrix. The inset is a deviation distribution histogram showing the difference of the measured values and the calculated values. (d) The confusion matrices of human emotion classification of the RAF-DB dataset. The accuracy of the prediction results for experiment and calculation shows good agreement.
In the MIONN, the 100 × 100 pixels, 8-bit grayscale images of human emotion are first converted to optical data flow using optical intensity modulation and then fed into the convolutional layer implemented by on-chip microring weight bank. Each four-microring photonic synapse in the on-chip weight bank represents a kernel in the convolutional layer, and four photonic synapses can perform convolution operation simultaneously. The convolution layer first convolutes the input images with four photonic kernels and the results are then processed with batch normalization and ReLU nonlinear activation to generate feature maps. As the kernels slides along the input images, the photonic synapses can extract key features used for recognition. To visualize which facial features the photonic synapses are sensitive to, class activation mappings (CAMs) of six different human emotion images are yielded and depicted in Figure 5(b). These CAMs display the sensitive areas of photonic synapses by heat map, which helps us investigate key features that contribute the recognition results. After serializing into a feature vector through flatten operation, the convolutional results are fed into the fully-connected layer and then the final output is obtained by the softmax function. The final output gives the predicted probabilities of the six emotion labels of the input image, and the largest of them is taken as the recognition result. Figure 5(b) also provides the top-3 probability of six human emotion images obtained by both the MIONN and a typical 64-bit digital computer (Intel(R) Core(TM) i9-12900K CPU, 64 GB RAM). By extracting key features of the human face such as mouth, nose, and the corner of the eyes using photonic synapses, the MIONN has successfully achieved the recognition of six basic human emotions and the recognition results are comparable to those of digital computers.
To further evaluate the computational accuracy of the MIONN, the human emotion image labelled as ‘anger’ is chosen to execute convolution operation in the first convolutional layer, and the comparison of the optical and theoretical convolution results are presented in Figure 5(c). The scatter points are closely distributed along the diagonal line and most of the points are within the blue shaded area, i.e., the deviations are within the range of [−0.05, 0.05], indicating that the measured and theoretical values are in excellent agreement. We then randomly feed 1000 images from the RAF-DB human emotion dataset into both the MIONN and the digital computer to test the accuracy classification tasks. Since the size of the input matrix is much larger than the MNIST handwritten digital dataset, which is frequently used to benchmark in the field of optical computing (see Supplementary Note 9 for details), we only perform the first convolutional layer with the microring weight bank for proof-of-concept in this test. The diffusion matrices with six categories of human emotion (‘surprise’, ‘fear’, ‘disgust’, ‘happiness’, ‘sadness’, and ‘anger’) obtained by both the MIONN (marked as ‘experiment’) and the digital computer (marked as ‘calculated’) are displayed in Figure 5(d). The blind test set accuracies of the MIONN and the digital computer are 78.5 and 82.3 %, respectively. We simulate the impact of Gaussian noise with different standard deviations on the inference accuracy of CNN and find that only when the standard deviation of Gaussian noise is less than 0.1, the inference accuracy does not decrease significantly (see Supplementary Note 10 for details). Results show that the classification accuracy of the MIONN is comparable to that of a digital computer, and the ability of the MIONN to recognize human emotions is verified.
4 Discussion
In our proof-of-concept experiments, the MIONN demonstrate the successful recognition of human facial expression, which is one of the most effective and universal ways for humans to express emotions and intentions. To take full advantage of the parallelism of photonic computing, we use an integrated DKS microcomb as the light source, and tens of generated frequency lines can be used as independent channels to load data to support optical tensor convolution of large size input images with richer details. On-chip microring weight bank with compact footprint can be straightforwardly configured to simulate the weights of convolutional kernels. Thanks to calibration procedures specifically designed for optical frequency comb and microring weight bank, our MIONN achieves a weighting precision of up to 8 bits, which is comparable to state-of-the-art electronic neuromorphic processors. In addition, the MIONN architecture is highly scalable, the number of channels and the size of the on-chip weight bank can be flexibly scaled, and its throughput can be significantly improved by adopting frequency lines with smaller intervals and more compact microring layout.
To demonstrate the scalability of the MIONN architecture, we further develop a packaged 16 × 16 photonic weight bank chip and a base with integrated circuit control modules. In this prototype, the photonic chip and electrical base can be considered as a photonic-electronic AI computing engine for high-throughput tasks. Figure 6(a) shows the photonic-electronic AI computing engine packaged with complete optical and electrical I/O, and the zoom-in photo of the photonic chip is shown in Figure 6(b). Optical signals are connected on and off the chip through a vertically coupled grating array, and 512-channel high-speed electrical I/O is achieved through standard pluggable connecting finger interconnected with external control circuits instead of bulky Jumper wires. Figure 6(c) displays the micrograph of the photonic chip, each microring in the 16 × 16 on-chip weight bank has an equal spacing of 0.14 mm, so the footprint of a microring is 0.14 × 0.14 = 0.0196 mm2. The structure of the microring is designed to match the frequency lines with 50 GHz interval, providing more than 80 channels available in the commonly used C-band to support AI applications requiring high throughput. Since the microcomb chip with a frequency spacing of 50 GHz is to be fabricated, systematic test of the photonic-electronic AI computing engine has not been conducted for the time being.

Prototype of photonic-electronic AI computing engine. (a) The photonic-electronic AI computing engine packaged with complete optical and electrical I/O. 512-channel high-speed electrical I/O is achieved through standard pluggable connecting finger interconnected with external control circuits instead of bulky Jumper wires. (b) The zoom-in photo of the packaged photonic chip. (c) Micrograph of the 16 × 16 on-chip microring weight bank.
Table 1 summarizes the comparison between our MIONN architecture and other neuromorphic computing hardware in terms of key performance metrics. Due to the scalability, we evaluate the MIONN with small-scale (N = 4) and large-scale (N = 16) on-chip microring weight bank. The small-scale MIONN can achieve a throughput of 3.2 TOPS, a computing density of 10.2 TOPS/mm2, and an energy efficiency of 1.21 TOPS/W. As the scale increases, the advantages of throughput and energy efficiency of MIONN architecture begin to emerge. Large-scale MIONN has the potential to support a throughput of 51.2 TOPS and a high energy efficiency of 4.18 TOPS/W. The energy consumption here does not include external benchtop instruments such as arbitrary waveform generator, oscilloscope, and the computer used to train neural networks. We also conduct statistics on the overall energy consumption with external benchtop instruments, and detailed information on key metrics of the MIONN can be found in Supplementary Note 11. Since the throughput is directly proportional to the size of the on-chip weight bank, the scalable MIONN architecture can be designed with highly flexible scaling to meet the needs of different application scenarios. The reconfigurable and compact microring is an ideal basic unit for neuromorphic computing because it can scale up largely, and thus support high computing density architecture. Since the process of optical computing is almost completely passive and only requires very low power consumption to maintain the resonant state of microrings, the MIONN has high energy efficiency. In contrast to bulky space diffractive systems and fiber optic systems, the key components of MIONN are integrated and can be fabricated in CMOS-compatible platforms, which is important for the miniaturization and commercialization of computing hardware.
Comparison of the MIONN and other neuromorphic computing hardware.
Technology | Throughout (TOPS)a | Computing density (TOPS/mm2)b | Energy efficiency (TOPS/W)b | Precision (bits) | Integration |
---|---|---|---|---|---|
MZI mesh [22] | 6.4 | N/A | N/A | 8 | Yes |
MZI mesh [26] | 7.2 | 9.88 | 1.84 | 4 | Yes |
MRR array [48] | 1.6 | 4.21 | 0.7 | >5.5 | Yes |
MRR array [31] | 2.4 | 13.88 | 0.3 | 7 | Yes |
WDM + PCM [28] | 28.8 | 1.2 | 0.4 | 5 | Yes |
InP SOA [49] | 1.6 | N/A | 0.59 | 4.5 | Yes |
WDM + ONN [29] | 11.32 | N/A | 1.25 | 8 | No |
This work | 3.2 (N = 4) | 10.2 | 1.21 (N = 4) | 8 | Yes |
51.2 (N = 16) | 4.18 (N = 16) |
-
aThe throughput is estimated at a typical detection rate of 100 GHz. bRegarding only the optical comb and weight bank chips.
Furthermore, the key performance metrics of MIONN architecture can be further improved, and here we discuss several feasible ways to further exploit the potential of MIONN. Implementing nonlinear activation functions in the optical domain can avoid frequent photoelectric conversions and improve the practicability of optical neuromorphic hardware. Although the nonlinear activation function in this work is performed by the electronic hardware, two type of optical nonlinear function chips based on Ge/Si hybrid structure have been successfully experimentally demonstrated in our recent works [50, 51], and will be integrated into the MIONN architecture as an important module to compensate for the lack of optical nonlinear functions. To reduce the power consumption, the on-chip comb generation can reach as low as 98 mW [52] using extremely low loss silicon nitride waveguides. On the other hand, conventional thermo-optical phase shifters can be replaced by phase shifters based on phase change material (PCM). PCM-based devices can operate without static energy consumption [28, 53, 54], and PCM only requires energy when switching to different states. High-speed germanium-on-silicon PDs can be directly integrated to the on-chip weight bank to reduce computing latency [55]. In addition, external control circuits can be connected to the photonic chip through advanced 3D stacking or optoelectronic fusion technologies to achieve higher I/O rate and higher integration.
5 Conclusions
In summary, we propose a photonic-electronic MIONN architecture based on CMOS-compatible platform for neuromorphic computing. The MIONN computes through light propagation, and naturally has higher parallelism and lower energy consumption compared to traditional electrical neural networks. Enabled by dozens of channels provided by the integrated optical microcomb, we develop a prototype photonic-electronic AI computing engine with a potential throughput up to 51.2 TOPS, a computing density of 10.2 TOPS/mm2, and a high energy efficiency of 4.18 TOPS/W. The auxiliary laser heating method and automatic feedback control ensure rapid generation and long-term stability of the microcomb (less than 2 dB power fluctuation within 60 min), and a self-calibration procedure for the on-chip microring weight bank is developed to achieve 8 bits weighting precision comparable to that of microelectronic neuromorphic processors. In proof-of-concept experiments, our MIONN is able to perform tensor convolution operations with high accuracy in the optical domain, and can run a multi-layer CNN model to successfully recognize human emotion images with a blind test set accuracy of 78.5 %. The proposed MIONN provides a promising direction for next-generation computing hardware for compute-intense AI applications.
Funding source: National Natural Science Foundation of China
Award Identifier / Grant number: 62075075, 62275088, U21A20511
Funding source: Innovation Project of Optics Valley Laboratory
Award Identifier / Grant number: OVL2021BG001
Funding source: National Key Research and Development Program of China
Award Identifier / Grant number: 2022YFB2804200
-
Research funding: This work was supported by the National Key Research and Development Program of China (2022YFB2804200), the National Natural Science Foundation of China (62075075, 62275088, U21A20511), the Innovation Project of Optics Valley Laboratory (Grant No. OVL2021BG001).
-
Author contributions: All the authors have accepted responsibility for the entire content of this submitted manuscript and approved submission.
-
Conflict of interest: The authors declare no conflicts of interest.
-
Data availability: All data supporting the findings of this study are available from the corresponding author upon reasonable request.
References
[1] S. Ji, W. Xu, M. Yang, and K. Yu, “3D convolutional neural networks for human action recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 1, pp. 221–231, 2013, https://doi.org/10.1109/tpami.2012.59.Suche in Google Scholar
[2] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015, https://doi.org/10.1038/nature14539.Suche in Google Scholar PubMed
[3] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Commun. ACM, vol. 60, no. 6, pp. 84–90, 2017, https://doi.org/10.1145/3065386.Suche in Google Scholar
[4] S. Grigorescu, B. Trasnea, T. Cocias, and G. Macesanu, “A survey of deep learning techniques for autonomous driving,” J. Field Robot., vol. 37, no. 3, pp. 362–386, 2020, https://doi.org/10.1002/rob.21918.Suche in Google Scholar
[5] B. R. Kiran, I. Sobh, V. Talpaert, et al.., “Deep reinforcement learning for autonomous driving: a survey,” IEEE Trans. Intell. Transport. Syst., vol. 23, no. 6, pp. 4909–4926, 2022, https://doi.org/10.1109/tits.2021.3054625.Suche in Google Scholar
[6] S. Mozaffari, O. Y. Al-Jarrah, M. Dianati, P. Jennings, and A. Mouzakitis, “Deep learning-based vehicle behavior prediction for autonomous driving applications: a review,” IEEE Trans. Intell. Transport. Syst., vol. 23, no. 1, pp. 33–47, 2022, https://doi.org/10.1109/tits.2020.3012034.Suche in Google Scholar
[7] A. Chatterjee, U. Gupta, M. K. Chinnakotla, et al.., “Understanding emotions in text using deep learning and big data,” Comput. Hum. Behav., vol. 93, pp. 309–317, 2019, https://doi.org/10.1016/j.chb.2018.12.029.Suche in Google Scholar
[8] T. Brown, B. Mann, N. Ryder, et al.., “Language models are few-shot learners,” Adv. Neural Inf. Process. Syst., vol. 33, pp. 1877–1901, 2020.Suche in Google Scholar
[9] N. Kurian, J. M. Cherian, N. A. Sudharson, K. G. Varghese, and S. Wadhwa, “AI is now everywhere,” Br. Dent. J., vol. 234, no. 2, pp. 72, 2023, https://doi.org/10.1038/s41415-023-5461-1.Suche in Google Scholar PubMed
[10] N. S. Kim, T. Austin, D. Blaauw, et al.., “Leakage current: Moore’s law meets static power,” Computer, vol. 36, no. 12, pp. 68–75, 2003, https://doi.org/10.1109/mc.2003.1250885.Suche in Google Scholar
[11] J. Cheng, H. Zhou, and J. Dong, “Photonic matrix computing: from fundamentals to applications,” Nanomaterials, vol. 11, no. 7, p. 1683, 2021, https://doi.org/10.3390/nano11071683.Suche in Google Scholar PubMed PubMed Central
[12] B. J. Shastri, A. N. Tait, T. Ferreira de Lima, et al.., “Photonics for artificial intelligence and neuromorphic computing,” Nat. Photonics, vol. 15, no. 2, pp. 102–114, 2021, https://doi.org/10.1038/s41566-020-00754-y.Suche in Google Scholar
[13] R. Xu, P. Lv, F. Xu, and Y. Shi, “A survey of approaches for implementing optical neural networks,” Opt. Laser Technol., vol. 136, p. 106787, 2021, https://doi.org/10.1016/j.optlastec.2020.106787.Suche in Google Scholar
[14] M. A. Al-Qadasi, L. Chrostowski, B. J. Shastri, and S. Shekhar, “Scaling up silicon photonic-based accelerators: challenges and opportunities,” APL Photonics, vol. 7, no. 2, p. 020902, 2022, https://doi.org/10.1063/5.0070992.Suche in Google Scholar
[15] T. Badloe, S. Lee, and J. Rho, “Computation at the speed of light: metamaterials for all-optical calculations and neural networks,” Adv. Photonics, vol. 4, no. 6, p. 064002, 2022, https://doi.org/10.1117/1.ap.4.6.064002.Suche in Google Scholar
[16] C. Huang, V. J. Sorger, M. Miscuglio, et al.., “Prospects and applications of photonic neural networks,” Adv. Phys. X, vol. 7, no. 1, p. 1981155, 2022, https://doi.org/10.1080/23746149.2021.1981155.Suche in Google Scholar
[17] J. Wu, X. Lin, Y. Guo, et al.., “Analog optical computing for artificial intelligence,” Engineering, vol. 10, pp. 133–145, 2022, https://doi.org/10.1016/j.eng.2021.06.021.Suche in Google Scholar
[18] P. Xu and Z. Zhou, “Silicon-based optoelectronics for general-purpose matrix computation: a review,” Adv. Photonics, vol. 4, no. 4, p. 044001, 2022, https://doi.org/10.1117/1.ap.4.4.044001.Suche in Google Scholar
[19] H. L. Zhou, J. J. Dong, J. W. Cheng, et al.., “Photonic matrix multiplication lights up photonic accelerator and beyond,” Light Sci. Appl., vol. 11, no. 1, pp. 1–21, 2022, https://doi.org/10.1038/s41377-022-00717-8.Suche in Google Scholar PubMed PubMed Central
[20] Y. Bai, X. Xu, M. Tan, et al.., “Photonic multiplexing techniques for neuromorphic computing,” Nanophotonics, vol. 12, no. 5, pp. 795–817, 2023, https://doi.org/10.1515/nanoph-2022-0485.Suche in Google Scholar
[21] X. Xu, W. Han, M. Tan, et al.., “Neuromorphic computing based on wavelength-division multiplexing,” IEEE J. Sel. Top. Quantum Electron., vol. 29, no. 2, pp. 1–12, 2023, https://doi.org/10.1109/jstqe.2022.3203159.Suche in Google Scholar
[22] Y. Shen, N. C. Harris, S. Skirlo, et al.., “Deep learning with coherent nanophotonic circuits,” Nat. Photonics, vol. 11, no. 7, pp. 441–446, 2017, https://doi.org/10.1038/nphoton.2017.93.Suche in Google Scholar
[23] T. W. Hughes, M. Minkov, Y. Shi, and S. Fan, “Training of photonic neural networks through in situ backpropagation and gradient measurement,” Optica, vol. 5, no. 7, pp. 864–871, 2018, https://doi.org/10.1364/optica.5.000864.Suche in Google Scholar
[24] H. Zhou, Y. Zhao, X. Wang, et al.., “Self-configuring and reconfigurable silicon photonic signal processor,” ACS Photonics, vol. 7, no. 3, pp. 792–799, 2020, https://doi.org/10.1021/acsphotonics.9b01673.Suche in Google Scholar
[25] H. Zhou, Y. Zhao, G. Xu, et al.., “Chip-scale optical matrix computation for PageRank algorithm,” IEEE J. Sel. Top. Quantum Electron., vol. 26, no. 2, pp. 1–10, 2020, https://doi.org/10.1109/jstqe.2019.2943347.Suche in Google Scholar
[26] H. Zhang, M. Gu, X. D. Jiang, et al.., “An optical neural chip for implementing complex-valued neural network,” Nat. Commun., vol. 12, no. 1, p. 457, 2021, https://doi.org/10.1038/s41467-020-20719-7.Suche in Google Scholar PubMed PubMed Central
[27] A. N. Tait, T. F. de Lima, E. Zhou, et al.., “Neuromorphic photonic networks using silicon photonic weight banks,” Sci. Rep., vol. 7, no. 1, p. 7430, 2017, https://doi.org/10.1038/s41598-017-07754-z.Suche in Google Scholar PubMed PubMed Central
[28] J. Feldmann, N. Youngblood, M. Karpov, et al.., “Parallel convolutional processing using an integrated photonic tensor core,” Nature, vol. 589, no. 7840, pp. 52–58, 2021, https://doi.org/10.1038/s41586-020-03070-1.Suche in Google Scholar PubMed
[29] X. Xu, M. Tan, B. Corcoran, et al.., “11 TOPS photonic convolutional accelerator for optical neural networks,” Nature, vol. 589, no. 7840, pp. 44–51, 2021, https://doi.org/10.1038/s41586-020-03063-0.Suche in Google Scholar PubMed
[30] J. Cheng, Y. Zhao, W. Zhang, et al.., “A small microring array that performs large complex-valued matrix-vector multiplication,” Front. Optoelectron., vol. 15, no. 1, p. 15, 2022, https://doi.org/10.1007/s12200-022-00009-4.Suche in Google Scholar PubMed PubMed Central
[31] S. Xu, J. Wang, S. Yi, and W. Zou, “High-order tensor flow processing using integrated photonic circuits,” Nat. Commun., vol. 13, no. 1, p. 7970, 2022, https://doi.org/10.1038/s41467-022-35723-2.Suche in Google Scholar PubMed PubMed Central
[32] B. Bai, Q. Yang, H. Shu, et al.., “Microcomb-based integrated photonic processing unit,” Nat. Commun., vol. 14, no. 1, p. 66, 2023, https://doi.org/10.1038/s41467-022-35506-9.Suche in Google Scholar PubMed PubMed Central
[33] J. Cheng, Z. He, Y. Guo, et al.., “Self-calibrating microring synapse with dual-wavelength synchronization,” Photon. Res., vol. 11, no. 2, pp. 347–356, 2023, https://doi.org/10.1364/prj.478370.Suche in Google Scholar
[34] X. Lin, Y. Rivenson, N. T. Yardimci, et al.., “All-optical machine learning using diffractive deep neural networks,” Science, vol. 361, no. 6406, p. 1004, 2018, https://doi.org/10.1126/science.aat8084.Suche in Google Scholar PubMed
[35] T. Zhou, X. Lin, J. Wu, et al.., “Large-scale neuromorphic optoelectronic computing with a reconfigurable diffractive processing unit,” Nat. Photonics, vol. 15, no. 5, pp. 367–373, 2021, https://doi.org/10.1038/s41566-021-00796-w.Suche in Google Scholar
[36] J. Li, Y.-C. Hung, O. Kulce, D. Mengu, and A. Ozcan., “Polarization multiplexed diffractive computing: all-optical implementation of a group of linear transformations through a polarization-encoded diffractive network,” Light Sci. Appl., vol. 11, no. 1, p. 153, 2022, https://doi.org/10.1038/s41377-022-00849-x.Suche in Google Scholar PubMed PubMed Central
[37] T. Wang, S.-Y. Ma, L. G. Wright, et al.., “An optical neural network using less than 1 photon per multiplication,” Nat. Commun., vol. 13, no. 1, p. 123, 2022, https://doi.org/10.1038/s41467-021-27774-8.Suche in Google Scholar PubMed PubMed Central
[38] T. Fu, Y. Zang, Y. Huang, et al.., “Photonic machine learning with on-chip diffractive optics,” Nat. Commun., vol. 14, no. 1, p. 70, 2023, https://doi.org/10.1038/s41467-022-35772-7.Suche in Google Scholar PubMed PubMed Central
[39] J. Li, T. Gan, B. Bai, et al.., “Massively parallel universal linear transformations using a wavelength-multiplexed diffractive optical network,” Adv. Photonics, vol. 5, no. 1, p. 016003, 2023, https://doi.org/10.1117/1.ap.5.1.016003.Suche in Google Scholar
[40] C. Huang, S. Bilodeau, T. F. D. Lima, et al.., “Demonstration of scalable microring weight bank control for large-scale photonic integrated circuits,” APL Photonics, vol. 5, no. 4, p. 040803, 2020, https://doi.org/10.1063/1.5144121.Suche in Google Scholar
[41] Y. Jiang, W. J. Zhang, F. Yang, and Z. Y. He, “Photonic convolution neural network based on interleaved time-wavelength modulation,” J. Lightwave Technol., vol. 39, no. 14, pp. 4592–4600, 2021, https://doi.org/10.1109/jlt.2021.3076070.Suche in Google Scholar
[42] W. Zhang, C. Huang, H.-T. Peng, et al.., “Silicon microring synapses enable photonic deep learning beyond 9-bit precision,” Optica, vol. 9, no. 5, pp. 579–584, 2022, https://doi.org/10.1364/optica.446100.Suche in Google Scholar
[43] A. Jha, C. Huang, T. F. deLima, et al.., “Nanophotonic cavity based synapse for scalable photonic neural networks,” IEEE J. Sel. Top. Quantum Electron., vol. 28, no. 6, pp. 1–8, 2022, https://doi.org/10.1109/jstqe.2022.3179983.Suche in Google Scholar
[44] H. Zhou, Y. Geng, W. Cui, et al.., “Soliton bursts and deterministic dissipative Kerr soliton generation in auxiliary-assisted microcavities,” Light Sci. Appl., vol. 8, no. 1, p. 50, 2019, https://doi.org/10.1038/s41377-019-0161-y.Suche in Google Scholar PubMed PubMed Central
[45] Y. Geng, W. Cui, J. Sun, et al.., “Enhancing the long-term stability of dissipative Kerr soliton microcomb,” Opt. Lett., vol. 45, no. 18, pp. 5073–5076, 2020, https://doi.org/10.1364/ol.400656.Suche in Google Scholar
[46] A. N. Tait, H. Jayatilleka, T. F. De Lima, et al.., “Feedback control for microring weight banks,” Opt. Express, vol. 26, no. 20, pp. 26422–26443, 2018, https://doi.org/10.1364/oe.26.026422.Suche in Google Scholar PubMed
[47] S. Li and W. Deng, “Real-world affective faces database (RAF-DB),” Available at: http://www.whdeng.cn/RAF/model1.html.Suche in Google Scholar
[48] C. R. Huang, S. Fujisawa, T. F. de Lima, et al.., “A silicon photonic-electronic neural network for fibre nonlinearity compensation,” Nat. Electron., vol. 4, no. 11, pp. 837–844, 2021, https://doi.org/10.1038/s41928-021-00661-2.Suche in Google Scholar
[49] B. Shi, N. Calabretta, and R. Stabile, “Deep neural network through an InP SOA-based photonic integrated cross-connect,” IEEE J. Sel. Top. Quantum Electron., vol. 26, no. 1, pp. 1–11, 2020, https://doi.org/10.1109/jstqe.2019.2945548.Suche in Google Scholar
[50] B. Wu, H. Li, W. Tong, J. Dong, and X. Zhang, “Low-threshold all-optical nonlinear activation function based on a Ge/Si hybrid structure in a microring resonator,” Opt. Mater. Express, vol. 12, no. 3, pp. 970–980, 2022, https://doi.org/10.1364/ome.447330.Suche in Google Scholar
[51] H. Li, B. Wu, W. Tong, J. Dong, and X. Zhang, “All-optical nonlinear activation function based on germanium silicon hybrid asymmetric coupler,” IEEE J. Sel. Top. Quantum Electron., vol. 29, no. 2, pp. 1–6, 2023, https://doi.org/10.1109/jstqe.2022.3166510.Suche in Google Scholar
[52] B. Stern, X. Ji, Y. Okawachi, A. L. Gaeta, and M. Lipson, “Battery-operated integrated frequency comb generator,” Nature, vol. 562, no. 7727, pp. 401–405, 2018, https://doi.org/10.1038/s41586-018-0598-9.Suche in Google Scholar PubMed
[53] J. Feldmann, N. Youngblood, C. D. Wright, H. Bhaskaran, and W. H. P. Pernice, “All-optical spiking neurosynaptic networks with self-learning capabilities,” Nature, vol. 569, no. 7755, pp. 208–214, 2019, https://doi.org/10.1038/s41586-019-1157-8.Suche in Google Scholar PubMed PubMed Central
[54] X. Li, N. Youngblood, C. Ríos, et al.., “Fast and reliable storage using a 5 bit, nonvolatile photonic memory cell,” Optica, vol. 6, no. 1, pp. 1–6, 2019, https://doi.org/10.1364/optica.6.000001.Suche in Google Scholar
[55] Y. Shi, J. Ren, G. Chen, et al.., “Nonlinear germanium-silicon photodiode for activation and monitoring in photonic neuromorphic networks,” Nat. Commun., vol. 13, no. 1, p. 6048, 2022, https://doi.org/10.1038/s41467-022-33877-7.Suche in Google Scholar PubMed PubMed Central
Supplementary Material
This article contains supplementary material (https://doi.org/10.1515/nanoph-2023-0298).
© 2023 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.
Artikel in diesem Heft
- Frontmatter
- Review
- Review: tunable nanophotonic metastructures
- Research Articles
- Diffusion probabilistic model based accurate and high-degree-of-freedom metasurface inverse design
- Human emotion recognition with a microcomb-enabled integrated optical neural network
- Plasmonic sensing using Babinet’s principle
- Drift current-induced tunable near-field energy transfer between twist magnetic Weyl semimetals and graphene
- Manipulation on radiation angles via spatially organized multipoles with vertical split-ring resonators
- Multi-faceted plasmonic nanocavities
- Broadband large-angle beam scanning with dynamic spin energy distribution based on liquid crystal cascaded bilayer metasurface
- Vector vortex beams sorting of 120 modes in visible spectrum
- Wavelength multiplexing infrared metasurfaces for protein recognition and trace detection
Artikel in diesem Heft
- Frontmatter
- Review
- Review: tunable nanophotonic metastructures
- Research Articles
- Diffusion probabilistic model based accurate and high-degree-of-freedom metasurface inverse design
- Human emotion recognition with a microcomb-enabled integrated optical neural network
- Plasmonic sensing using Babinet’s principle
- Drift current-induced tunable near-field energy transfer between twist magnetic Weyl semimetals and graphene
- Manipulation on radiation angles via spatially organized multipoles with vertical split-ring resonators
- Multi-faceted plasmonic nanocavities
- Broadband large-angle beam scanning with dynamic spin energy distribution based on liquid crystal cascaded bilayer metasurface
- Vector vortex beams sorting of 120 modes in visible spectrum
- Wavelength multiplexing infrared metasurfaces for protein recognition and trace detection