Abstract
The intrinsic complexity of nonlinear optical phenomena offers a fundamentally new resource to analog brain-inspired computing, with the potential to address the pressing energy requirements of artificial intelligence. We introduce and investigate the concept of nonlinear inference capacity in optical neuromorphic computing in highly nonlinear fiber-based optical Extreme Learning Machines. We demonstrate that this capacity scales with nonlinearity to the point where it surpasses the performance of a deep neural network model with five hidden layers on a scalable nonlinear classification benchmark. By comparing normal and anomalous dispersion fibers under various operating conditions and against digital classifiers, we observe a direct correlation between the system’s nonlinear dynamics and its classification performance. Our findings suggest that image recognition tasks, such as MNIST, are incomplete in showcasing deep computing capabilities in analog hardware. Our approach provides a framework for evaluating and comparing computational capabilities, particularly their ability to emulate deep networks, across different physical and digital platforms, paving the way for a more generalized set of benchmarks for unconventional, physics-inspired computing architectures.
1 Introduction
The rapid advancement of artificial intelligence has sparked renewed interest in brain-inspired hardware, particularly optical implementations that promise energy-efficient solutions for AI acceleration and intelligent edge sensing. Traditional computing architectures face significant challenges when executing analog neural networks, resulting in substantial energy, water, and computational time requirements for operating large networks on conventional digital processors (i.e., CPUs, GPUs, TPUs). Optical approaches have garnered particular attention due to their intrinsic parallelism and scalability across multiple optical degrees of freedom, offering reduced energy consumption [1].
Unlike electronic solutions, a primary challenge in realizing competitive neuromorphic optical hardware lies in the all-optical implementation of nonlinearity to circumvent the electro-optical bottleneck toward deep all-optical architectures [2], [3].
Nonlinearity is crucial for emulating synaptic switching behavior in information networks and enabling advanced learning capabilities, including improved accuracy and generalization. In deep networks, this issue is often addressed through multiple layers, which offer higher nonlinear mapping capabilities of the model. The concept of “nonlinear mapping capability” refers to a system’s ability to transform input data into higher-dimensional spaces where previously inseparable patterns become linearly separable. Moving to hardware, physical substrates can inherently provide complex nonlinear responses through their natural dynamics without the need for additional layers or computational resources.
In optics, taking advantage of the nonlinearity offered by light–matter interactions in the media, while actively studied [4], is widely assumed to be complicated and inefficient due to its high-power demands [5].
Several recent studies suggest alternatives to realize nontrivial nonlinearity in optical systems. These include electronic feedback loops to electro-optic modulators [6], [7], saturable absorption [8] complex active gain dynamics in mode-competitive cavities [9], and, most recently, repeated linear encoding through multiple scattering of input data [10], [11].
Neuromorphic computing with wave dynamics offers a promising approach to the elegant utilization of natural nonlinear dynamics in physical substrates. In optics, the concept aligns with Extreme Learning Machines (ELM) – a form of reservoir computing [12], [13] without internal recurrence. Unlike other optical implementations, the learning machine comprises multiple virtual nodes that are coupled by the intrinsic feed-forward propagation dynamics of a single optical component, such as a fiber. This concept, recently proposed by Marcucci et al. [14], utilizes process-intrinsic, nonlinear mode interactions from natural waveguide dynamics as a computing resource. Experimental demonstrations include nonlinearly coupling spatial modes in multimode fibers [15] and spectral frequency generation in single-mode fibers [16], [17], [18].
However, analog wave computers have a common difficulty: the systems are generally complex and cannot be easily mapped to practical computing models, a critical challenge lies in the diversity of existing approaches and the general inability to measure a system’s nonlinear mapping capabilities independent of specific tasks and to correlate them with learning abilities. This transformation is fundamental to solving complex tasks. Current approaches in the optical community empirically test new nonlinear systems on this capability through task-specific image recognition benchmarks. Here, it is common practice to compare, e.g., accuracies across system configurations to demonstrate improvements and to relate those improvements to the scaling in the learning behavior of deeper, hence more nonlinear neural networks computer models.
In particular, the MNIST dataset is frequently used as a benchmark. Yet, it requires only low nonlinearity to separate the ten classes, as a logistic regression can achieve approximately 92 % accuracy [19]. Similarly, linear Support Vector Machines (SVMs), which leverage kernel tricks to project input data into higher-dimensional spaces for linear separation, demonstrate comparable performance. Furthermore, performance improvements, even if using big networks, cannot be clearly attributed to higher nonlinearity given by the number of hidden layers (i.e., deepness), but might also result from increased connectivity, as we will discuss further down in this paper. The question of nonlinearity’s relevance in neural networks also remains largely unexplored in computer science, with only few investigations into nonbinary nonlinear activation functions, such as a multistep perceptron [20] or trainable splines [21].
This work attempts to better understand and quantify the nonlinear inference capacity using a scalable, task-independent dataset and validate it using two different optical fiber processors. This investigation is particularly relevant for comparing computational capabilities across different physical substrates. We begin with illustrating the operation principle of a frequency-based ELM in a single optical fiber, as an arbitrary kernel machine; we refer to it from this chapter onward as fiber-optical ELM. We then assess the inference capabilities of two different nonlinear systems distinguished by their dispersion properties and corresponding nonlinear dynamics (i.e., self-phase modulation and soliton fission).
We assess the nonlinear inference capacity of the two different fiber-optical machines using a scalable spiral dataset, where increasing the angular span progressively challenges separability. These comparisons were conducted under different system configurations, including variations in fiber types, different numbers of system read-outs, and different power levels. Observations on datasets are compared to digital classifier models, like neural networks and support vector machines, providing a strong basis for drawing parallels between the depth of digital networks and the nonlinear inference capacity of our analog computing systems.
2 Materials and methods
2.1 Experimental setup
We conducted experiments using a fiber-optical extreme learning machine in the frequency domain. The experimental system (see Figure 1d) comprises a femtosecond laser source (Toptica DFC) operating at 80 MHz repetition rate with 70 nm bandwidth centered at 1,560 nm, followed by an erbium-doped fiber amplifier (Thorlabs EDFA300P) operated at 25 % pump current to compensate for system losses. Data are encoded using a polarization-maintaining programmable spectral filter (Coherent Waveshaper 1000A/X) operating in the extended C- to L-band (1,528–1,602 nm), which allows to independently modify spectral phase and amplitude across 400 frequency channels for both dispersion compensation and phase information imprinting. Placing the Waveshaper after the EDFA enables precise controlling of the attenuation, and, thus, the input power to the nonlinear fiber circumventing gain saturation effects and nonlinear deterioration of the encoded phase information before being processed. The initially chirped pulses are precompressed through a dispersion-compensating fiber (Thorlabs PMDCF, 1 m length) to 400 fs before entering the spectral filter. A particle swarm optimization algorithm in conjunction with optical autocorrelation measurements has been utilized to further optimize the input pulse phase to obtain 135 fs [22], [23] before entering the main processing unit – a highly nonlinear fiber. Two nonlinear fibers were in use: (1) Thorlabs HN1550, 5 m length with all-normal dispersion (−1 ps/nm/km @1,550 nm), and 10.8 W−1km−1 nonlinear coefficient, and (2) Thorlabs PMHN1, 5 m length, with anomalous dispersion (+1 ps/nm/km @1,550 nm), and 10.7 W−1km−1 nonlinear coefficient. The output is measured using an optical spectral analyzer (Yokogawa AQ6375E) and analyzed on a standard office computer. All components are fiber-coupled and polarization-maintaining, ensuring stable operation in a compact footprint.

Illustration of the data flow in the fiber-based neuromorphic system using an example from the spiral dataset. (a) Input data: four sample points were selected from four different spirals. (b) Corresponding data encoding: the first half of the spectral encoding range (limited by the WaveShaper) encodes the X 1-coordinate of a data tuple; the second half encodes the X 2-coordinate. (c) Encoding phase after multiplying it with a constant phase scale factor and an arbitrary but fixed mask. (d) The experimental setup used for processing. A computer is used as I/O device and is not part of a feedback loop. (e) Linear spectral intensities at fiber output corresponding to the four sample inputs. (f) Linear spectral intensities at the selected, optimized search bins serving as system read-outs. (g) Prediction scores obtained by multiplying the read-outs with the trained weight matrix. Per sample, scores are sorted from class 1 to 4, from top to bottom. The highest values in a vector of four (i.e., argmax(Y score)) determine the predicted class. (h) Prediction results: points represent the predictions, while circles around these points indicate the true class labels, the examples contain one misclassification indicated by the red cross.
We illustrate how the fiber-optical ELM processes information with the example of a four-arm spiral dataset. For the spiral classification benchmark, the inputs consist solely of two coordinates for each data point, with each coordinate tuple assigned to one point of the four spirals.
Further details about the dataset are provided in the results section. The coordinate tuples are encoded into the spectral phase of a femtosecond pulse using a encoded signal, where the first half of the encoded signal corresponds to the X 1-coordinate and the second half corresponds to the X 2-coordinate, as illustrated in Figure 1b. To enhance the system’s sensitivity to the encoded inputs, the encoded signal is multiplied by a random mask, shown in Figure 1c. This approach is common for optical reservoir computers [24], [25], [26], [27], as it leverages the system’s sensitivity to phase jumps rather than absolute phase, ensuring a maximized system response. As a result, the output signals corresponding to different encodings become more distinguishable.
2.2 Supervised offline training
Following the principles of extreme learning machines [20], we train only the output layer. This method leverages the fiber-optical ELM’s inherent ability to perform complex, high-dimensional transformations efficiently, reducing the computational burden associated with traditional training. Hence, to translate the output spectra of the system into an interpretable inference, we train a weight matrix based on selecting a subset from all wavelength recordings. From this chapter onward, we call the locations of the wavelength windows “search bins” and their corresponding intensities “read-outs.” We emphasize that only a subset of the spectral outputs is used to avoid overfitting (cf. Figure 3c, ND case from 50 bins and above). Also, using only a low number of “read-outs” is taking a further step toward an all-optical neural network, as considering fewer outputs eases the practical implementation of an optical weight bank.
We select a user-defined number of search bins n using an optimization algorithm called Equal Search, a straightforward method that identifies combinations of equally spaced search bins, as described in ref. [17]. From the resulting combinations, we choose the one that minimizes the mean square error (
where X is the read-outs vectors (intensities at selected search bins of dimension R n × R k with k samples), and Y is the task’s label vectors (ground truth of dimension R m × R k ). We experimented with incorporating the ridge term while training the system; however, we found it consistently yielded worse results (e.g., see long-term measurements in Figure 2A in the Supplementary Material). We assume that sampling the output bins in combination with our Equal Search keeps model complexity low and effectively reduces feature collinearity in the covariance matrix while maintaining only the most informative spectral features, rendering regularization unnecessary and even counterproductive.
After training the weight matrix, classification is performed by acquiring the output values at the best combination of search bins identified before. These values are stored in a vector and multiplied by the weight matrix, producing another vector, referred to as the prediction score Y score = WX. A winner-takes-all decision, i.e., Y class = argmax(Y score) is then performed on the Prediction Score vector, classifying the test data based on the index of the highest score, as shown in Figure 1g and h. We compare systems based on task-specific classification accuracy in percent (%), which is defined by the ratio of number of correct class predictions over the total number of samples.
Figure 2 illustrates the nonlinear projection capabilities of our fiber-optical neural network through measured spectral response to input data tuples (X 1, X 2) sampled from four distinct classes in a two-dimensional coordinate system. The Equal Search algorithm identified four optimal search bins (Figure 2a–d) demonstrating highly distinct separation between the data classes in the spectral domain. Particularly noteworthy are the projections shown in Figure 2b and c, where the nonlinear fiber transformation achieves nearly ideal plane-wise separation of the classes in three dimensions, providing clear evidence of the system’s powerful separation capabilities. In contrast, Figure 2e and f presents randomly selected wavelength windows where the class separation is poor or nonexistent, with data points from different classes showing significant overlap.

3D plots illustrating the relationship between output spectral intensity of a supercontinuum from an anomalous dispersive fiber and the input coordinates for all given samples. (a–d) Logarithmic spectral intensity versus input coordinates (X1, X2) at the optimized, selected search bins, demonstrating the system’s intrinsically distinguishable response to different classes. (e–f) Spectral intensity versus input coordinates at two randomly selected wavelength windows. All data shown are >10 dB above the spectrometer’s noise floor.
This behavior effectively showcases the fundamental principle proposed by Huang et al. in their original ELM framework: ELMs can be understood as arbitrary kernel machines that, in optimal cases, provide hidden nodes with kernel response functions capable of transforming complex problems into linearly separable ones [20]. Following this interpretation, the performance of an ELM improves with access to a greater variety of kernel functions – a key insight that leads to our hypothesis that physical systems exhibiting richer nonlinear dynamics should outperform those with weaker nonlinearity in classification tasks.
3 Results and discussion
3.1 Scalable nonlinear classification
In this section, we demonstrate the capability of a fiber-optical ELM to effectively handle highly nonlinear tasks, exemplified by the spiral classification benchmark. The spiral classification task has emerged as a critical benchmark in the machine learning community [28] for evaluating neural networks’ capability to handle tangled data points and to learn nonlinear decision boundaries. This two-dimensional dataset, featuring interleaved spiral arms over an angular span of 2π, is intrinsically not linearly separable and poses a particular challenge for optical computing systems due to the inherent difficulty in implementing nonlinear activation functions in the optical domain. However, recent breakthroughs have demonstrated significant progress in this area, with recent notable achievements including on-chip complex-valued neural networks reaching 95 % accuracy on a 2-arm spiral classification [29] and a three-layer optoelectronic neural network achieving 86 % accuracy on a 4-arm spiral classification [30].
Building upon these advances, we introduce an enhanced version of this benchmark that increases the hardness of the task. We generalized the spiral dataset by a free parameter, called the maximum angular span θ max, which controls the complexity of the spiral benchmark, making it particularly effective for illustrating the impact of the network depth and the system’s nonlinear dynamics on the inference capabilities of digital and optical neuromorphic systems, respectively, as later shown in Figures 5a and 6. As such, it is an excellent general measure of the nonlinear inference capacity of new neuromorphic physical substrates.
The data points are generated using analytic equations (Eq. (2)), allowing flexibility in defining the number of phase-shifted spiral arms (number of classes, controlled by index k) and their wraps around the center (controlling the complexity of the task). The number of wraps defined by a full 360° turn around the center is controlled by θ max, where the number of wraps equals θ max/2π. For the presented results, we generated 250 points per spiral, yielding 1,000 total points across four spirals. From this set, we used 800 randomly selected points (200 per spiral) to train the system and kept 200 points (50 points per spiral) for testing after training (4:1 data split ratio).
We evaluated our fiber-optical ELM on the spiral benchmark, with its increasing nonlinearity as θ max increases (e.g., 0.5π → 10π), which is an ideal task to demonstrate the fiber-optical ELM’s ability to handle highly nonlinear problems. Figure 3 illustrates the classification results for θ max = 10π, corresponding to 5 wraps around the center per spiral. Figure 3a and b shows the average and the standard deviation (STD) of the spectral intensities for both fiber types, normal dispersion (ND) fiber and anomalous dispersion (AD) fiber, respectively, measured under the same excitation conditions (i.e., 18.45 mW, 136 fs, cp. Table 1). The difference in standard deviations between the two cases arises from using different scale factors in the encoding phase for the two fibers.

4-arms spiral classification benchmark in normal and anomalous dispersion fibers. (a) Average and standard deviation (STD) of output spectral intensities for the spirals dataset in the normal dispersion (ND) case. (b) Average and STD of output spectral intensities for the spirals dataset in the anomalous dispersion (AD) case. (c) Classification test accuracy across all classes achieved for both fiber types as a function of the number of search bins. (d) Classification results for the ND case using 50 search bins. (e) Classification results for the AD case using 50 search bins.
Pulse width and peak power at the input of the highly nonlinear fiber (HNLF) for various encoding conditions (attenuation), calculated using average power and pulse width (both are measured before propagating through the fiber). The pulse width was measured using an APE autocorrelator (pulseCheck NX 50). The average of 16 scans over a range of 5 ps was fitted to a Lorentzian function to determine the pulse width.
| Encoding attenuation applied (dB) | Averag power measured (mW) | Pulse width measured (fs) | Peak power calculated (W) |
|---|---|---|---|
| 0 | 18.56 | 136 | 1,705.88 |
| 2 | 11.66 | 146 | 998.29 |
| 4 | 7.41 | 155 | 597.58 |
| 6 | 4.72 | 157.5 | 374.60 |
| 8 | 3.04 | 161 | 236.02 |
We chose the scale factors according to the highest classification accuracy obtained during our experiments with various scale factors (see Figure 4A in the Supplementary Material). Specifically, we use π/2 for the ND fiber and π/16 for the AD fiber, resulting in a larger STD for the ND case. Our Equal Search analysis allows us to compare the system’s performance for various numbers of search bins. In Figure 3c, we observe increasing the number of search bins increases system accuracy for both fibers. The AD fiber achieves superior performance on the spiral benchmark (>95 % accuracy for 35 bins and up) compared to ND fiber, which reaches the best accuracy of 91.5 % with 50 bins before going into overfitting (i.e., test accuracy diverges from training accuracy).
3.2 Handwritten digit recognition
Next, we test the system’s performance on a low nonlinear task. We identified the MNIST dataset to be suitable for this purpose as it is known to be solvable to a high degree by using linear classifiers (cf. Figure 1A). The MNIST dataset is a widely used collection of handwritten digits from 0 to 9 (i.e., 10 classes), stored as 28 × 28 pixels gray-valued images, that serves as a standard benchmark in machine learning. The dataset used in this study is sourced from the TensorFlow library (tensorflow_datasets).
We trained the system on the first 2,100 images and tested it on 300 images previously unseen by the system using both normal and anomalous dispersion fibers. To encode the MNIST images, we first crop an 18 × 18-pixel window to meet the data input limitations of the Waveshaper. The cropped images were then flattened row-wise from top to bottom into 1D signals corresponding to pixel intensities. These 1D signals were multiplied by a random mask and a phase scale factor before being added to an optimized phase profile, yielding the total encoding phase (cf. Figure 3Ad), which the Waveshaper adds on the optical pulse, hence encoding the input information, before propagating through the fiber. At the output layer, the system follows the same procedure as in the spiral benchmark. It stores the output read-outs at the newly selected search bins and translates these read-outs into an interpretable prediction score by multiplying the read-outs vector with a trained weight matrix. The highest value in the prediction score (i.e., argmax(Y score)) determines the predicted class. Figure 3A in the Supplementary Material illustrates all input and output levels for an MNIST test image.
Unlike the spiral benchmark, the system demonstrates similar performance on MNIST when using ND fiber (89.33 % accuracy) and AD fiber (87.3 % accuracy), as shown in Figure 4d and e, respectively. Both cases surpass the baseline of 83.7 % for linear regression on the data [31] and perform similarly to a support vector machine (SVM) with a linear kernel, which achieves about 91 % in accuracy with only 20 support vectors (cp. Figure 1A in the Supplementary Material). Classification accuracy further improves with an increased number of search bins, as illustrated in Figure 4c.

MNIST digit classification benchmark in normal and anomalous dispersion fibers. (a, b) Average and standard deviation (STD) of measured output spectral intensities for the MNIST dataset in the (a) normal dispersion (ND) case and (b) anomalous dispersion (AD) case. (c) Classification test accuracy achieved across 300 MNIST test samples across all classes as a function of the number of search bins for both fiber types. (d, e) Confusion matrices of our systems for unseen test data for the (d) ND case (achieved accuracy 89.33 %) and (e) AD case (achieved accuracy 87.3 %) using 150 search bins for both cases.
We hypothesize that the higher nonlinear mapping introduced by the AD fiber does not significantly affect system performance for a task with low nonlinearity, such as MNIST. The slightly better performance observed with the ND fiber in our setup can be attributed to a more extensive optimization process during the selection of the optimal encoding phase scale factor. Specifically, when using the ND fiber, we conducted multiple experiments with different phase scale factors and selected the one that achieved the highest accuracy. For both fibers, a phase scale factor of
The confusion matrices and per-class accuracy for both cases are presented in Figure 4d and e. A notable observation is the misclassification between the digit classes 3 and 5 and classes 4 and 9, which likely arises from their similar 1D intensity profiles.
To assess robustness, we conducted a long-term evaluation involving training the system on 3,000 images and testing it on 35,000 images over an acquisition period of 35 h. The system maintained an accuracy of 84.69 %, with only a modest 4.5 % drop over the extended testing period, underscoring its reliability and resilience. Further details of this robustness analysis are provided in Figure 2A in the Supplementary Material.
3.3 Scaling nonlinear inference capacity
We further investigate the impact of input power (consequently the system’s nonlinear mapping capability) on the performance of our fiber-optical ELM handling increasingly harder task, by changing the attenuation applied on the input signals (0, 2, 4, 6, 8 dB), while progressively increasing the spirals’ θ max (0.5π, 2π, 4π, 8π, 10π) and hence the complexity of the classification task. We benchmark our system against digital neural networks with different configurations. To ensure a fair comparison between optical and digital, we fed both systems with the exact same data at a fixed training–testing split ratio.
The results demonstrate the critical role of network depth in addressing highly nonlinear tasks with digital neural networks. As shown in Figure 5a, increasing the number of layers in deep networks, while keeping the total number of nodes constant, increases the classification accuracy of the spiral data for large θ max significantly.

Nonlinear inference capacity scaling behavior in digital neural networks and fiber-optical extreme learning machines in two dispersive regimes. (a) Best test accuracy on 200 spiral data samples achieved by digital classifiers (a linear kernel support vector machine with 100 support vectors, and neural networks in different configurations (cp. Table 1A); all trained for 1,000 training epochs) and our fiber-optical ELM using 100 search bins for increasing nonlinear problem hardness in the spiral task, defined by the maximum angular span θ max. (b, c) Test accuracies on 200 spiral data samples as a function of system nonlinearity (or attenuation) and maximum angular span for both, (b) normal dispersion and (c) anomalous dispersion.
The experimental results in Figure 5a highlight the fiber-optical ELM’s exceptional efficiency in handling such highly nonlinear tasks. In particular, the AD fiber can outperform a deep network with 5 hidden layers and 1,024 nodes for θ max > 10π, with the ND fiber being just a little below this performance.
Furthermore, Figure 5b illustrates the accuracy achieved using the ND fiber on the spiral benchmark with different maximum angular spans (θ max = 0.5π, 2π, 4π, 8π, 10π), under different attenuation levels (0, 2, 4, 6, 8 dB), represented as rows and columns in a colormap, respectively. Figure 5c presents the corresponding results for the AD fiber. Both cases use 50 search bins.
We notice, for both dispersion domains, that applying attenuation (decreasing the input power) has a pronounced effect on system performance for highly nonlinear tasks (e.g., spirals with high θ max) but less for low nonlinear tasks (e.g., spirals with low θ max). This performance degradation with decreasing input power levels stems from the reduced variety in nonlinear system response per frequency, which forms the nonlinear mapping capability of the systems. Increasing the number of search bins mitigates the loss in nonlinear mapping induced by attenuation, which is in accordance with previous observations on comparable problems [17].
Notably, the AD fiber shows particularly better performance at lower input power levels for more complex tasks compared to the ND fiber, though with greater fluctuations. These fluctuations are most evident in the unexpectedly high accuracy for θ max = 8π at attenuation of 6 dB (input power = 6.7 dBm). In particular, we observe that the AD domain features a significantly slower decline in the test accuracies of the spiral dataset with decreasing power. The AD fiber also maintains remarkably high classification accuracy as the spiral task complexity increases. Even at an angular span of 24π (the maximum tested in our experiments), the AD fiber system still achieves 87.5 % accuracy despite the task becoming undersampled (see Figure 5A).
We attribute the better performance of the AD fiber to the larger variety of nonlinear responses per frequency, which form the pool of functions the bin search algorithm can sample to build the nonlinear decision boundary. The richer complexity of nonlinear transformations sets the soliton fission broadening process apart from the more regulated and symmetric broadening behavior of self-phase modulation. The sensitivity of the fission process also leads to a larger variance in the test results in Figure 5c and, hence, to a less consistent trend in the decline of accuracy with system nonlinearity. Bin search techniques that allow for a more flexible positioning of the bins may reduce this effect.
It becomes apparent that the system performance depends not only on the pulse power but on the system nonlinearity in general. The system nonlinearity can be generalized by the soliton number:
γ is the nonlinear coefficient (W−1km−1), P 0 is the peak power, T 0 is the measured pulse width (seconds), β 2 is the dispersion coefficient (s2/m), and f rep is the laser repetition rate in (Hz). Table 1 shows the power and autocorrelation measurements for the presented results.
The soliton number corresponds to a normalized nonlinear gain parameter of a waveguide system, which is proportional with peak power P 0 and, hence, inversely to the attenuation level. We emphasize parametrizing the system’s nonlinear mapping capability using the soliton number rather than input power alone, as it reflects both the peak power and the pulse width of the input signal, hence the signal’s coherence, which is essential for supercontinuum generation [32].
Increasing the optical system’s nonlinearity (indicated by the soliton number) in an optical neuromorphic system has a similar impact on the achieved results as increasing the depth of a digital neural network. This can be seen in the similarity of the performance trends between an optical system and the digital model on a highly nonlinear task (e.g., spiral benchmark with θ max = 10π) and a low nonlinear task represented by MNIST in Figure 6.

Performance comparison of optical-fiber ELM and digital neural networks handling high and low nonlinear tasks. (a) Performance trends of the fiber-optical ELM on a highly nonlinear task (spiral benchmark) and a low nonlinear task (MNIST) under varying attenuation levels. (b) Performance trends of a multilayer neural network with different numbers of hidden layers but the same total number of nodes (400) evaluated on the spiral benchmark and MNIST datasets.
Figure 6a presents the test accuracy achieved for MNIST and the spiral benchmark (with θ max = 10π) at various attenuation levels. The results are shown for the ND fiber trained using 150 search bins (solid lines) and 50 search bins (dotted lines). Increasing the system’s nonlinearity does not enhance its performance on the MNIST dataset. In contrast, higher nonlinearity significantly improves performance on the spiral benchmark with a high angular span (θ max = 10π).
Notably, a similar trend is evident when solving these benchmarks using digital neural networks. Figure 6b illustrates the achieved test accuracy for deep neural networks for an increasing number of hidden layers, yet at a constant total number of nodes (400). While MNIST is solvable with 10 nodes (cf. Figure 1Aa in the Supplementary Material), a significantly higher number of nodes is needed to solve the spiral benchmark. The 400 nodes have been distributed across the hidden layers as shown in Table 2A.
The model results in Figure 6b show, as expected, that increasing the network depth has a negligible impact on the test accuracy achieved for MNIST. Here, more layers only help in convergence, enabling the model to stabilize with fewer training epochs. In contrast, consistent with the earlier results shown in Figure 5a, deeper networks provide a significant advantage when handling highly nonlinear tasks, such as the spiral benchmark with θ max = 10π. Specifically, 4- and 5-hidden-layer networks with 400 nodes (see Figure 6b) outperform 2- and 3-hidden-layer networks with 1,024 nodes (see Figure 5a), highlighting the importance of network depth before connectivity when handling complex nonlinear tasks.
4 Conclusions
Our results suggest that while the performance of fiber-optical ELMs stalls on the MNIST dataset, they excel in handling highly nonlinear tasks. The comparable scaling behavior in classification performance between fiber ELMs regarding system nonlinearity (Figure 6a) and digital neural networks regarding the number of layers (Figure 6b) indicates that nonlinear fibers can mimic depth in neuro-inspired computing. This highlights their suitability for specialized applications where conventional optical neural networks, which are often shallow (i.e., single-layer), face limitations.
To date, the performance gain in optical neural networks is traditionally shown in improvements in image classification results across various datasets, including MNIST, fashion-MNIST, ImageNet. While these datasets are widely used, they fail to isolate the performance gains achieved through increased nonlinearity from those resulting from increased network connectivity. We demonstrated that measuring the system’s nonlinear inference capacity through scalable nonlinear benchmarks, such as the spiral dataset, provides a less ambiguous way to compare the computational depth of unconventional neuromorphic hardware. The spiral benchmark’s key strength lies in its parametric nature, allowing controlled increments in task complexity that directly test a system’s nonlinear inference capacity. This approach has proven especially valuable in two critical aspects: First, it enables quantitative comparison between different analog computing substrates, whose computational operations are often difficult to enumerate in terms of the number of performed computations. Second, it establishes a direct bridge between physical implementations and digital models, allowing us to correlate the scaling of physical system dynamics with the depth of digital neural networks. These findings may pave the way for a framework for developing standardized benchmarks that can effectively evaluate both conventional and physics-inspired computing architectures, particularly in their ability to handle increasingly complex nonlinear tasks.
We furthermore observe a direct separation of the spiral classes in selected output channels of our systems, which uniquely showcase nonlinear Schrödinger systems as genuinely arbitrary kernel machines, providing a unique substrate for implementing the ELM computing framework. The soliton number serves as a valuable, platform-independent measure of the nonlinearity as it provides a generalizable nonlinear gain parameter for wave systems that can be reduced to a nonlinear Schrödinger equation of the form
with complex-valued field a, and normalized space Z and time T. This may include mathematically similar systems such as the Korteweg–de Vries (KdV) equation (e.g., reservoirs based on shallow-water waves [33]) or the Fitzhugh–Nagumo (FN) equation used to describe membrane dynamics in nerve cells [34]. By incorporating this parameter, we gain a more nuanced understanding of the system’s behavior and its ability to process complex inputs effectively. Therefore, we hypothesize that it may provide a standardized metric for defining the nonlinear state across physical systems and comparing the nonlinear inference capacity. Nonetheless, it is important to acknowledge that it is not a universally predictive measure of a system’s computational capabilities, since those capabilities can still vary from system to system (e.g., from anomalous to normal dispersion) at the same soliton number, depending on the underlying dynamics.
Through our work, we hope to encourage ongoing exploration of generalizable performance metrics, as the connection between data-driven inference performance and the measures of intricate physical dynamics may prove vital in uncovering the genuine computational performance of natural systems. Information measures, such as Shannon Entropy [35], Fischer Information [36], or Information Processing Capacity [37], may provide a profound start, yet their application to analog feedforward ELMs remains unclear.
Our findings collectively highlight the promise of the fiber-optical Schrödinger system as a scalable, efficient solution for neuromorphic computing. Their capability to handle highly nonlinear tasks with a compact and stable design makes them a strong candidate for advancing future computing systems, in particular in a more open approach to unconventional neuromorphic computing frameworks that go beyond the perceptron architecture.
Funding source: Carl-Zeiss-Stiftung
Award Identifier / Grant number: P2021-05-025
Acknowledgments
We acknowledge that this work was made possible by funding from the Carl Zeiss Foundation through the NEXUS program (project P2021-05-025).
-
Research funding: Carl-Zeiss Stiftung, Nexus program (P2021-05-025).
-
Author contributions: SS did the data acquisition, data analysis, modeling, and coding and contributed to writing, proof reading, figure editing, and methods development. MC led the project, supervised the data acquisition, data analysis, modeling, and coding, and contributed to writing, proof reading, figure editing, and methods development. MC and BF designed the setup. BF also contributed to optimizing the experiments, coding, proof reading, methods development, and figure editing. MM contributed to optimizing the experiments, coding, and proof reading. GRC contributed to data analysis, coding, and proof reading. TB contributed to modeling and proof reading. All authors have accepted responsibility for the entire content of this manuscript and consented to its submission to the journal, reviewed all the results and approved the final version of the manuscript.
-
Conflict of interest: The authors declare no conflict of interest regarding the publication of this paper.
-
Data availability: The raw spiral benchmark data used are available on our Github page. https://github.com/Smart-Photonics-IPHT/09_Create-Data-Set_Create-Spiral-Data.
References
[1] S. Shekhar, et al.., “Roadmapping the next generation of silicon photonics,” Nat. Commun., vol. 15, no. 1, p. 751, 2024, https://doi.org/10.1038/s41467-024-44750-0.Search in Google Scholar PubMed PubMed Central
[2] B. J. Shastri, A. N. Tait, M. A. Nahmias, B. Wu, and P. R. Prucnal, “Spatiotemporal pattern recognition with cascadable graphene excitable lasers,” in 2014 IEEE Photonics Conference, IEEE, 2014, pp. 573–574.10.1109/IPCon.2014.6995270Search in Google Scholar
[3] S. Bandyopadhyay, et al.., “Single-chip photonic deep neural network with forward-only training,” Nat. Photonics, vol. 18, no. 12, pp. 1335–1343, 2024, https://doi.org/10.1038/s41566-024-01567-z.Search in Google Scholar
[4] O. Destras, S. Le Beux, F. G. De Magalhães, and G. Nicolescu, “Survey on activation functions for optical neural networks,” ACM Comput. Surv., vol. 56, no. 2, pp. 1–30, 2023. https://doi.org/10.1145/3607533.Search in Google Scholar
[5] D. Marković, A. Mizrahi, D. Querlioz, and J. Grollier, “Physics for neuromorphic computing,” Nat. Rev. Phys., vol. 2, no. 9, pp. 499–510, 2020, https://doi.org/10.1038/s42254-020-0208-2.Search in Google Scholar
[6] S. Ortín, et al.., “A unified framework for reservoir computing and extreme learning machines based on a single time-delayed neuron,” Sci. Rep., vol. 5, no. 1, p. 14945, 2015, https://doi.org/10.1038/srep14945.Search in Google Scholar PubMed PubMed Central
[7] F. Duport, B. Schneider, A. Smerieri, M. Haelterman, and S. Massar, “All-optical reservoir computing,” Opt. Express, vol. 20, no. 20, pp. 22783–22795, 2012, https://doi.org/10.1364/OE.20.022783.Search in Google Scholar PubMed
[8] K. Gupta, Y.-R. Lee, Y.-C. Cho, and W. Choi, “Multilayer optical neural network using saturable absorber for nonlinearity,” Opt. Commun., vol. 577, p. 131471, 2025, https://doi.org/10.1016/j.optcom.2024.131471.Search in Google Scholar
[9] A. Skalli, et al.., “Photonic neuromorphic computing using vertical cavity semiconductor lasers,” Opt. Mater. Express, vol. 12, no. 6, p. 2395, 2022, https://doi.org/10.1364/OME.450926.Search in Google Scholar
[10] C. C. Wanjura and F. Marquardt, “Fully nonlinear neuromorphic computing with linear wave scattering,” Nat. Phys., vol. 20, no. 9, pp. 1434–1440, 2024, https://doi.org/10.1038/s41567-024-02534-9.Search in Google Scholar
[11] F. Xia, et al.., “Nonlinear optical encoding enabled by recurrent linear scattering,” Nat. Photonics, pp. 1–9, 2024.Search in Google Scholar
[12] G. Tanaka, et al.., “Recent advances in physical reservoir computing: a review,” Neural Netw., vol. 115, pp. 100–123, 2019, https://doi.org/10.1016/j.neunet.2019.03.005.Search in Google Scholar PubMed
[13] G. Van Der Sande, D. Brunner, and M. C. Soriano, “Advances in photonic reservoir computing,” Nanophotonics, vol. 6, no. 3, pp. 561–576, 2017. https://doi.org/10.1515/nanoph-2016-0132.Search in Google Scholar
[14] G. Marcucci, D. Pierangeli, and C. Conti, “Theory of neuromorphic computing by waves: machine learning by rogue waves, dispersive shocks, and solitons,” Phys. Rev. Lett., vol. 125, no. 9, p. 093901, 2020, https://doi.org/10.1103/PhysRevLett.125.093901.Search in Google Scholar PubMed
[15] U. Teğin, M. Yıldırım, l̇. Oğuz, C. Moser, and D. Psaltis, “Scalable optical learning operator,” Nat. Comput. Sci., vol. 1, no. 8, pp. 542–549, 2021, https://doi.org/10.1038/s43588-021-00112-0.Search in Google Scholar PubMed
[16] T. Zhou, F. Scalzo, and B. Jalali, “Nonlinear Schrödinger kernel for hardware acceleration of machine learning,” J. Lightwave Technol., vol. 40, no. 5, pp. 1308–1319, 2022, https://doi.org/10.1109/JLT.2022.3146131.Search in Google Scholar
[17] B. Fischer, et al.., “Neuromorphic computing via fission-based broadband frequency generation,” Adv. Sci., vol. 10, no. 35, p. 2303835, 2023, https://doi.org/10.1002/advs.202303835.Search in Google Scholar PubMed PubMed Central
[18] M. Zajnulina, A. Lupo, and S. Massar, “Weak kerr nonlinearity boosts the performance of frequency-multiplexed photonic extreme learning machines: a multifaceted approach,” Opt. Express, vol. 33, no. 4, pp. 7601–7619, 2025, https://doi.org/10.1364/OE.503279.Search in Google Scholar
[19] A. Palvanov and Y. I. Cho, “Comparisons of deep learning algorithms for MNIST in real-time environment,” Int. J. Fuzzy Log. Intell. Syst.vol. 18, no. 2, pp. 126–134, 2018, https://doi.org/10.5391/IJFIS.2018.18.2.126.Search in Google Scholar
[20] G.-B. Huang, Q.-Y. Zhu, and C.-K. Siew, “Extreme learning machine: theory and applications,” Neurocomputing, vol. 70, no. 1, pp. 489–501, 2006, https://doi.org/10.1016/j.neucom.2005.12.126.Search in Google Scholar
[21] X. Gao, et al.., “Learning continuous piecewise non-linear activation functions for deep neural networks,” in 2023 IEEE International Conference on Multimedia and Expo (ICME), 2023, pp. 1835–1840.10.1109/ICME55011.2023.00315Search in Google Scholar
[22] A. Efimov, M. D. Moores, N. M. Beach, J. L. Krause, and D. H. Reitze, “Adaptive control of pulse phase in a chirped-pulse amplifier,” Opt. Lett., vol. 23, no. 24, pp. 1915–1917, 1998, https://doi.org/10.1364/OL.23.001915.Search in Google Scholar PubMed
[23] B. Fischer, M. Müftüoglu, and M. Chemnitz, “Machine-aided near-transform-limited pulse compression in fully fiber-interconnected systems for efficient spectral broadening,” in Nonlinear Optics and its Applications 2024, vol. 13004, J. M. Dudley, A. C. Peacock, B. Stiller, and G. Tissoni, Eds., International Society for Optics and Photonics, SPIE, 2024, p. 1300402.10.1117/12.3022291Search in Google Scholar
[24] L. Appeltant, et al.., “Information processing using a single dynamical node as complex system,” Nat. Commun., vol. 2, no. 1, p. 468, 2011, https://doi.org/10.1038/ncomms1476.Search in Google Scholar PubMed PubMed Central
[25] L. Larger, et al.., “Photonic information processing beyond turing: an optoelectronic implementation of reservoir computing,” Opt. Express, vol. 20, no. 3, p. 3241, 2012, https://doi.org/10.1364/OE.20.003241.Search in Google Scholar PubMed
[26] Y. Paquot, et al.., “Optoelectronic reservoir computing,” Sci. Rep., vol. 2, no. 1, p. 287, 2012, https://doi.org/10.1038/srep00287.Search in Google Scholar PubMed PubMed Central
[27] R. Martinenghi, S. Rybalko, M. Jacquot, Y. K. Chembo, and L. Larger, “Photonic nonlinear transient computing with multiple-delay wavelength dynamics,” Phys. Rev. Lett., vol. 108, no. 24, p. 244101, 2012, https://doi.org/10.1103/PhysRevLett.108.244101.Search in Google Scholar PubMed
[28] H. Chang and D.-Y. Yeung, “Robust path-based spectral clustering,” Pattern Recogn., vol. 41, no. 1, pp. 191–203, 2008, https://doi.org/10.1016/j.patcog.2007.04.010.Search in Google Scholar
[29] H. Zhang, et al.., “An optical neural chip for implementing complex-valued neural network,” Nat. Commun., vol. 12, no. 1, p. 457, 2021, https://doi.org/10.1038/s41467-020-20719-7.Search in Google Scholar PubMed PubMed Central
[30] A. Song, S. Nikhilesh Murty Kottapalli, R. Goyal, B. Schölkopf, and P. Fischer, “Low-power scalable multilayer optoelectronic neural networks enabled with incoherent light,” Nat. Commun., vol. 15, no. 1, p. 10692, 2024, https://doi.org/10.1038/s41467-024-55139-4.Search in Google Scholar PubMed PubMed Central
[31] M. Hary, D. Brunner, L. Leybov, P. Ryczkowski, J. M. Dudley, and G. Genty, “Principles and metrics of extreme learning machines using a highly nonlinear fiber,” 2025. http://arxiv.org/abs/2501.05233.Search in Google Scholar
[32] D. Castelló-Lurbe, et al.., “Measurement of the soliton number in guiding media through continuum generation,” Opt. Lett., vol. 45, no. 16, pp. 4432–4435, 2020, https://doi.org/10.1364/OL.399382.Search in Google Scholar PubMed
[33] G. Marcucci, P. Caramazza, and S. Shrivastava, “A new paradigm of reservoir computing exploiting hydrodynamics,” Phys. Fluids, vol. 35, no. 7, p. 071703, 2023, https://doi.org/10.1063/5.0157919.Search in Google Scholar
[34] A. C. Cevikel, A. Bekir, and O. Guner, “Exploration of new solitons solutions for the fitzhugh–nagumo-type equations with conformable derivatives,” Int. J. Mod. Phys. B, vol. 37, no. 23, p. 2350224, 2023, https://doi.org/10.1142/S0217979223502247.Search in Google Scholar
[35] M. Zajnulina, “Shannon entropy helps optimize the performance of a frequency-multiplexed extreme learning machine,” 2025. http://arxiv.org/abs/2408.07108.10.1364/opticaopen.26566942.v1Search in Google Scholar
[36] S. Rotter, “The concept of Fisher information in scattering problems and neural networks,” in Nonlinear Optics and its Applications 2024, vol. PC13004, SPIE, 2024, p. PC1300401.10.1117/12.3022535Search in Google Scholar
[37] J. Dambre, D. Verstraeten, B. Schrauwen, and S. Massar, “Information processing capacity of dynamical systems,” Sci. Rep., vol. 2, p. 514, 2012, https://doi.org/10.1038/srep00514.Search in Google Scholar PubMed PubMed Central
Supplementary Material
This article contains supplementary material (https://doi.org/10.1515/nanoph-2025-0045).
© 2025 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.
Articles in the same Issue
- Frontmatter
- Editorial
- Guided nonlinear optics for information processing
- Research Articles
- All-optical nonlinear activation function based on stimulated Brillouin scattering
- Photonic neural networks at the edge of spatiotemporal chaos in multimode fibers
- Principles and metrics of extreme learning machines using a highly nonlinear fiber
- Nonlinear inference capacity of fiber-optical extreme learning machines
- Optical neuromorphic computing via temporal up-sampling and trainable encoding on a telecom device platform
- In-situ training in programmable photonic frequency circuits
- Training hybrid neural networks with multimode optical nonlinearities using digital twins
- Reliable, efficient, and scalable photonic inverse design empowered by physics-inspired deep learning
- Intermodal all-optical pulse switching and frequency conversion using temporal reflection and refraction in multimode fibers
- Modulation instability control via evolutionarily optimized optical seeding
- Review
- From signal processing of telecommunication signals to high pulse energy lasers: the Mamyshev regenerator case
Articles in the same Issue
- Frontmatter
- Editorial
- Guided nonlinear optics for information processing
- Research Articles
- All-optical nonlinear activation function based on stimulated Brillouin scattering
- Photonic neural networks at the edge of spatiotemporal chaos in multimode fibers
- Principles and metrics of extreme learning machines using a highly nonlinear fiber
- Nonlinear inference capacity of fiber-optical extreme learning machines
- Optical neuromorphic computing via temporal up-sampling and trainable encoding on a telecom device platform
- In-situ training in programmable photonic frequency circuits
- Training hybrid neural networks with multimode optical nonlinearities using digital twins
- Reliable, efficient, and scalable photonic inverse design empowered by physics-inspired deep learning
- Intermodal all-optical pulse switching and frequency conversion using temporal reflection and refraction in multimode fibers
- Modulation instability control via evolutionarily optimized optical seeding
- Review
- From signal processing of telecommunication signals to high pulse energy lasers: the Mamyshev regenerator case