Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Virtual reservoir computer using an optical resonator

Open Access Open Access

Abstract

Reservoir computing is a machine learning approach that enables us to use recurrent neural networks without involving the complexity of training algorithms and make hardware implementation possible. We present a novel photonic architecture of a reservoir computer that employs a nonlinear node and a resonator to implement a virtual recurrent neural network. This resonator behaves as an echo generator component that substitutes the delay line in delay-based reservoir computers available in the literature. The virtual neural network formed in our implementation is fundamentally different from the delay-based reservoir computers. Different virtual architectures based on the FSR and the Finesse of the resonator are investigated to provide higher performance depending on the task. We test the performance of our proposed reservoir by 3 benchmark tasks, signal classification, nonlinear channel equalization, and memory capacity. Our system enhances the memory capacity of the reservoir compared to the delay-based systems and shows reliable results in signal classification and nonlinear channel equalization.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

Artificial neural networks (ANNs) are inspired from the brain’s operation and are able to process highly complex computational tasks such as speech recognition, image recognition, and time series prediction which are difficult and time-consuming for von Neumann computers. The two main categories of neural networks are feedforward and recurrent neural networks (RNNs). In feedforward ANNs, connections (equivalent to axons and synapses) between neurons steer the signal from input to output in one direction, while in RNNs, feedback connections are also present. Due to the complexity of training RNNs (finding the correct connection weights between neurons) [1], reservoir computers (RCs) have been introduced in recent years as an attractive solution [2,3]. A reservoir computer consists of three layers: the input layer, the reservoir, and the output layer. The reservoir itself is typically a fixed RNN which does not need training. The enticing feature is that only the output layer is trained, thus the reservoir computing approach has become particularly attractive for hardware realization of RNNs as most of the connections of neurons are fixed in the design stage and does not need reconfiguration (training).

Optical implementations of RCs have gained considerable research attention in recent years and various architectures have been introduced thanks to low power consumption and crosstalk, high bandwidth, and high-speed computing in optics [47]. The fixed reservoir layer in an optical RC may be implemented with actual (physical) neurons or virtual neurons. In the physical implementation, nodes (neurons) are spatially distributed and physically connected e.g. for instance by using waveguides on an integrated chip. Different optical elements such as semiconductor optical amplifiers [8,9], ring resonators [10], passive optical structures [11], and photonic crystal cavities [12] have been used as physical nodes in these physical implementation. For high computing performance, typically a high number of nodes is required in the reservoir layer ($10^2$ to $10^3$) which complicates the integration of such realizations significantly.

To circumvent these demanding issues regarding the high number of physical nodes required in an RC, the concept of virtual nodes by means of time multiplexing was introduced in [13]. In this approach by using a delay line, virtual nodes are connected successively in time which are then fed back into the nonlinear node, yielding a particular form of RNN topology [14,15]. The schematic of the virtual reservoir using delay line is shown in Fig. 1(a). The input signal ($u(t)$) containing bits of data is multiplied by a mask signal. This mask is of higher data rate (pulses of smaller width called chips). The masked signal is the input signal of the reservoir layer which is made of a single nonlinear node and a delay line. At each chip, the input signal is combined with the delayed version of itself which has travelled through the delay line. With the proper choice of the delay, one can successfully connect the virtual nodes. The equivalent network of a virtual RC for 5 virtual nodes is shown in Fig. 1(b). Different approaches have been adopted to realize delay-based virtual reservoir computers e.g. electronic circuits [15], opto-electronic systems [13,16], all-optical systems [17,18], and laser systems [1922]. Although these delay-based implementations reduce the number of physical nodes, but unfortunately they are in need of considerable time delay whenever a large number of virtual nodes is required and thus are difficult to implement. The most straight forward approach to realize this time delay is using a transmission line, e.g. an optical fiber, which requires long lengths (in the orders of meters to kilometers) [1719] and rendering them unfavorable if not impossible for integrated circuits. In more elegant implementations, a delay line is substituted by a resonator and thereby the physical length of the delay element is reduced from meters in [23] to millimeters and centimeters in [2426]. Still the length of the structure which determines the delay required by the delay line is dictated by the number of virtual nodes.

 figure: Fig. 1.

Fig. 1. (a). Schematic of delay-based reservoir. (b). The virtual equivalent network of the delay-based reservoir for $N=5$. (c). Schematic of resonator-based reservoir. (d). The virtual equivalent network of the resonator-based reservoir for $N=5$. The strength of the connections is shown by their darkness and brightness. The weaker the connections, the lighter their color. The time delay of the connections is demonstrated using the number of triangles on each arrow.

Download Full Size | PDF

This inspires us to seek alternative virtual node based solutions which can be more easily integrated. Our idea is to use an echo generator replacing the lengthy delay line in Fig. 1(c). This echo generator block is realized in our proposal with a simple yet properly designed multi-mode Fabry-Perot resonator, in which the signal goes back and forth and generates an echo at each round trip. In a single-mode resonator; on the other hand, the signal is trapped within the resonator before gradually coming out of it. The former acts as an echo generator block while the latter imitates a delay line. Aside from the significant size reduction if one employs a compact multi-mode resonator, the dynamics of the system also allows a more versatile topology as depicted in Fig. 1(d). The use of resonator causes fundamental differences in the architecture of the virtual neural network. This means that unlike the previous implementations in which merely two neurons were connected at any moment, more than two neurons are connected at the same time. This can improve performance in some tasks such as memory capacity and nonlinear channel equalization which is demonstrated throughout our paper.

The paper is structured as follows. The resonator is a fundamental element in our structure which determines the equivalent virtual network of the reservoir computer. The details related to the multi-mode resonator are presented in section 2. Also, the effect of resonator parameters on the virtual network is investigated in this section. Our proposed structure and the equivalent virtual network of the reservoir computer is presented in section 3 and the performance of this RC under some standard tasks are presented in section 4. Finally, the conclusions are given in section 5.

2. Resonator in the multi-mode regime

Resonators are most commonly operate in the single-mode regime such that the input signal bandwidth is smaller than the resonator bandwidth. In this regime, the output wave shape of a single pulse after propagation through the resonator will be unchanged and the output of the resonator is the delayed version of the input. If the signal bandwidth is larger than the single-mode resonator bandwidth, a distorted and slightly delayed version of the input pulse reaches the output. Neither of these behaviours are desired for our case. We seek an output for the resonator which contains one or more delayed copies of the input pulse, with sufficient delay between the input and the last copy or echo of the input pulse. This requires to use the resonator in a multi-mode fashion, and the input pulse bandwidth to span multiple resonances of the resonator, to create proper echoes and delay. Different types of resonators such as bandpass and stopband structures are candidates such an echo generator block. In this paper, without the loss of generality and for simplicity’s sake, we use Fabry-Perot resonators which can be used in transmission and reflection mode.

First, we assume a Fabry-Perot resonator constructed by two mirrors with reflection coefficients $r_1$, $r_2$ and transmission coefficients $t_1$ and $t_2$ which are spaced L apart from each other. Two dielectric slabs can be used as the mirrors of the Fabry-Perot. For a slab having a length equal to an odd multiples of quarter wavelength, it can be found that the reflection and transmission coefficients of mirror will be $r_i$ and $jt_i$ such that $t_i^2 + r_i ^2 =1$. The medium between the two mirrors is filled homogeneously with a material of $n_r$ refractive index. In this model of the resonator, we assume that the variations of the refractive index and the length of the resonator over the bandwidth of the input signal can be ignored. Also, diffraction losses at the mirrors due to finite dimensions are neglected.

The impulse response of this Fabry-Perot resonator can be written as

$$T (t) ={-}t_1 t_2 \sum_{n=0}^{\infty} {r_1} ^n {r_2} ^n \delta(t-n t_{RT} - \dfrac{t_{RT}}{2}),$$
in which $t_{RT} = {2Ln_r}/{c}$ is the round-trip time of light propagating in the resonator and $c$ is the speed of light [27].

The impulse response of the resonator shows that the output of the resonator can be echoes of the input pulse if the pulse width in the time domain is smaller than the round trip time of the resonator. This means that in the frequency domain, the pulse bandwidth should be wider than the resonator bandwidth which includes multiple modes of the resonator as depicted in Fig. 2(a). In this regime, if a modulated Gaussian pulse with temporal width $\tau _p$ is incident on the resonator, the output of the resonator will be that shown in Fig. 2(c). The output contains the echoes of the input pulse with decaying amplitudes.

 figure: Fig. 2.

Fig. 2. (a). Transmission coefficient of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain when the bandwidth of Gaussian pulse is wider than resonator bandwidth so that the bandwidth of the pulse contains several modes of the resonator. Both the curves are normalized. (b) Transmission phase of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain. (c) Input Gaussian pulse and the output of the resonator in time domain.

Download Full Size | PDF

If one of the mirrors of the Fabry-Perot resonator is a perfect mirror, the reflection coefficient of the resonator can be written as:

$$R(t) = r_1 \delta(t) - t_1 ^2 \sum_{n=0}^{\infty} {r_1} ^{n} \delta(t-(n+1) t_{RT})$$

In reflection mode, in order for the resonator to reflect echoes of the input signal, the input signal pulse width must include several resonant modes, as before. The response of this resonator to a Gaussian pulse and the resonator response is shown in Fig. 3. The comparison of two Figs. 2 and 3 shows that the reduction rate of echoes in reflection mode resonator is lower than transmission mode, as expected. The FSR for a Fabry-Perot is defined as: $FSR = c/2n_rL$ and the finesse is approximated by:

$$F = \dfrac{\pi \sqrt{R_1 R_2 }}{1-\sqrt{R_1 R_2}},$$
in which $R_1 = |r_1|^2$ [27]. A Fabry-Perot resonator is characterized by these two parameters which determine the length of the resonator and the reflection coefficients of its mirrors. The length between two mirrors in the resonator determines the time interval between echoes at its output. The longer the resonator, the longer the time interval between echoes. The amplitude of the echoes is controlled by the reflection coefficient of the mirrors. The higher the reflection coefficient of the mirrors, the longer the wave stays inside the resonator. The reduction rate of echoes at the resonator output is determined by the reflection coefficient of the mirrors. In order to find a resonator-based reservoir with good performance, it is sufficient to train and test the system with different values for these two parameters.

 figure: Fig. 3.

Fig. 3. (a). Reflection coefficient of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain when the bandwidth of Gaussian pulse is wider than resonator bandwidth so that the bandwidth of the pulse contains several modes of the resonator. Both the curves are normalized. (b) Reflection phase of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain. (c) Input Gaussian pulse and the reflected signal from the resonator in time domain.

Download Full Size | PDF

3. Proposed architecture of reservoir computer

Two possible architectures of resonator based RC is shown in Fig. 4 and 5 which use Fabry-Perot resonator in transmission and reflection mode respectively. Our proposed RC consists of input, reservoir, and output layers. In both architectures, a laser is an element in the input layer and the optical source that feeds the $LiNbO_3$ Mach-Zehnder modulator (MZM). The reservoir layer of both implementations consists of: a Mach-Zehnder Modulator (MZM) as the nonlinear node, a Fabry-Perot resonator, an attenuator, and an electronic feedback loop. The MZM has a $\sin$ functionality and the phase $\phi$ determines the working point of the modulator which has been set to $\pi /4$. The modulated optical signal is fed to the resonator and a photodetector (PD) is employed to detect the resonator output. The modulation signal is generated using an arbitrary wave generator (AWG) in the input layer. The delayed versions of the input signal are added to the modulated signal and are applied to the MZM. In the architecture with the resonator in transmission mode, a significant portion of the power is reflected by the resonator, which is not favorable. The isolator avoids back reflections of the resonator. But, in Fig. 5 the resonator reflects all power and the input and output paths are separated by a circulator. In both implementations, the output of the reservoir layer is detected using a PD in the output layer and the electronic signal is sampled using a 100 Giga Sample/s digital oscilloscope. The number of nodes is equal to 50 neurons and the reflectivity of two mirrors of Fabry-Perot resonator and its FSR are two parameters that are investigated to achieve good performance.

 figure: Fig. 4.

Fig. 4. Schematic of our proposed RC in transmission mode.

Download Full Size | PDF

 figure: Fig. 5.

Fig. 5. Schematic of our proposed RC in reflection mode.

Download Full Size | PDF

3.1 Input layer

In the input layer, to connect the original signal to the virtual nodes, a pre-processing step is needed. First, the original signal $u(t)$ undergoes a sample and hold procedure with a sampling time interval $T$. Then, the resulting signal $u[n]$ is multiplied with a periodic masking signal $m(t)$. The masking signal has $N$ different values in each interval $T=N\theta$ and is a periodic signal with period $T$. The time interval $T$ and $\theta$ are called bit time and chip time respectively and $N$ is the number of virtual neurons.

3.2 Reservoir layer

The reservoir consists of a MZM as the nonlinear node, a resonator, and an electronic feedback loop which connects the echoes from the resonator output to the nonlinear node.

3.2.1 Topology of the equivalent virtual network

In order to determine the difference between a delay-based reservoir and a resonator-based reservoir, we consider their equivalent virtual neural networks. For delay-based reservoir computing in the unsynchronized regime, the reservoir state is formulated as follows:

$$\textbf{x}\left [n\right ] = \mathrm{f}_{\mathrm{NL}} \left( \alpha \textbf{W}_\mathrm{res_1}^k \textbf{x} \left [n-\ell\right ] +\alpha \textbf{W}_\mathrm{res_{2}}^k \textbf{x} \left [n-\ell-1 \right] + \beta \textbf{M} {\mathrm{u}} \left [n \right ] \right ),$$
where the delay length of the feedback loop is $\tau = (\ell N+k)\theta$. In this equation $x[n]$ is a vector of $N$ dimension which shows the state of the reservoir at time step $n$. $u[n]$ is the input signal and $M$ is a vector with $N$ values of the masking signal. $W_{res_1}^k$ and $W_{res_2}^k$ are the weight matrices of connections with delay of $\ell$ and $\ell +1$ time steps, respectively. The $W_{res_1}^k$ is a matrix having all of the $k$’th sub-diameter elements equal to unity, in which $k=1$ begins from the top sub-diameter below the original diameter. The $W_{res_2}^k$ is a matrix having all of the $k$’th sub-diameter elements equal to unity, in which $k=1$ begins from the top sub-diameter. For example, these two matrices are in the form of Eq. 5 when $k = 1$. The first diagonal below the main diagonal in the matrix $W_{res_1}^1$ have elements equal to unity and the $N-1$ diagonal above the main diagonal in the matrix $W_{res_2}^1$ also has an element equal to unity. Therefore, in the delay-based reservoir, only two time steps are simultaneously accessible.
$$W_{res_1}^1=\left(\begin{array}{ccccc} 0 & 0 & \cdots & 0 & 0\\ 1 & 0 & \cdots & 0 & 0 \\ 0 & 1 & \cdots & 0 & 0 \\ \vdots & \vdots & \ddots & \vdots & \vdots \\ 0 & 0 & \cdots & 1 & 0 \end{array}\right), W_{res_2}^1=\left(\begin{array}{ccccc} 0 & 0 & \cdots & 1\\ 0 & 0 & \cdots & 0 \\ 0 & 0 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & 0 \end{array}\right)$$

In the resonator-based reservoir with a multi-mode Fabry-Perot resonator, the input signal is combined with delayed versions of the reservoir node states. Every echo from the resonator acts as a delay line and therefore the reservoir states are calculated according to the following equation:

$$\textbf{x}\left [n\right ] = \mathrm{f}_{\mathrm{NL}} \Big (t_1 t_2\alpha \big( \sum _{q = 0} ^{\infty} {r_1} ^{q} {r_2} ^{q} \textbf{W}_\mathrm{res_{1}}^{k_q} \textbf{x} \left [n-\ell_q\right ] +{r_1} ^{q } {r_2} ^{q}\textbf{W}_\mathrm{res_{2}}^{k_q} \textbf{x} \left [n-\ell_q-1 \right] \big)+ \beta \textbf{M} {\mathrm{u}} \left [n \right ] \Big),$$
in which $q$ is a counter on echo number, $\ell _q$ and $k_q$ are determined using the delay of the echo $q$, and the matrices $W_{res_{1,2}}^k$ are $N\times N$ matrices which have been generalized to govern the recursive series of echoes and are constructed as explained in Eq. (4). The amplitude of the farther echoes damps gradually. Therefore, as in Fig. 1(d) is shown, the connections in the equivalent network are infinite and the strength of the connections damps as the delay of the connection increases.

3.2.2 Dynamics of the reservoir

In order to simulate the dynamics of the reservoir layer, the time domain reservoir output should be determined. The laser signal with frequency $f_0$ and amplitude $E_0$ enters the input port of the modulator. The output signal of the modulator is written as follows:

$$\begin{aligned} x(t) & = \dfrac{E_0}{2}\left \lbrace 1+\exp \left(j \left( \phi + m(t)u(t) + \left \vert x(t) \divideontimes T\left(t\right) \right \vert ^2 \right)\right) \right \rbrace e^{j\omega _c t}\\ & = \lbrace X(t) \rbrace e^{j\omega _c t}. \end{aligned}$$

After substituting Eq. (1) in (7), the low frequency part of reservoir state, $X(t)$, can be expressed as:

$$\begin{aligned} x(t) & = \dfrac{E_0}{2} \Bigg \lbrace 1+\exp \Big( {j\big( \phi + m(t)u(t) + \left \vert t_1 t_2 \sum_{n=0}^{\infty} {{r_1} ^n {r_2} ^n x\left(t-n t_{RT} - \dfrac{t_{RT}}{2} \right) } \right \vert ^2 \big)} \Big) \Bigg \rbrace e^{j\omega _c t}\\ & \begin{aligned} = \dfrac{E_0}{2} \Bigg \lbrace 1 & +\exp \Big(j\big( \phi +m(t)u(t) \\ & + \left \vert t_1 t_2 \sum_{n=0}^{\infty} {{r_1} ^n {r_2} ^n X\left(t-n t_{RT} - \dfrac{t_{RT}}{2} \right) e^{{-}j\omega _c \left(n t_{RT} + \dfrac{t_{RT}}{2} \right)}} \right \vert ^2 \big)\Big) \Bigg \rbrace e^{j\omega _c t}. \end{aligned} \end{aligned}$$

It should be noted that for the reflective implementation $T\left (t\right )$ should be replaced with $R\left (t\right )$. The reservoir output after photodiode is calculated as $x_{PD} = |x(t)|^2$ and is sampled in the middle of each chip time interval.

3.3 Output layer

The output of the reservoir $y[n]$ is a linear combination of reservoir states as: $y[n] = \textbf {W}_\mathrm {out}\: \textbf {x}[n]$, which $W_{out}$ is the weight matrix of the output layer and the output weights are determined by minimizing the mean square error (MSE) which is defined as follows:

$$\mathrm{MSE} = \dfrac{1}{N_{\mathrm{samp}}} \sum _{n=1} ^{N_{\mathrm{samp}}} { \left( y_{\mathrm{desired}} [n] - y[n] \right ) ^2},$$
where $y_{\mathrm {desired}}$ is the target output of the RC and $N_{\mathrm {samp}}$ is the number of input signal samples. This minimization is performed using an offline ridge regression algorithm. After the training phase, the matrix $W_{out}$ is fixed and we expose the testing data to our reservoir computer and the performance of the reservoir computer is evaluated by comparing the reservoir output and the expected output. Usually, this comparison is performed by calculating the normalized mean square error (NMSE) which is $MSE/var(y)$.

4. Simulation results

We evaluated the performance of our proposed RC on three benchmark tasks: signal classification, memory capacity, and nonlinear channel equalization. The number of virtual nodes in our simulations is $N=50$ and two parameters input gain $\alpha$ and feedback gain $\beta$ are optimized for each reservoir.

4.1 Signal classification

In the signal classification task, the input is a combination of square and triangular waves and the target output is 0 while the input is triangular wave and 1 otherwise. For every triangle sample, 10 voltage levels are used. Each of the training and test datasets consists of 1000 samples randomly chosen between square and triangle samples. We test our proposed RC in transmission mode using this task. Figure 6(a) shows the NMSE by varying the mirror reflectivity of the resonator. The FSR of the resonator is equal to 12.5 GHz. This plot shows that the NMSE is low for the reflectivity values below 0.95 and increases as the reflectivity increases. This is because by increasing the reflectivity, the echo amplitudes damp slowly and results in a network with long delay connection which connects input samples of the reservoir to each other. However, in the signal classification task, the arrangement of triangular and square samples is random and does not correlate with each other and also long delay connections are not required in this task. This is while, the strength of the important first echoes is low. Therefore, the performance of the RC decreases as the reflectivity increases. It should be noted that the NMSE obtained is several orders of magnitude better than the NMSE reported in [13]. We also add noise to the input signal of our RC and train the output layer using the noisy data. The NMSE obtained in different values of reflectivity is shown in Fig. 6(b) which is below 0.01 for some values of mirror reflectivity. The target output and the output of our RC for mirror reflectivity 0.91 is demonstrated in Fig. 7. Figure 7(a) shows the results for noiseless case which reveals that the output and target signals perfectly match and Fig. 7(b) shows the noisy case on which the output follows the target with a slight difference.

 figure: Fig. 6.

Fig. 6. Signal classification Task for the RC in transmission mode. (a) NMSE as a function of mirror reflectivity for signal classification without considering the effect of noise.(b) NMSE as a function of mirror reflectivity for signal classification with a noisy signal, SNR is 30 dB.

Download Full Size | PDF

 figure: Fig. 7.

Fig. 7. Signal classification Task for the RC in transmission mode. (a) The output of the reservoir and the target output for ideal signal classification. (b) The output of the reservoir and the target output for noisy signal classification.

Download Full Size | PDF

The performance of our reservoir computer when the reflectivities of the two mirrors of the resonator are 0.91 and 0.95 is studied under noisy signal classification task and the obtained results are shown in Fig. 8. For low SNR values the classification becomes harder and the performance of the reservoir is severely affected.

 figure: Fig. 8.

Fig. 8. Noisy signal classification Task for the RC in transmission mode. NMSE as a function of SNR for two mirror reflectivities 0.91 and 0.95.

Download Full Size | PDF

4.2 Nonlinear channel equalization

In this task, the reservoir computer is used as an equalizer of a nonlinear with memory channel. Assume we want to send 4 symbols $\{-3,-1,+1,+3\}$ on a wireless communication channel. These symbols will distort after propagating on the channel and the receiver, receives distorted symbols. In this task, the model of the communication channel is as follows, first a sequence of symbols $d[n]$ passes through a memory channel,

$$\begin{aligned} q[n]= & 0.08 d[n+2]-0.12 d[n+1]+d[n]+0.18 d[n-1] \\ & -0.1 d[n-2]+0.091 d[n-3]-0.05 d[n-4] \\ & +0.04 d[n-5]+0.03 d[n-6]+0.01 d[n-7]. \end{aligned}$$

Then it will transform nonlinearly as,

$$u[n]=q[n]+0.036 q[n]^{2}-0.011 q[n]^{3}+\nu[n],$$
in which $\nu [n]$ is zero-mean white Gaussian noise and its power is adjusted to the signal-to-noise (SNR) from 12-32 dB. This distorted signal is the input of the reservoir computer. In this task, the reservoir is trained to estimate the input of the channel which is $d[n]$. In order to evaluate the performance of our reservoir computer, the symbol error rate (SER) is calculated. Figure 9 and Fig. 10 show the SER for different values of mirror reflectivity of Fabry-Perot resonator in transmission and reflection modes respectively. The FSR of the resonators are 12.5 GHz, 6.25 GHz, and 3.125 GHz which provides the echo distance of $2\theta$, $4\theta$, and $6\theta$, respectively. The SER decreases when the echo interval increases. The reason for the increase in reservoir performance is that by increasing the echo interval, the stronger connections have longer latency. As a result, the reservoir gains richer dynamics. Also, this figure shows that when echo distance equals $2\theta$ and the reflectivity is in the range $[0.8, 0.94]$, the error is less than other values. For lower reflectivities, the amplitude of the output echoes damp more rapidly and the number of echoes with significant amplitude decreases. Therefore, it is possible that none of echoes result in delay larger than the period T. But as the reflectivity approaches 1, the damping rate of echoes decreases which leads to long-delay connections. A good reservoir should have fading memory to slowly reduce the effect of the past input samples. Therefore the output becomes independent of much earlier inputs. Due to the fact that increasing the distance of echoes, changes the virtual neural network and makes the connections with longer delays more stronger, the performance of the reservoir changes and the SER will be minimum in another reflectivity value. In other hand, as the echo distance increases to $6 \theta$, the SER is less dependant to the reflectivity values. Figure 9 and 10 show that the $\mathrm {SER} < 10^{-4}$ is possible which is the best values reported in papers yet. The length of an air-filled Fabry-Perot resonator which causes the echo distance equal to $6\theta$ will be 3.6 cm. It should be however noted that the air-filled Fabry-Perot resonator is merely an easy to design example that can be substituted with other types of electromagnetic resonators, e.g. micro ring arrays, point defects in photonic crystals, etc. that would have smaller foot print and can be much more easily integrated. Also, Fig. 11 and 12 demonstrate the SER as the SNR ranges from $12$ to $32$ dB. The value of the reflectivity of two mirrors is 0.91 and 0.55 in Fig. 11 and 12, respectively where the SER has its minimum value.

 figure: Fig. 9.

Fig. 9. Nonlinear channel equalization for the RC in transmission mode. Symbol error rate as a function of reflectivity of mirrors in transmission mode. The results are for 5 different runs for each reflectivity value. SNR is set to 32 dB.

Download Full Size | PDF

 figure: Fig. 10.

Fig. 10. Nonlinear channel equalization for the RC in reflection mode. Symbol error rate as a function of reflectivity of mirror in reflection mode. The results are for 5 different runs for each reflectivity value. SNR is set to 32 dB.

Download Full Size | PDF

 figure: Fig. 11.

Fig. 11. Nonlinear channel equalization for the RC in transmission mode. Symbol error rate as a function of SNR in transmission mode with constant mirror reflectivity of 0.91. The results are for 5 different runs for each reflectivity value.

Download Full Size | PDF

 figure: Fig. 12.

Fig. 12. Nonlinear channel equalization for the RC in reflection mode. Symbol error rate as a function of SNR in transmission mode with constant mirror reflectivity of 0.55. The results are for 5 different runs for each reflectivity value.

Download Full Size | PDF

4.3 Memory capacity

This task is introduced to determine the network memory capacity [28]. In this task, the reservoir computer is trained to generate delayed input signal. The input signal $u[n]$ is a uniformly distributed random series in the interval $[-1,+1]$. The desired output $u[n-k]$ is the input signal shifted k time steps in the past. In order to evaluate the k-delay memory capacity, the reservoir is trained for each k so that the memory capacity related to delay k maximizes. The total linear memory capacity is obtained by summing the memory capacities over all delays as follow:

$$M C=\sum_{k=1}^{\infty} M C_{k},$$
where the k-delay memory capacity $MC_{i}$ is defined as:
$$\begin{aligned} M C_{k} & =\max _{W_{k} ^{out}} d\left[W_{k} ^{out}](u(n-k), y_{k}(n)\right) \\ & =\max _{W_{k} ^{out}} \frac{\operatorname{cov}^{2}\left(u(n-k), y_{k}(n)\right)}{\sigma^{2}\left (u(n-k)\right) \sigma^{2}(y_{k}(n))}. \end{aligned}$$

Here, $d\left [W_k^{out}\right ]$ is the determination coefficient and is defined as the squared correlation coefficient of two signals. The k-delay memory capacity is in the range $[0,1]$ and the summation on $MC_k$ will be greater than 1. The memory capacity of our reservoir in transmission mode is shown in Fig. 13 for different values of mirror reflectivity and the echo distance. The memory capacity for echo distance equal to $2\theta$ ranges from 2 to 3 and for $R \in [0.9, 0.98]$ is larger than other values. Also, the memory capacity becomes larger by increasing the echo distance. As the echo distance increases the connections with longer delays are stronger. This results show that our reservoir has considerably higher memory than the one presented in [25].

 figure: Fig. 13.

Fig. 13. Memory capacity versus mirror reflectivity for the RC in transmission mode.

Download Full Size | PDF

5. Conclusion

We have proposed a novel photonic RC architecture which uses a multi-mode resonator instead of the lengthy delay line to realize virtual nodes required in the neural network. The resonator generates multiple echoes, each of them behaves like a delay line. Therefore, the virtual neural network is fundamentally different from the delay based reservoir computers. This new implementation causes richer dynamics that provides considerably higher memory capacity which is beneficial for tasks with high dynamics. Our proposed RC shows good performance in other standard tasks like signal classification and nonlinear channel equalization.

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. R. Pascanu, T. Mikolov, and Y. Bengio, “On the difficulty of training recurrent neural networks,” in International conference on machine learning, (PMLR, 2013), pp. 1310–1318.

2. W. Maass, T. Natschläger, and H. Markram, “Real-time computing without stable states: A new framework for neural computation based on perturbations,” Neural Comput. 14(11), 2531–2560 (2002). [CrossRef]  

3. H. Jaeger, “The echo state approach to analysing and training recurrent neural networks-with an erratum note,” Bonn, Germany: German National Research Center for Information Technology GMD Technical Report 148, 13 (2001).

4. H. J. Caulfield and S. Dolev, “Why future supercomputing requires optics,” Nat. Photonics 4(5), 261–263 (2010). [CrossRef]  

5. R. S. Tucker, “The role of optics in computing,” Nat. Photonics 4(7), 405 (2010). [CrossRef]  

6. D. A. Miller, “The role of optics in computing,” Nat. Photonics 4(7), 406 (2010). [CrossRef]  

7. D. Woods and T. J. Naughton, “Photonic neural networks,” Nat. Phys. 8(4), 257–259 (2012). [CrossRef]  

8. K. Vandoorne, W. Dierckx, B. Schrauwen, D. Verstraeten, R. Baets, P. Bienstman, and J. Van Campenhout, “Toward optical signal processing using photonic reservoir computing,” Opt. Express 16(15), 11182–11192 (2008). [CrossRef]  

9. K. Vandoorne, J. Dambre, D. Verstraeten, B. Schrauwen, and P. Bienstman, “Parallel reservoir computing using optical amplifiers,” IEEE Trans. Neural Netw. 22(9), 1469–1481 (2011). [CrossRef]  

10. F. Denis-Le Coarer, M. Sciamanna, A. Katumba, M. Freiberger, J. Dambre, P. Bienstman, and D. Rontani, “All-optical reservoir computing on a photonic chip using silicon-based ring resonators,” IEEE J. Sel. Top. Quantum Electron. 24(6), 1–8 (2018). [CrossRef]  

11. K. Vandoorne, P. Mechet, T. Van Vaerenbergh, M. Fiers, G. Morthier, D. Verstraeten, B. Schrauwen, J. Dambre, and P. Bienstman, “Experimental demonstration of reservoir computing on a silicon photonics chip,” Nat. Commun. 5(1), 3541 (2014). [CrossRef]  

12. M. A. A. Fiers, T. Van Vaerenbergh, F. Wyffels, D. Verstraeten, B. Schrauwen, J. Dambre, and P. Bienstman, “Nanophotonic reservoir computing with photonic crystal cavities to generate periodic patterns,” IEEE Trans. Neural Netw. Learning Syst. 25(2), 344–355 (2014). [CrossRef]  

13. Y. Paquot, F. Duport, A. Smerieri, J. Dambre, B. Schrauwen, M. Haelterman, and S. Massar, “Optoelectronic reservoir computing,” Sci. Rep. 2(1), 287 (2012). [CrossRef]  

14. Y. Paquot, J. Dambre, B. Schrauwen, M. Haelterman, and S. Massar, “Reservoir computing: a photonic neural network for information processing,” in Nonlinear Optics and Applications IV, vol. 7728 (International Society for Optics and Photonics, 2010), p. 77280B.

15. L. Appeltant, M. C. Soriano, G. Van der Sande, J. Danckaert, S. Massar, J. Dambre, B. Schrauwen, C. R. Mirasso, and I. Fischer, “Information processing using a single dynamical node as complex system,” Nat. Commun. 2(1), 468 (2011). [CrossRef]  

16. L. Larger, M. C. Soriano, D. Brunner, L. Appeltant, J. M. Gutiérrez, L. Pesquera, C. R. Mirasso, and I. Fischer, “Photonic information processing beyond turing: an optoelectronic implementation of reservoir computing,” Opt. Express 20(3), 3241–3249 (2012). [CrossRef]  

17. F. Duport, B. Schneider, A. Smerieri, M. Haelterman, and S. Massar, “All-optical reservoir computing,” Opt. Express 20(20), 22783–22795 (2012). [CrossRef]  

18. A. Dejonckheere, F. Duport, A. Smerieri, L. Fang, J.-L. Oudar, M. Haelterman, and S. Massar, “All-optical reservoir computer based on saturation of absorption,” Opt. Express 22(9), 10868–10881 (2014). [CrossRef]  

19. D. Brunner, M. C. Soriano, C. R. Mirasso, and I. Fischer, “Parallel photonic information processing at gigabyte per second data rates using transient states,” Nat. Commun. 4(1), 1364 (2013). [CrossRef]  

20. K. Hicke, M. A. Escalona-Morán, D. Brunner, M. C. Soriano, I. Fischer, and C. R. Mirasso, “Information processing using transient dynamics of semiconductor lasers subject to delayed feedback,” IEEE J. Sel. Top. Quantum Electron. 19(4), 1501610 (2013). [CrossRef]  

21. R. M. Nguimdo, G. Verschaffelt, J. Danckaert, and G. Van der Sande, “Fast photonic information processing using semiconductor lasers with delayed optical feedback: Role of phase dynamics,” Opt. Express 22(7), 8672–8686 (2014). [CrossRef]  

22. R. M. Nguimdo, G. Verschaffelt, J. Danckaert, and G. Van der Sande, “Simultaneous computation of two independent tasks using reservoir computing based on a single photonic nonlinear node with optical feedback,” IEEE Trans. Neural Netw. Learning Syst. 26(12), 3301–3307 (2015). [CrossRef]  

23. Q. Vinckier, F. Duport, A. Smerieri, K. Vandoorne, P. Bienstman, M. Haelterman, and S. Massar, “High-performance photonic reservoir computer based on a coherently driven passive cavity,” Optica 2(5), 438–446 (2015). [CrossRef]  

24. H. Zhang, X. Feng, B. Li, Y. Wang, K. Cui, F. Liu, W. Dou, and Y. Huang, “Integrated photonic reservoir computing based on hierarchical time-multiplexing structure,” Opt. Express 22(25), 31356–31370 (2014). [CrossRef]  

25. K. Takano, C. Sugano, M. Inubushi, K. Yoshimura, S. Sunada, K. Kanno, and A. Uchida, “Compact reservoir computing with a photonic integrated circuit,” Opt. Express 26(22), 29424–29439 (2018). [CrossRef]  

26. K. Harkhoe, G. Verschaffelt, A. Katumba, P. Bienstman, and G. Van der Sande, “Demonstrating delay-based reservoir computing using a compact photonic integrated chip,” Opt. Express 28(3), 3086–3096 (2020). [CrossRef]  

27. A. E. Siegman, “Lasers university science books,” Mill Valley, CA 37, 169 (1986).

28. H. Jaeger, “Tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF and the “echo state network” approach,” vol. 5 (GMD-Forschungszentrum Informationstechnik Bonn, 2002).

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (13)

Fig. 1.
Fig. 1. (a). Schematic of delay-based reservoir. (b). The virtual equivalent network of the delay-based reservoir for $N=5$. (c). Schematic of resonator-based reservoir. (d). The virtual equivalent network of the resonator-based reservoir for $N=5$. The strength of the connections is shown by their darkness and brightness. The weaker the connections, the lighter their color. The time delay of the connections is demonstrated using the number of triangles on each arrow.
Fig. 2.
Fig. 2. (a). Transmission coefficient of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain when the bandwidth of Gaussian pulse is wider than resonator bandwidth so that the bandwidth of the pulse contains several modes of the resonator. Both the curves are normalized. (b) Transmission phase of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain. (c) Input Gaussian pulse and the output of the resonator in time domain.
Fig. 3.
Fig. 3. (a). Reflection coefficient of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain when the bandwidth of Gaussian pulse is wider than resonator bandwidth so that the bandwidth of the pulse contains several modes of the resonator. Both the curves are normalized. (b) Reflection phase of the resonator (blue curve) and optical Gaussian pulse (red curve) in frequency domain. (c) Input Gaussian pulse and the reflected signal from the resonator in time domain.
Fig. 4.
Fig. 4. Schematic of our proposed RC in transmission mode.
Fig. 5.
Fig. 5. Schematic of our proposed RC in reflection mode.
Fig. 6.
Fig. 6. Signal classification Task for the RC in transmission mode. (a) NMSE as a function of mirror reflectivity for signal classification without considering the effect of noise.(b) NMSE as a function of mirror reflectivity for signal classification with a noisy signal, SNR is 30 dB.
Fig. 7.
Fig. 7. Signal classification Task for the RC in transmission mode. (a) The output of the reservoir and the target output for ideal signal classification. (b) The output of the reservoir and the target output for noisy signal classification.
Fig. 8.
Fig. 8. Noisy signal classification Task for the RC in transmission mode. NMSE as a function of SNR for two mirror reflectivities 0.91 and 0.95.
Fig. 9.
Fig. 9. Nonlinear channel equalization for the RC in transmission mode. Symbol error rate as a function of reflectivity of mirrors in transmission mode. The results are for 5 different runs for each reflectivity value. SNR is set to 32 dB.
Fig. 10.
Fig. 10. Nonlinear channel equalization for the RC in reflection mode. Symbol error rate as a function of reflectivity of mirror in reflection mode. The results are for 5 different runs for each reflectivity value. SNR is set to 32 dB.
Fig. 11.
Fig. 11. Nonlinear channel equalization for the RC in transmission mode. Symbol error rate as a function of SNR in transmission mode with constant mirror reflectivity of 0.91. The results are for 5 different runs for each reflectivity value.
Fig. 12.
Fig. 12. Nonlinear channel equalization for the RC in reflection mode. Symbol error rate as a function of SNR in transmission mode with constant mirror reflectivity of 0.55. The results are for 5 different runs for each reflectivity value.
Fig. 13.
Fig. 13. Memory capacity versus mirror reflectivity for the RC in transmission mode.

Equations (13)

Equations on this page are rendered with MathJax. Learn more.

T ( t ) = t 1 t 2 n = 0 r 1 n r 2 n δ ( t n t R T t R T 2 ) ,
R ( t ) = r 1 δ ( t ) t 1 2 n = 0 r 1 n δ ( t ( n + 1 ) t R T )
F = π R 1 R 2 1 R 1 R 2 ,
x [ n ] = f N L ( α W r e s 1 k x [ n ] + α W r e s 2 k x [ n 1 ] + β M u [ n ] ) ,
W r e s 1 1 = ( 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 0 ) , W r e s 2 1 = ( 0 0 1 0 0 0 0 0 0 0 0 0 )
x [ n ] = f N L ( t 1 t 2 α ( q = 0 r 1 q r 2 q W r e s 1 k q x [ n q ] + r 1 q r 2 q W r e s 2 k q x [ n q 1 ] ) + β M u [ n ] ) ,
x ( t ) = E 0 2 { 1 + exp ( j ( ϕ + m ( t ) u ( t ) + | x ( t ) T ( t ) | 2 ) ) } e j ω c t = { X ( t ) } e j ω c t .
x ( t ) = E 0 2 { 1 + exp ( j ( ϕ + m ( t ) u ( t ) + | t 1 t 2 n = 0 r 1 n r 2 n x ( t n t R T t R T 2 ) | 2 ) ) } e j ω c t = E 0 2 { 1 + exp ( j ( ϕ + m ( t ) u ( t ) + | t 1 t 2 n = 0 r 1 n r 2 n X ( t n t R T t R T 2 ) e j ω c ( n t R T + t R T 2 ) | 2 ) ) } e j ω c t .
M S E = 1 N s a m p n = 1 N s a m p ( y d e s i r e d [ n ] y [ n ] ) 2 ,
q [ n ] = 0.08 d [ n + 2 ] 0.12 d [ n + 1 ] + d [ n ] + 0.18 d [ n 1 ] 0.1 d [ n 2 ] + 0.091 d [ n 3 ] 0.05 d [ n 4 ] + 0.04 d [ n 5 ] + 0.03 d [ n 6 ] + 0.01 d [ n 7 ] .
u [ n ] = q [ n ] + 0.036 q [ n ] 2 0.011 q [ n ] 3 + ν [ n ] ,
M C = k = 1 M C k ,
M C k = max W k o u t d [ W k o u t ] ( u ( n k ) , y k ( n ) ) = max W k o u t cov 2 ( u ( n k ) , y k ( n ) ) σ 2 ( u ( n k ) ) σ 2 ( y k ( n ) ) .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.