Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Image edge detection with a photonic spiking VCSEL-neuron

Open Access Open Access

Abstract

We report both experimentally and in theory on the detection of edge features in digital images with an artificial optical spiking neuron based on a vertical-cavity surface-emitting laser (VCSEL). The latter delivers fast (< 100 ps) neuron-like optical spikes in response to optical inputs pre-processed using convolution techniques; hence representing image feature information with a spiking data output directly in the optical domain. The proposed technique is able to detect target edges of different directionalities in digital images by applying individual kernel operators and can achieve complete image edge detection using gradient magnitude. Importantly, the neuromorphic (brain-like) spiking edge detection of this work uses commercially sourced VCSELs exhibiting responses at sub-nanosecond rates (many orders of magnitude faster than biological neurons) and operating at the important telecom wavelength of 1300 nm; hence making our approach compatible with optical communication and data-centre technologies.

Published by The Optical Society under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

1. Introduction

Neuromorphic spiking models, developed to emulate the spiking responses of biological neurons in the brain [1], are seeing a rise in interest as new routes to novel computing paradigms are being explored. The brain, and its vast interconnected networks of neurons, is known to be capable of performing complex, human-like decision making more efficiently than traditional Von-Neumann computing architectures making the procurement of neuronal spike-based functionalities a key priority. Neuromorphic platforms based on electronic circuits for both research and technology applications have been developed in recent years with systems such as Neurogrid [2], SpiNNaker [3], TrueNorth [4] and more recently the BrainScaleS 2 neuromorphic chip [5] coming to fruition. These impressive systems have thrived due to the maturity of CMOS technology, however photonic approaches are also sought after by researchers due to their potential high energy efficiency, large bandwidth, low crosstalk, and fast spiking rates offered by light-based platforms. The push for photonic realisations has inspired a wave of optical spiking neuronal models based on numerous technologies such as phase change materials (PCMs) [6,7], resonant-tunnelling diodes (RTDs) [8,9], photonic cavities [10], optical modulators [11] and semiconductor lasers (SLs) [1225]. SL technology has given rise to a large number of spiking neuronal models with devices such as quantum dot lasers [12,13], micropillar lasers [14,15] and lasers with saturable absorber sections [16], to name a few (see [17] for a review).

Among the most promising SL neuronal models are vertical-cavity surface-emitting lasers (VCSELs) [1825]. These devices are reported to generate controllable excitable (spiking) dynamics along with key neuronal functionalities such as input thresholding and tonic spike activation [18,19]. The complex non-linear dynamics responsible for the excitable responses have also provided pathways to inhibitory dynamics [20], controllable spike rate-encoding for digital/analog-to-spike conversion at rates surpassing 1 GHz [21], and input integration [22,23] for realisations of logic and supervised pattern recognition systems. Additionally, theoretical studies into the achievable neuronal functionalities of VCSEL devices, based on the spin-flip model (SFM) [26,27], suggest their capability to perform learning based upon spike-timing dependent plasticity (STDP) as observed in biological neural networks [24,25]. The neuromorphic capabilities of these VCSEL-neurons are therefore exciting, and are only further enhanced by the inherent advantages of VCSELs, such as their affordability, high-speed and low-energy operation, ease of integration into 2D arrays of devices, telecom wavelength compatibility with data centre technologies (C&O telecom bands) and sub-nanosecond spiking dynamics, demonstrated at up to 7 orders of magnitude faster than biological timescales.

Image interpretation and classification is one of the characteristic tasks allowed by artificial neural networks. This complex functionality, realized computationally using convolutional neural networks (CNNs), applies the traditional image processing technique, convolution, in a large parallel network to achieve an overall recognition by assembling and comparing many smaller features [28]. Like neuromorphic spiking models, photonic realisations of CNNs have recently come under increasing investigation [2933] as again it is expected that the energy efficiency, speed and inherent parallelism offered by light-based technology will benefit the computationally intensive convolution task. At the heart of these complete systems, numerous photonic technologies such as ring resonators and photodiodes [29,30], Mach Zehnder interferometers [31,32], and PCMs [33] are being used for the creation of elements that perform weighting, summation, thresholding and activation functionalities.

In this work, we propose and demonstrate edge-feature image detection using a single artificial spiking VCSEL-neuron. The latter fires optical spiking patterns in response to pre-convolved optical inputs; hence representing edge-features (with different directionalities) in digital images with sub-ns long optical spikes. Specifically, we apply our photonic VCSEL-neuron technique for the edge detection of digital images at the key telecom wavelength of 1300 nm. Unlike reported implementations of CNNs, our technique utilizes time-division multiplexed optical inputs to enable operation with a single VCSEL-neuron; hence reducing hardware requirements, and provides output data in a spiking representation, directly in the optical domain. We provide both experimental findings and theoretical results modelling the anticipated spiking responses of the VCSEL-neuron based on the SFM [26,27]. This paper is organised as follows: in sections 2 & 3 we discuss the convolutional and experimental technique applied to achieve image edge-detection with a spiking photonic VCSEL-neuron. Section 4 describes the theoretical model used to validate the experimental findings. Section 5 provides theoretical and experimental results on vertical and horizontal image edge detection and in Section 6 we discuss and provide results of image gradient magnitude detection.

2. Image processing and convolution technique for image edge detection

In this work, image convolution is incorporated alongside an experimental realisation of a photonic spiking VCSEL-neuron (via modulated optical injection) for the first time, as shown in Fig. 1. In this section, we detail the pre-processing image convolution technique utilised (Fig. 1(a)). Initially, digital greyscale source images undergo conversion to positive (black) and negative (white) integer (1 and −1) matrices. Source images that contain different directional features were selected as shown in Fig. 2. The 28 × 28 pixel cross and saltire (Scotland’s national flag) images (Fig. 2(a)–(b)) were chosen as they contain horizontal, vertical and diagonal features. The larger 50 × 50 pixel image in Fig. 2(c) corresponds to the logo of our institute, the Institute of Photonics (IOP) at the University of Strathclyde. The IOP logo contains additional curved features and is much larger in size to illustrate the versatility of our work’s technique. In order to reveal edge information in these images, convolution is performed using a 2 × 2 kernel operator. The latter applies a weight to each pixel in a 2 × 2 region of the image and sums of all 4 weighted pixel values. The 4-value sum corresponds to the destination pixel value in the new convolved image (as shown in Fig. 1(a)). Different features can be targeted for recognition by applying different kernel operators and 2 × 2 kernels are scanned along every pixel in a row, and every row in an image (sliding window method). By comparing neighbouring pixels in this convolution process, we are able to identify features that best match our selected kernel operator. This process can be summarised using the equation:

$${g_{p,q}} = \sum\limits_{m = 0}^M {\sum\limits_{n = 0}^N {{f_{p + m,q + n}} \cdot {K_{m,n}}} }$$
where gp,q is the value of the destination pixel when the source image anchor-pixel fp,q is operated on by kernel K. A (M+1) x (N+1) pixel neighbourhood is operated on by the customisable kernel operator, in this work we set M = N = 1 to achieve a 2 × 2 neighbourhood array. No image buffer was used, hence both the number of rows and columns in the new convolved image is reduced by 1.

 figure: Fig. 1.

Fig. 1. Image convolution technique utilised to achieve spiking image edge detection and the experimental setup employed to implement it. Image processing procedure (a) and experimental realisation of the VCSEL-neuron (b). In (a), a black and white source image is converted into positive (1) and negative (−1) integers before it is multiplied by a 2 × 2 kernel operator. The resulting image is converted into a RZ image input where the destination pixel value is taken and encoded into the tunable laser’s optical intensity. In (b), light from a tunable laser is encoded with the convolved image input using a Mach Zehnder intensity modulator (MZ). The intensity encoded signal is injected into the spiking VCSEL-neuron whose response is collected via the optical circulator and analysed using a fast real-time oscilloscope. Two polarisation controllers (PC), an optical isolator (OI) and a variable optical attenuator (VOA) are used to control the light signals within the fibre-optic based experimental setup of this work.

Download Full Size | PDF

 figure: Fig. 2.

Fig. 2. Black and white source images of a cross (a), a saltire (Scotland’s national flag) (b) and the logo of the Strathclyde’s Institute of Photonics (IOP) (c). The vertical (Kernels 1–2), horizontal (Kernels 3–4) and diagonal (Kernels 5–8) operators are used to detect specific edge features in the source images. Source images (a) and (b) have a resolution of 28 × 28 pixels whereas source image (c) has a larger resolution of 50 × 50 pixels.

Download Full Size | PDF

The destination pixel value gp,q is injected into the VCSEL-neuron (Fig. 1(b)) which will then identify with ultrashort (sub-nanosecond) neuron-like spikes which pixels in the original source image contain the target feature. In order to generate the input signals entering the VCSEL-neuron, time-division multiplexing was used. In the time-multiplexed image input, each destination pixel was sequentially allocated the same configurable pixel duration. The latter was selected equal to 1.5 ns/pixel to coincide with the spiking refractory period of the VCSEL-neuron [22]. This, along with an inter-pixel return-to-zero (RZ) coding scheme, helped to ensure all neighbouring pixels were capable of triggering a single spiking event per input. All destination pixel values were held for 0.25 ns before returning to zero. The time-division multiplexed image input was produced in a 12 GSa/s, 5 GHz bandwidth, arbitrary waveform generator (AWG – Keysight M8190a) and fed to the VCSEL-neuron. The latter performs thresholding and spike activation operations to finally provide output data with a spiking representation (directly in the optical domain) of the edge-features present in the injected digital images, as shown in Fig. 1(b). We note that the image data input, here demonstrated at 1.5 ns/pixel (based on the refractory period of the commercially-obtained VCSEL-neuron) could be achieved at faster rates, e.g. sub-ns long pixel inputs, by further optimisation of device design and fabrication, which is beyond the scope of this study. Similarly, we note that overall processing rates can be improved by expanding from a single device to a multi-device architecture with parallel artificial neurons. These scaling-up approaches are of high interest and we plan to expand and explore the alternative system architectures in our future work.

3. Experimental VCSEL-neuron implementation for ultrafast spiking edge detection

The experimental setup used to implement the photonic VCSEL-neuron and perform spiking image edge detection is shown in the schematic diagram of Fig. 1(b). Similar to our previous work [23], modulated optical injection was used to trigger spiking responses from the VCSEL-neuron. Light from a 1300 nm tunable laser (TL – Santec TLS-210 V) was optically encoded with the time-division multiplexed image input (from the initial pre-processing step). A Mach Zehnder (MZ) intensity modulator was configured to introduce drops of intensity in the tunable laser’s light when subject to positive pulses from the image input. The encoded optical signal was passed to a coupler where a power meter (PM) made a measure of input power, and a circulator injected the optical signal into the VCSEL-neuron. An amplified 9 GHz bandwidth photodetector (Thorlabs PDA8GS) was used to collect the output of the VCSEL-neuron and a high-speed 13.5 GHz bandwidth, 40 GSa/s sampling rate, real-time oscilloscope (OSC - Agilent Infiniium DSO81304B) was used for temporal analysis. Throughout this work, the commercially sourced, fibre-pigtailed VCSEL device was biased with a current of 6.5 mA (Ith = 2.96 mA) and temperature stabilised at 298 K. Under these operating conditions, the VCSEL exhibited single mode lasing and the presence of two linear-orthogonally polarised modes, namely a main lasing (parallel) and a subsidiary-attenuated (orthogonal) polarisation mode. Optical injection was made at a frequency detuning (Δf) of −4.58 GHz from the peak of the subsidiary (orthogonal) mode, inducing polarisation switching during injection locking. An injection power of 152.7 µW was used to injection lock the device. Encoded intensity drops of sufficient amplitude were used to force the laser out of injection locking and into a regime of fast spike firing dynamics. The stability of experimental parameters such as laser bias, temperature, injection power and detuning is important to the controllable nature of the excitable dynamics with condition deviations affecting the overall performance of the spiking system. The VCSEL-neuron is therefore performing input thresholding and spike activation, allowing the system to reveal target feature information through the triggering of fast neuromorphic spiking events in the optical domain at the key telecom wavelength of 1300 nm.

4. Theoretical analysis of the VCSEL-neuron with the spin flip model (SFM)

We use a modified version of the well-known spin-flip model (SFM) [27] to evaluate theoretically the operation of the VCSEL-neuron and validate the experimental findings. The modified model used here includes an additional term that accounts for the convolution step, namely the injection of the time varying post-kernel image input. The modified rate equations are as shown below:

$$\begin{aligned} \frac{{d{E_{x,y}}}}{{dt}} &={-} ({k \pm {\gamma_a}} ){E_{x,y}} - i({k\alpha \pm {\gamma_p}} ){E_{x,y}}\\ &+ k({1 + i\alpha } )({N{E_{x,y}} \pm in{E_{x,y}}} )+ {k_{inj}}{E_{inj}}(t ){e^{i\Delta {\omega _x}t}} + {F_{x,y}} \end{aligned}$$
$$\frac{{dN}}{{dt}} ={-} {\gamma _N}[{N({1 + {{|{{E_x}} |}^2} + {{|{{E_y}} |}^2}} )- \mu + in({{E_y}E_x^\ast{-} {E_x}E_y^\ast } )} ]$$
$$\frac{{dn}}{{dt}} ={-} {\gamma _s}n - {\gamma _N}[{n({{{|{{E_x}} |}^2} - {{|{{E_y}} |}^2}} )+ iN({{E_y}E_x^\ast{-} {E_x}E_y^\ast } )} ]$$
where the subsidiary (orthogonally-polarised) and solitary (parallel-polarised) lasing modes of the VCSEL are represented by subscripts x and y respectively. The field amplitudes of the subsidiary and solitary modes are represented by Ex and Ey. N is the total carrier inversion between conduction and valence bands and n is the carrier inversion difference between spins of opposite polarity. γa is the gain anisotropy (dichroism) rate, γp is the linear birefringence rate, γN is the decay rate of the carrier inversion and γs is the spin-flip rate. k is the field decay rate, α is the linewidth enhancement rate and µ is the normalized pump current (µ = 1 represents the VCSEL’s threshold). Einj represents the post-kernel image input created during the convolution process and kinj is the injection strength. The angular frequency detuning is defined as Δωx = ωinj0, where the central frequency ω0 = (ωxy)/2 lies between the frequencies of the subsidiary ωx = ω0+αγap and the solitary mode ωy = ω0p-αγa. Δf = finj-fx is the frequency detuning between the injected field and the subsidiary mode, hence Δωx = 2πΔf+αγa-γ­p.

The spontaneous emission noise Fx and Fy are calculated as:

$${F_x} ={-} \sqrt {\frac{{{\beta _{sp}}{\gamma _N}}}{2}} \left( {\sqrt {N + n} {\xi_1} + \sqrt {N - n} {\xi_2}} \right)$$
$${F_y} ={-} i\sqrt {\frac{{{\beta _{sp}}{\gamma _N}}}{2}} \left( {\sqrt {N + n} {\xi_1} + \sqrt {N - n} {\xi_2}} \right)$$
where βsp represents the spontaneous emission strength and ξ1,2 represent two independent Gaussian white noise terms of zero mean and a unit variance. The model was solved using the fourth order Runge-Kutta method and the following parameters: γp = 128 ns−1, γa = 2 ns−1, γN = 0.5 ns−1, γs = 110 ns−1, α = 2, k = 185 ns−1, kinj = 15 ns−1 and βsp = 10−5.

5. Vertical and horizontal edge detection in source images using a spiking VCSEL-neuron

The detection of horizontal and vertical edge features was first tested using the ‘cross’ source image in Fig. 2(a). Kernels 1–4 (shown in Fig. 2) were sequentially applied to the source image and the resulting input values were experimentally injected into the VCSEL-neuron. Kernels 1–2 target vertical lines that transition from white-to-black (Kernel 1) and black-to-white (Kernel 2) pixels. Kernels 3–4 target in turn horizontal lines that transition from white-to-black (Kernel 3) and black-to-white (Kernel 4) pixels respectively. Figures 3(a-b) and 3(e-f) show respectively the spiking responses measured at the output of the VCSEL-neuron when applying individually Kernels 1–4. Image reconstruction maps are built to depict the collected time series from the VCSEL-neuron as intensity colour maps, where the spiking responses appear yellow and the resting state appears blue. Pixels with spiking responses should indicate the presence of the target feature in the source image. As expected, Figs. 3(a) and 3(b) demonstrate the triggering of a spike in response to vertical edges in the cross image. In Fig. 3(c) the image input, created using vertical Kernel 1, is shown for row 10 of Fig. 3(a). The image input injects a positive pulse for the detection of a matching target feature and a negative pulse for the detection of an inverse target feature. The corresponding VCSEL-neuron’s response, shown is Fig. 3(d), demonstrates that only the positive pulse triggers a fast ∼100 ps spike at the output of the VCSEL-neuron, highlighting the detection of the target feature. Therefore, applying Kernel 2 as shown in Fig. 3(b) successfully activates spiking responses to the opposite vertical edges in our image.

 figure: Fig. 3.

Fig. 3. Images built from the spiking responses obtained at the output of the VCSEL-neuron when Kernels 1 (a), 2 (b), 3 (e) and 4 (f) are applied to the cross image and injected into the system. The image input and output time series, corresponding to the selected rows in (a) and (e), are plotted at the bottom of the figure. Input (c) and output (d) correspond to row 10 of (a). Input (g) and output (h) correspond to row 12 of (e). Pixel duration is set to 1.5 ns/pixel in all cases.

Download Full Size | PDF

Similarly, just as Figs. 3(a) and 3(b) demonstrate vertical edge detection in the injected ‘cross’ patterned source image, Figs. 3(e) and 3(f) demonstrate in turn horizontal edge detection. The image input and system response for row 12 of Fig. 3(e) (Kernel 3) are shown in the time series of Figs. 3(g) and 3(h). Here, as Kernel 3 is scanned horizontally along that specific row of the source image, we create an input consisting of multiple positive target detections. The VCSEL-neuron responds to this input by firing multiple spiking events, one for each of the target detections. The spiking system is not triggered by the half-amplitude pulses (corresponding to corner edges) as the encoded input energy was not enough to cross the spiking activation threshold of the device. These results demonstrate that the experimental spiking system can threshold and activate spiking responses for horizontal edges when applying Kernels 3 and 4.

For comparison, Fig. 4 shows the theoretically calculated response of the VCSEL-neuron (using the model described in Section IV) when the same ‘cross’ image is injected into the model. From the calculated maps reconstructing the image (plotted in white and blue for distinction with the experimental findings of Fig. 3) we see excellent agreement with the measured results. The same number of spiking responses are activated when each kernel is applied to detect different vertical and horizontal image edge features. Also, a similar spiking threshold is achieved preventing the activation of corner edges giving an overall experimental accuracy of 100% in each of the 4 presented cases when compared with the numerical results. The spiking rate achieved in the theoretical results also showed a good correlation with the experiment allowing the model to operate at 1.5 ns/pixel. Therefore, both experiment and theory agree that convolutions performed by horizontal and vertical Kernels 1–4, can successfully be thresholded by the VCSEL-neuron to encode target edge-features with fast spiking responses directly in the optical domain. Additionally, we found that in both theory and experiment, the spike activation threshold could be controlled by varying injection power and frequency detuning, and that it could grant the detection of non-target features with smaller amplitude inputs (such as corners). Similarly, we expect that the control of the activation threshold can be used to apply this convolution technique to non-binary, greyscale images to selectively trigger the detection of lower contrast edge features.

 figure: Fig. 4.

Fig. 4. Simulations of the spiking response from the VCSEL-neuron, using the spin-flip model, when Kernels 1 (a), 2 (b), 3 (e) and 4 (f) are applied to the source ‘cross’ image before its injection into the system. Similar to Fig. 3, inputs and outputs are plotted for row 10 of (a) and row 12 of (e).

Download Full Size | PDF

6. Gradient-based edge detection in source images with a spiking VCSEL-neuron

In order to progress beyond the detection of one individual directional edge per input, we look towards performing gradient edge detection. Gradient is a vector with both a direction and a magnitude, which provide information about the rate of change of pixel intensity. Specifically calculating the magnitude of the gradient creates a set of data that can be injected into a VCSEL-neuron to perform spiking edge detection. The magnitude of the gradient |G(x,y)| can be calculated using the following equation:

$$|{G({x,y} )} |= \sqrt {G_x^2 + G_y^2}$$
where Gx and Gy are respectively the result of convolving a horizontal and a vertical kernel with the image. Both horizontal and vertical kernels must be 90° rotations of one another. By combining the results of two kernel operators in this way, we can detect edges in our images indiscriminately of their direction. Consequently, the convolution results of Kernels 1 and 3 were combined in this way to produce the gradient magnitude. The latter was taken in place of the destination pixel value when creating the image input for injection into the VCSEL-neuron. Gradient-based edge detection was performed on all three source images included in Fig. 2 and the results for both experimental and theoretical approaches are showcased in Figs. 5 and 6.

 figure: Fig. 5.

Fig. 5. Experimental (blue) and theoretical (white) responses when Kernel 1 and 3 are combined to detect gradient magnitude in the cross (a-d) and saltire (e-h) source images. (a,e) & (b,f) plot respectively experimental and theoretical image reconstruction maps obtained from the spiking responses at the VCSEL-neuron’s output. (c) & (d) plot respectively the experimental and theoretical time series of row 10 in (a) and (b). (g) and (h) plot respectively the experimental and theoretical time series of row 10 in (e) and (f).

Download Full Size | PDF

 figure: Fig. 6.

Fig. 6. Experimentally measured (a,b) and theoretically calculated (c,d) responses from the VCSEL-neuron illustrating the successful performance of gradient magnitude detection in the 50 × 50 pixel image of the Institute of Photonics (IOP) logo. Similar to Fig. 5, (b) and (d) plot the measured and calculated time series at the output of the VCSEL-neuron for row 25 in (a) and (c), respectively.

Download Full Size | PDF

Figures 5(a) and 5(c) show the experimental response from the VCSEL-neuron when gradient-based edge detection is performed on the ‘cross’ source image of Fig. 2(a). The time series demonstrates that we now have a spiking response for both vertical edges (white-to-black and black-to-white transitions) and the image maps, built from the VCSEL-neuron’s spiking output, reveal also the successful detection of the horizontal edges in the ‘cross’. Figures 5(b) and 5(d) also demonstrate the modelled activation of spiking events for each edge of the ‘cross’ image showing excellent agreement with the experimental results (an experimental accuracy of 99.18% when compared with the theoretical result). The experimental gradient edge detection of the ‘Saltire’ image (Fig. 2(b)) is shown in Fig. 5(e) illustrating also the successful detection of diagonal edges from a source image with our VCSEL-neuron. Despite gradient detection combining a horizontal and a vertical kernel, we see from the image map (Fig. 5(e)) and time series (Fig. 5(g)) that all diagonal edges were successfully detected in the ‘Saltire’ source image. The modelling of the VCSEL-neuron’s response provided in Figs. 5(f) and 5(h) again shows excellent agreement with the experimental results (experimental results are 99.86% accurate to the theoretical findings), revealing the successful detection of all the diagonal edges. Figure 6 shows the experimental and theoretical response of the VCSEL-neuron to the gradient edge detection of the ‘IOP’ logo source image included in Fig. 2(c). This larger 50 × 50 pixel image contained both straight and curved lines (as shown in Fig. 2(c)), distributed unevenly across the image background. The image map and time series reveal that the gradient detection successfully reveals every directionality of edge. Despite the larger image, the pixel duration remained consistent and only a larger overall time series was required. Again, the modelling of the VCSEL-neuron (Figs. 6(c) and 6(d)) showed excellent agreement with the experimental results with an overall experimental accuracy of 99.67% compared to the numerical results. Spiking edge detection can therefore be successfully performed by calculating the magnitude of the gradient and using the VCSEL-neuron to threshold the convolved inputs.

7. Conclusion

In summary, we demonstrate image edge detection using a single spiking VCSEL-neuron operating at the key telecom wavelength of 1300 nm. Our approach uses time-division multiplexing for image data injection; hence allowing for single VCSEL operation, that delivers the output data of image edge-features in a fast spiking representation (<100 ps long spikes) directly in the optical domain. The artificial neuronal model presented demonstrates the spiking edge detection of vertical, horizontal and diagonal straight lines using individual 2 × 2 kernel operators and traditional image convolution techniques. Building upon this, the identification and extraction of multiple straight and curved edges, irrespective of their directionality, was achieved in a single input run by calculating and thresholding image gradient magnitude. Furthermore, the experimental findings in this work are in excellent agreement with numerical simulations carried out using a modified version of the SFM. The commercially available and telecommunication compatible spiking photonic VCSEL-based neurons used in this proof-of-concept demonstration represent a first step towards the future development of high-speed neuromorphic photonic systems formed by interconnected networks of VCSEL neurons for enhanced image processing capability with additional functionalities (e.g. multiple input integration [23], photonic-hardware implemented weight control [24,25,32]) and pathways to help integrate with fields such as optical communications, directly in the optical domain.

Funding

Office of Naval Research Global (ONRGNICOP-N62909-18-1-2027); Horizon 2020 Framework Programme (828841-ChipAI-H2020-FETOPEN2018-2020); National Natural Science Foundation of China (61674119, 61974177); Engineering and Physical Sciences Research Council Doctoral Training Partnership (EP/N509760).

Acknowledgments

The authors would like to acknowledge Prof. A. Kemp and Prof. T. Ackemann from the University of Strathclyde for lending some of the equipment used in this work.

Disclosures

The authors declare no conflicts of interest.

Data Availability

All data underpinning this publication are openly available from the University of Strathclyde KnowledgeBase at https://doi.org/10.15129/551ef443-84fd-4c3a-8a1b-d381d0788cf2.

References

1. A. S. Cassidy, J. Georgiou, and A. G. Andreou, “Design of silicon brains in the nano-CMOS era: spiking neurons, learning synapses and neural architecture optimization,” Neural Networks 45, 4–26 (2013). [CrossRef]  

2. B. V. Benjamin, P. Gao, E. McQuinn, S. Choudhary, A. R. Chandrasekaran, J.-M. Bussat, R. Alvarez-Icaza, J. V. Arthur, P. A. Merolla, and K. Boahen, “Neurogrid: a mixed-analog-digital multichip system for large-scale neural simulations,” Proc. IEEE 102(5), 699–716 (2014). [CrossRef]  

3. S. Furber, F. Galluppi, S. Temple, and L. Plana, “The SpiNNaker Project,” Proc. IEEE 102(5), 652–665 (2014). [CrossRef]  

4. M. V. DeBole, B. Taba, A. Amir, F. Akopyan, A. Andreopoulos, W. P. Risk, J. Kusnitz, C. Ortega Otero, T. K. Nayak, R. Appuswamy, P. J. Carlson, A. S. Cassidy, P. Datta, S. K. Esser, G. J. Garreau, K. L. Holland, S. Lekuch, M. Mastro, J. McKinstry, C. di Nolfo, B. Paulovicks, J. Sawada, K. Schleupen, B. G. Shaw, J. L. Klamo, M. D. Flickner, J. V. Arthur, and D. S. Modha, “TrueNorth: accelerating from zero to 64 million neurons in 10 years,” Comput. 52(5), 20–29 (2019). [CrossRef]  

5. S. Friedmann, J. Schemmel, A. Grubl, A. Hartel, M. Hock, and K. Meier, “Demonstrating Hybrid Learning in a Flexible Neuromorphic Hardware System,” IEEE Trans. Biomed. Circuits Syst. 11(1), 128–142 (2017). [CrossRef]  

6. J. Feldmann, N. Youngblood, C. D. Wright, H. Bhaskaran, and W. H. P. Pernice, “All-optical Spiking Neurosynaptic Networks with Self-learning Capabilities,” Nature 569(7755), 208–214 (2019). [CrossRef]  

7. I. Chakraborty, G. Saha, A. Sengupta, and K. Roy, “Toward Fast Neural Computing Using All-Photonic Phase Change Spiking Neurons,” Sci. Rep. 8(1), 12980 (2018). [CrossRef]  

8. B. Romeira, J. Javaloyes, C. N. Ironside, J. M. L. Figueiredo, S. Balle, and O. Piro, “Excitability and Optical Pulse Generation in Semiconductor Lasers Driven by Resonant Tunneling Diode Photo-detectors,” Opt. Express 21(18), 20931 (2013). [CrossRef]  

9. B. Romeira, J. M. L. Figueiredo, and J. Javaloyes, “Delay Dynamics of Neuromorphic Optoelectronic Nanoscale Resonators: Perspectives and Applications,” Chaos 27(11), 114323 (2017). [CrossRef]  

10. M. Brunstein, A. M. Yacomotti, I. Sagnes, F. Raineri, L. Bigot, and A. Levenson, “Excitability and self-pulsing in a photonic crystal nanocavity,” Phys. Rev. A 85(3), 031803 (2012). [CrossRef]  

11. J. K. George, A. Mehrabian, R. Amin, J. Meng, T. F. de Lima, A. N. Tait, B. J. Shastri, T. El-Ghazawi, P. R. Prucnal, and V. J. Sorger, “Neuromorphic Photonics with Electro-absorption Modulators,” Opt. Express 27(4), 5181 (2019). [CrossRef]  

12. C. Mesaritakis, A. Kapsalis, A. Bogris, and D. Syvridis, “Artificial Neuron Based on Integrated Semiconductor Quantum Dot Mode-locked Lasers,” Sci. Rep. 6(1), 39317 (2016). [CrossRef]  

13. M. Dillane, I. Dubinkin, N. Fedorov, T. Erneux, D. Goulding, B. Kelleher, and E. A. Viktorov, “Excitable Interplay Between Lasing Quantum Dot States,” Phys. Rev. E 100(1), 012202 (2019). [CrossRef]  

14. F. Selmi, R. Braive, G. Beaudoin, I. Sagnes, R. Kuszelewicz, T. Erneux, and S. Barbay, “Spike Latency and Response Properties of an Excitable Micropillar Laser,” Phys. Rev. E 94(4), 042219 (2016). [CrossRef]  

15. V. A. Pammi, K. Alfaro-Bittner, M. G. Clerc, and S. Barbay, “Photonic Computing with Single and Coupled Spiking Micropillar Lasers,” IEEE J. Sel. Top. Quantum Electron. 26(1), 1–7 (2020). [CrossRef]  

16. S. Barbay, R. Kuszelewicz, and A. M. Yacomotti, “Excitability in a Semiconductor Laser with Saturable Absorber,” Opt. Lett. 36(23), 4476–4478 (2011). [CrossRef]  

17. P. R. Prucnal, B. J. Shastri, and T. Ferreira de Lima, “Recent Progress in Semiconductor Excitable Lasers for Photonic Spike Processing,” Adv. Opt. Photonics 8(2), 228–299 (2016). [CrossRef]  

18. B. Garbin, G. Tissoni, and S. Barland, “Excitable Pulses and Diffusion of Localized Stated in a Driven Semiconductor Laser with Delay,” CAP 7(3), 96–101 (2018). [CrossRef]  

19. A. Hurtado and J. Javaloyes, “Controllable Spiking Patterns in Long-wavelength Vertical Cavity Surface Emitting Lasers for Neuromorphic Photonics Systems,” Appl. Phys. Lett. 107(24), 241103 (2015). [CrossRef]  

20. J. Robertson, T. Deng, J. Javaloyes, and A. Hurtado, “Controlled Inhibition of Spiking Dynamics in VCSELs for Neuromorphic Photonics: Theory and Experiments,” Opt. Lett. 42(8), 1560–1563 (2017). [CrossRef]  

21. M. Hejda, J. Robertson, J. Bueno, and A. Hurtado, “Spike-based Information Encoding in Vertical Cavity Surface Emitting Lasers for Neuromorphic Photonic Systems,” JPhys Photonics 2(4), 044001 (2020). [CrossRef]  

22. A. Dolcemascolo, B. Garbin, B. Peyce, R. Veltz, and S. Barland, “Resonator Neuron and Triggering Multipulse Excitability in Laser with Injected Signal,” Phys. Rev. E 98(6), 062211 (2018). [CrossRef]  

23. J. Robertson, M. Hejda, J. Bueno, and A. Hurtado, “Ultrafast Optical Integration and Pattern Classification for Neuromorphic Photonics Based on Spiking VCSEL Neurons,” Sci. Rep. 10(1), 6098 (2020). [CrossRef]  

24. S. Xiang, Y. Zhang, J. Gong, X. Guo, L. Lin, and Y. Hao, “STDP-based Unsupervised Spike Pattern Learning in a Photonic Spiking Neural Network with VCSELs and VCSOAs,” IEEE J. Sel. Top. Quantum Electron. 25(6), 1–9 (2019). [CrossRef]  

25. S. Xiang, J. Gong, Y. Zhang, X. Guo, Y. Han, A. Wen, and Y. Hao, “Numerical Implementation of Wavelength-dependent Photonic Spike Timing Dependent Plasticity Based on VCSOA,” IEEE J. Quantum Electron. 54(6), 1–7 (2018). [CrossRef]  

26. J. Martin-Regalado, F. Prati, M. San Miguel, and N. B. Abraham, “Polarization Properties of Vertical-cavity Surface-emitting Lasers,” IEEE J. Quantum Electron. 33(5), 765–783 (1997). [CrossRef]  

27. H. Susanto, K. Schires, M. J. Adams, and I. D. Henning, “Spin-flip Model of Spin-polarized Vertical-cavity Surface-emitting Lasers: Asymptotic Analysis, Numerics, and Experiments,” Phys. Rev. A 92(6), 063838 (2015). [CrossRef]  

28. M. El-Sayed, Y. A. Estaitia, and M. A. Khafagy, “Automated Edge Detection Using Convolutional Neural Network,” IJACSA 4(10), 11–17 (2013). [CrossRef]  

29. V. Bangari, B. A. Marquez, H. Miller, A. N. Tait, M. A. Nahmias, T. F. de Lima, H. Peng, P. R. Prucnal, and B. J. Shastri, “Digital Electronics and Analog Photonics for Convolutional Neural Networks (DEAP-CNNs),” IEEE J. Sel. Top. Quantum Electron. 26(1), 1–13 (2020). [CrossRef]  

30. A. Mehrabian, Y. Al-Kabani, V. J. Sorger, and T. El-Ghazawi, “PCNNA: A Photonic Convolutional Neural Network Accelerator,” in 2018 31st IEEE International System-on-Chip Conference (2018), pp. 169–173.

31. Y. Shen, N. C. Harris, S. Skirlo, M. Prabhu, T. Baehr-Jones, M. Hochberg, X. Sun, S. Zhao, H. Larochelle, D. Englund, and M. Soljačić, “Deep Learning with Coherent Nanophotonic Circuits,” Nat. Photonics 11(7), 441–446 (2017). [CrossRef]  

32. S. Xu, J. Wang, R. Wang, J. Chen, and W. Zou, “High-accuracy Optical Convolution Unit Architecture for Convolutional Neural Networks by Cascaded Acousto-optical Modulator Arrays,” Opt. Express 27(14), 19778–19787 (2019). [CrossRef]  

33. J. Feldmann, N. Youngblood, M. Karpov, H. Gehring, X. Li, M. Le Gallo, X. Fu, A. Lukashchuk, A. Raja, J. Liu, D. Wright, A. Sebastian, T. Kippenberg, W. Pernice, and H. Bhaskaran, “Parallel Convolution Processing Using an Integrated Photonic Tensor Core,” arXiv:2002.00281v1 (2020).

Data Availability

All data underpinning this publication are openly available from the University of Strathclyde KnowledgeBase at https://doi.org/10.15129/551ef443-84fd-4c3a-8a1b-d381d0788cf2.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1.
Fig. 1. Image convolution technique utilised to achieve spiking image edge detection and the experimental setup employed to implement it. Image processing procedure (a) and experimental realisation of the VCSEL-neuron (b). In (a), a black and white source image is converted into positive (1) and negative (−1) integers before it is multiplied by a 2 × 2 kernel operator. The resulting image is converted into a RZ image input where the destination pixel value is taken and encoded into the tunable laser’s optical intensity. In (b), light from a tunable laser is encoded with the convolved image input using a Mach Zehnder intensity modulator (MZ). The intensity encoded signal is injected into the spiking VCSEL-neuron whose response is collected via the optical circulator and analysed using a fast real-time oscilloscope. Two polarisation controllers (PC), an optical isolator (OI) and a variable optical attenuator (VOA) are used to control the light signals within the fibre-optic based experimental setup of this work.
Fig. 2.
Fig. 2. Black and white source images of a cross (a), a saltire (Scotland’s national flag) (b) and the logo of the Strathclyde’s Institute of Photonics (IOP) (c). The vertical (Kernels 1–2), horizontal (Kernels 3–4) and diagonal (Kernels 5–8) operators are used to detect specific edge features in the source images. Source images (a) and (b) have a resolution of 28 × 28 pixels whereas source image (c) has a larger resolution of 50 × 50 pixels.
Fig. 3.
Fig. 3. Images built from the spiking responses obtained at the output of the VCSEL-neuron when Kernels 1 (a), 2 (b), 3 (e) and 4 (f) are applied to the cross image and injected into the system. The image input and output time series, corresponding to the selected rows in (a) and (e), are plotted at the bottom of the figure. Input (c) and output (d) correspond to row 10 of (a). Input (g) and output (h) correspond to row 12 of (e). Pixel duration is set to 1.5 ns/pixel in all cases.
Fig. 4.
Fig. 4. Simulations of the spiking response from the VCSEL-neuron, using the spin-flip model, when Kernels 1 (a), 2 (b), 3 (e) and 4 (f) are applied to the source ‘cross’ image before its injection into the system. Similar to Fig. 3, inputs and outputs are plotted for row 10 of (a) and row 12 of (e).
Fig. 5.
Fig. 5. Experimental (blue) and theoretical (white) responses when Kernel 1 and 3 are combined to detect gradient magnitude in the cross (a-d) and saltire (e-h) source images. (a,e) & (b,f) plot respectively experimental and theoretical image reconstruction maps obtained from the spiking responses at the VCSEL-neuron’s output. (c) & (d) plot respectively the experimental and theoretical time series of row 10 in (a) and (b). (g) and (h) plot respectively the experimental and theoretical time series of row 10 in (e) and (f).
Fig. 6.
Fig. 6. Experimentally measured (a,b) and theoretically calculated (c,d) responses from the VCSEL-neuron illustrating the successful performance of gradient magnitude detection in the 50 × 50 pixel image of the Institute of Photonics (IOP) logo. Similar to Fig. 5, (b) and (d) plot the measured and calculated time series at the output of the VCSEL-neuron for row 25 in (a) and (c), respectively.

Equations (7)

Equations on this page are rendered with MathJax. Learn more.

g p , q = m = 0 M n = 0 N f p + m , q + n K m , n
d E x , y d t = ( k ± γ a ) E x , y i ( k α ± γ p ) E x , y + k ( 1 + i α ) ( N E x , y ± i n E x , y ) + k i n j E i n j ( t ) e i Δ ω x t + F x , y
d N d t = γ N [ N ( 1 + | E x | 2 + | E y | 2 ) μ + i n ( E y E x E x E y ) ]
d n d t = γ s n γ N [ n ( | E x | 2 | E y | 2 ) + i N ( E y E x E x E y ) ]
F x = β s p γ N 2 ( N + n ξ 1 + N n ξ 2 )
F y = i β s p γ N 2 ( N + n ξ 1 + N n ξ 2 )
| G ( x , y ) | = G x 2 + G y 2
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.