TY - JOUR
T1 - Reprogrammable Electro-Optic Nonlinear Activation Functions for Optical Neural Networks
AU - Williamson, Ian A.D.
AU - Hughes, Tyler W.
AU - Minkov, Momchil
AU - Bartlett, Ben
AU - Pai, Sunil
AU - Fan, Shanhui
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/1
Y1 - 2020/1
N2 - We introduce an electro-optic hardware platform for nonlinear activation functions in optical neural networks. The optical-to-optical nonlinearity operates by converting a small portion of the input optical signal into an analog electric signal, which is used to intensity -modulate the original optical signal with no reduction in processing speed. Our scheme allows for complete nonlinear ON-OFF contrast in transmission at relatively low optical power thresholds and eliminates the requirement of having additional optical sources between each of the layers of the network Moreover, the activation function is reconfigurable via electrical bias, allowing it to be programmed or trained to synthesize a variety of nonlinear responses. Using numerical simulations, we demonstrate that this activation function significantly improves the expressiveness of optical neural networks, allowing them to perform well on two benchmark machine learning tasks: learning a multi-input exclusive-OR (XOR) logic function and classification of images of handwritten numbers from the MNIST dataset. The addition of the nonlinear activation function improves test accuracy on the MNIST task from 85% to 94%.
AB - We introduce an electro-optic hardware platform for nonlinear activation functions in optical neural networks. The optical-to-optical nonlinearity operates by converting a small portion of the input optical signal into an analog electric signal, which is used to intensity -modulate the original optical signal with no reduction in processing speed. Our scheme allows for complete nonlinear ON-OFF contrast in transmission at relatively low optical power thresholds and eliminates the requirement of having additional optical sources between each of the layers of the network Moreover, the activation function is reconfigurable via electrical bias, allowing it to be programmed or trained to synthesize a variety of nonlinear responses. Using numerical simulations, we demonstrate that this activation function significantly improves the expressiveness of optical neural networks, allowing them to perform well on two benchmark machine learning tasks: learning a multi-input exclusive-OR (XOR) logic function and classification of images of handwritten numbers from the MNIST dataset. The addition of the nonlinear activation function improves test accuracy on the MNIST task from 85% to 94%.
KW - Optical neural networks
KW - electro-optic modulators
KW - feedforward neural networks
KW - intensity modulation
KW - machine learning
KW - neuromorphic computing
KW - nonlinear optics
KW - phase modulation
KW - photodetectors
UR - http://www.scopus.com/inward/record.url?scp=85068938982&partnerID=8YFLogxK
U2 - 10.1109/JSTQE.2019.2930455
DO - 10.1109/JSTQE.2019.2930455
M3 - Article
AN - SCOPUS:85068938982
SN - 1077-260X
VL - 26
JO - IEEE Journal of Selected Topics in Quantum Electronics
JF - IEEE Journal of Selected Topics in Quantum Electronics
IS - 1
M1 - 7700412
ER -