TY - JOUR
T1 - Physics-Aware Analytic-Gradient Training of Photonic Neural Networks
AU - Zhan, Yuancheng
AU - Zhang, Hui
AU - Lin, Hexiang
AU - Chin, Lip Ket
AU - Cai, Hong
AU - Karim, Muhammad Faeyz
AU - Poenar, Daniel Puiu
AU - Jiang, Xudong
AU - Mak, Man Wai
AU - Kwek, Leong Chuan
AU - Liu, Ai Qun
N1 - Publisher Copyright:
© 2024 The Authors. Laser & Photonics Reviews published by Wiley-VCH GmbH.
PY - 2024/4
Y1 - 2024/4
N2 - Photonic neural networks (PNNs) have emerged as promising alternatives to traditional electronic neural networks. However, the training of PNNs, especially the chip implementation of analytic gradient descent algorithms that are recognized as highly efficient in traditional practice, remains a major challenge because physical systems are not differentiable. Although training methods such as gradient-free and numerical gradient methods are proposed, they suffer from excessive measurements and limited scalability. State-of-the-art in situ training method is also cost-challenged, requiring expensive in-line monitors and frequent optical I/O switching. Here, a physics-aware analytic-gradient training (PAGT) method is proposed that calculates the analytic gradient in a divide-and-conquer strategy, overcoming the difficulty induced by chip non-differentiability in the training of PNNs. Multiple training cases, especially a generative adversarial network, are implemented on-chip, achieving a significant reduction in time consumption (from 31 h to 62 min) and a fourfold reduction in energy consumption, compared to the in situ method. The results provide low-cost, practical, and accelerated solutions for training hybrid photonic-digital electronic neural networks.
AB - Photonic neural networks (PNNs) have emerged as promising alternatives to traditional electronic neural networks. However, the training of PNNs, especially the chip implementation of analytic gradient descent algorithms that are recognized as highly efficient in traditional practice, remains a major challenge because physical systems are not differentiable. Although training methods such as gradient-free and numerical gradient methods are proposed, they suffer from excessive measurements and limited scalability. State-of-the-art in situ training method is also cost-challenged, requiring expensive in-line monitors and frequent optical I/O switching. Here, a physics-aware analytic-gradient training (PAGT) method is proposed that calculates the analytic gradient in a divide-and-conquer strategy, overcoming the difficulty induced by chip non-differentiability in the training of PNNs. Multiple training cases, especially a generative adversarial network, are implemented on-chip, achieving a significant reduction in time consumption (from 31 h to 62 min) and a fourfold reduction in energy consumption, compared to the in situ method. The results provide low-cost, practical, and accelerated solutions for training hybrid photonic-digital electronic neural networks.
KW - on-chip training
KW - optical computing
KW - photonic integrated chip
KW - photonic neural networks
UR - http://www.scopus.com/inward/record.url?scp=85185662693&partnerID=8YFLogxK
U2 - 10.1002/lpor.202300445
DO - 10.1002/lpor.202300445
M3 - Journal article
AN - SCOPUS:85185662693
SN - 1863-8880
VL - 18
JO - Laser and Photonics Reviews
JF - Laser and Photonics Reviews
IS - 4
M1 - 2300445
ER -