2019
DOI: 10.1364/oe.27.037150
|View full text |Cite
|
Sign up to set email alerts
|

Efficient training and design of photonic neural network through neuroevolution

Abstract: Recently, optical neural networks (ONNs) integrated in photonic chips has received extensive attention because they are expected to implement the same pattern recognition tasks in the electronic platforms with high efficiency and low power consumption. However, the current lack of various learning algorithms to train the ONNs obstructs their further development. In this article, we propose a novel learning strategy based on neuroevolution to design and train the ONNs. Two typical neuroevolution algorithms are … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
27
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 49 publications
(27 citation statements)
references
References 55 publications
0
27
0
Order By: Relevance
“…Adjoint variable method (AVM) (Hughes et al 2018) is proposed to model the circuit state as a partial-differential-equation-controlled linear system, and directly measures the exact gradient via in situ light intensity measurement. Evolutionary algorithms, e.g., particle swarm optimization and genetic algorithm, are demonstrated to train MZIs on chip (Zhang et al 2019). A stochastic zeroth-order gradient descent based method FLOPS (Gu et al 2020a) has been proposed to improve the training efficiency by 3-5× compared with previous methods.…”
Section: Onn Architecture and Training Methodsmentioning
confidence: 99%
“…Adjoint variable method (AVM) (Hughes et al 2018) is proposed to model the circuit state as a partial-differential-equation-controlled linear system, and directly measures the exact gradient via in situ light intensity measurement. Evolutionary algorithms, e.g., particle swarm optimization and genetic algorithm, are demonstrated to train MZIs on chip (Zhang et al 2019). A stochastic zeroth-order gradient descent based method FLOPS (Gu et al 2020a) has been proposed to improve the training efficiency by 3-5× compared with previous methods.…”
Section: Onn Architecture and Training Methodsmentioning
confidence: 99%
“…Recently, another approach to realise optical neural networks was based on Mach-Zehnder interferometers (MZIs) to calculate matrix products [25,26], see Figure 6b. By carefully manipulating a specific phase shift between a coherent pair of incoming light pulses allow to multiply a two-element vector, encoded in the amplitude of the pulses, by a two-by-two matrix [27,28]. An array of the interferometers can then perform arbitrary matrix operations, which is widely used, for example, in the boson sampling approach.…”
Section: All-optical Neural Networkmentioning
confidence: 99%
“…In addition, the high tunability of graphene metalines can be combined with inverse design technology to design a more efficient photonics device [28]- [30]. In general, inverse design is converted to the optimization problem which can be solved by gradient based methods (such as adjoint variable method (AVM) [31]- [32]), gradient free methods (such as genetic algorithm (GA) [33]- [37]) and model based methods (such as machine learning [38]- [40]). As a representative algorithm of gradient based methods, AVM not only designs linear devices but also optimizes for the nonlinear devices in frequency domain, but it requires physical background to derive the gradient of objective function [31]- [32].…”
Section: Introductionmentioning
confidence: 99%
“…However, in order to train the model whose inputs are physical parameters and outputs are electromagnetic responses, it requires a significant amount of time to generate the training instances and test instances [38]. In comparison to the gradient based methods and model based methods, gradient free methods, which depend on search strategy and evolutionary strategy, are simple, effective and parallelizable [33]- [37]. As a result, although GA easily falls into local optimum and demands tremendous computational time, it has been applied in the inverse design for many photonics devices, such as polarization beam splitters [33], polarization rotators [35], diodes [34], waveguide crossings [36], optical neural networks [37] and so on.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation