“…Deep, convolutional, and residual feed-forward NN structures are discussed in [4], [7], [8], and [9], which all rely on a similar input data configuration with decomposed I and Q inputs, and the same approach has been employed for linearizing a load-modulated balanced PA [10], beamforming, or MIMO transmitters [11], [12], [13], for joint DPD and PAPR reduction [14], or self-interference cancellation in fullduplex radio [15], [16]. Recurrent NN (RNN) structures have also been studied as an alternative to feed-forward NNs, suited especially for strong PA memory effects [17], [18], [19], [20]. RNNs are, however, complex to train, since the recurrent structures need to be unrolled to ensure temporal consistency during training.…”