“…Moreover, to address the challenges of gradient explosion and gradient disappearance, the gate recurrent unit (GRU) is used to learn time series information, as it offers computational efficiency compared to the long short-term memory (LSTM) [34]. Furthermore, the phases of actions are incorporated into the network construction to make the feedback model dependent on the evolution of phases [1], enabling improved scalability of the skill model in the time domain. Fig.…”