Artificial neural network (ANN) has been commonly used to deal with many problems. However, since this algorithm applies backpropagation algorithms based on gradient descent (GD) technique to look for the best solution, the network may face major risks of being entrapped in local minima. To overcome those drawbacks of ANN, in this work, we propose a novel ANN working parallel with metaheuristic algorithms (MAs) to train the network. The core idea is that first, (1) GD is applied to increase the convergence speed. (2) If the network is stuck in local minima, the capacity of the global search technique of MAs is employed. (3) After escaping from local minima, the GD technique is applied again. This process is applied until the target is achieved. Additionally, to increase the efficiency of the global search capacity, a hybrid of particle swarm optimization and genetic algorithm (PSOGA) is employed. The effectiveness of ANNPSOGA is assessed using both numerical models and measurement. The results demonstrate that ANNPSOGA provides higher accuracy than traditional ANN, PSO, and other hybrid ANNs (even a higher level of noise is employed) and also considerably decreases calculational cost compared with PSO.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.