Search citation statements
Paper Sections
Citation Types
Year Published
Publication Types
Relationship
Authors
Journals
The Arithmetic Optimization Algorithm (AOA) is a recently proposed metaheuristic algorithm that has been shown to perform well in several benchmark tests. The AOA is a metaheuristic that uses the main arithmetic operators' distribution behavior, such as multiplication, division, subtraction, and addition. This paper proposes a binary version of the Arithmetic Optimization Algorithm (BAOA) to tackle the feature selection problem in classification. The algorithm's search space is converted from a continuous to a binary one using the sigmoid transfer function to meet the nature of the feature selection task. The classifier uses a method known as the wrapper-based approach K-Nearest Neighbors (KNN), to find the best possible solutions. This study uses 18 benchmark datasets from the University of California, Irvine (UCI) repository to evaluate the suggested binary algorithm's performance. The results demonstrate that BAOA outperformed the Binary Dragonfly Algorithm (BDF), Binary Particle Swarm Optimization (BPSO), Binary Genetic Algorithm (BGA), and Binary Cat Swarm Optimization (BCAT) when various performance metrics were used, including classification accuracy, selected features as well as the best and worst optimum fitness values.
The Arithmetic Optimization Algorithm (AOA) is a recently proposed metaheuristic algorithm that has been shown to perform well in several benchmark tests. The AOA is a metaheuristic that uses the main arithmetic operators' distribution behavior, such as multiplication, division, subtraction, and addition. This paper proposes a binary version of the Arithmetic Optimization Algorithm (BAOA) to tackle the feature selection problem in classification. The algorithm's search space is converted from a continuous to a binary one using the sigmoid transfer function to meet the nature of the feature selection task. The classifier uses a method known as the wrapper-based approach K-Nearest Neighbors (KNN), to find the best possible solutions. This study uses 18 benchmark datasets from the University of California, Irvine (UCI) repository to evaluate the suggested binary algorithm's performance. The results demonstrate that BAOA outperformed the Binary Dragonfly Algorithm (BDF), Binary Particle Swarm Optimization (BPSO), Binary Genetic Algorithm (BGA), and Binary Cat Swarm Optimization (BCAT) when various performance metrics were used, including classification accuracy, selected features as well as the best and worst optimum fitness values.
Deep Reinforcement Learning (DRL) allows agents to make decisions in a specific environment based on a reward function, without prior knowledge. Adapting hyperparameters significantly impacts the learning process and time. Precise estimation of hyperparameters during DRL training poses a major challenge. To tackle this problem, this study utilizes Grey Wolf Optimization (GWO), a metaheuristic algorithm, to optimize the hyperparameters of the Deep Deterministic Policy Gradient (DDPG) algorithm for achieving optimal control strategy in two simulated Gymnasium environments provided by OpenAI. The ability to adapt hyperparameters accurately contributes to faster convergence and enhanced learning, ultimately leading to more efficient control strategies. The proposed DDPG-GWO algorithm is evaluated in the 2DRobot and MountainCarContinuous simulation environments, chosen for their ease of implementation. Our experimental results reveal that optimizing the hyperparameters of the DDPG using the GWO algorithm in the Gymnasium environments maximizes the total rewards during testing episodes while ensuring the stability of the learning policy. This is evident in comparing our proposed DDPG-GWO agent with optimized hyperparameters and the original DDPG. In the 2DRobot environment, the original DDPG had rewards ranging from -150 to -50, whereas, in the proposed DDPG-GWO, they ranged from -100 to 100 with a running average between 1 and 800 across 892 episodes. In the MountainCarContinuous environment, the original DDPG struggled with negative rewards, while the proposed DDPG-GWO achieved rewards between 20 and 80 over 218 episodes with a total of 490 timesteps.
This research article presents the Multi-Objective Hippopotamus Optimizer (MOHO), a unique approach that excels in tackling complex structural optimization problems. The Hippopotamus Optimizer (HO) is a novel approach in meta-heuristic methodology that draws inspiration from the natural behaviour of hippos. The HO is built upon a trinary-phase model that incorporates mathematical representations of crucial aspects of Hippo’s behaviour, including their movements in aquatic environments, defense mechanisms against predators, and avoidance strategies. This conceptual framework forms the basis for developing the multi-objective (MO) variant MOHO, which was applied to optimize five well-known truss structures. Balancing safety precautions and size constraints concerning stresses on individual sections and constituent parts, these problems also involved competing objectives, such as reducing the weight of the structure and the maximum nodal displacement. The findings of six popular optimization methods were used to compare the results. Four industry-standard performance measures were used for this comparison and qualitative examination of the finest Pareto-front plots generated by each algorithm. The average values obtained by the Friedman rank test and comparison analysis unequivocally showed that MOHO outperformed other methods in resolving significant structure optimization problems quickly. In addition to finding and preserving more Pareto-optimal sets, the recommended algorithm produced excellent convergence and variance in the objective and decision fields. MOHO demonstrated its potential for navigating competing objectives through diversity analysis. Additionally, the swarm plots effectively visualize MOHO’s solution distribution of MOHO across iterations, highlighting its superior convergence behaviour. Consequently, MOHO exhibits promise as a valuable method for tackling complex multi-objective structure optimization issues.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.