over a decade ago, the first comprehensive review on the state of benchmarking in Genetic Programming (GP) analyzed the mismatch between the problems that are used to test the performance of GP systems and real-world problems. Since then, several benchmark suites in major GP problem domains have been proposed over time, which were able to fill some of the major gaps. In the framework of the first review about the state of benchmarking in GP, logic synthesis was classified as one of the major GP problem domains. However, a diverse and accessible benchmark suite for logic synthesis is still missing in the field of GP. In this work, we take a first step towards a benchmark suite for logic synthesis that covers different types of Boolean functions that are commonly used for the evaluation of GP systems. We also present baseline results that have been obtained by former work and in our evaluation experiments by using Cartesian Genetic Programming.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.