“…Shannon information entropy (H) is defined as an expected amount of information in a random variable [38], and is widely used, together with its successors, as a measure of information value in differ-ent fields, including chemistry, medicine [9,21,24], robotics and machine learning. In deep belief networks, the maximum entropy learning algorithm provides better generalization capability than the maximum likelihood learning approach, ensuring a less biased distribution and robust to over-fitting predictive model [23].…”