“…These views shed light on the potential adaptive value of cost-avoidance mechanisms.Keywords: cognitive effort, information theory, active inference, predictive coding, efficient coding, computational neuroscience Recent advances in artificial intelligence and computational neuroscience have led to formalization of cognition as a bounded rationality process (Friston, 2010;Kingma and Welling, 2013;Ortega and Braun, 2013;Tishby et al, 2000;Tkačik and Bialek, 2014).According to this view, rather than aiming systematically at the optimal solution to computational problems, cognitive processes trade off performance with computational costs.Remarkably, the way computational costs are formalized across these different studies is very consistent, despite their different approaches. In fact, whether one starts from an inference problem, in which the evidence for a model is maximized given some data (Genewein et al, 2015;Kingma and Welling, 2013;Tishby et al, 2000), or whether one is more generally attempting to minimize the entropy of future states (Friston, 2010), or whether one takes a decision making perspective, in which expected utility is maximized (Ortega et al, 2015), or even from the point of view of thermodynamics (Ortega and Braun, 2013;Sengupta et al, 2013), information cost is framed as a measure of divergence between an initial belief (or prior probability distribution over a variable of interest x, such as expected reward) and an updated belief (or posterior probability distribution over the same variable x) obtained after receiving new data. This measure of difference between probability distributions, called the Kullback-Leibler (KL) divergence, represents the amount of information one needs to collect in order to update the prior to the posterior ( for probability distributions P and Q).…”