“…While this finding might seem trivial at first, it has far-reaching consequences, as it brings forward questions about the reproducibility of inferences about the mapping between brain activity and cognitive states that are drawn from the interpretation of cognitive decoding decisions of DL models. Recent empirical work in DL research has demonstrated that the convergence of DL models, and thereby the specifics of their learned mapping between input data and target signal, is dependent on many non-deterministic aspects of the training process, such as random seeds and random weight initializations (Dodge et al, 2019, Henderson et al, 2018, Lucic et al, 2018, Reimers and Gurevych, 2017 as well as the specific choices for other hyper-parameters, such as individual layer specifications and optimization algorithms Lucic et al (2018), Melis et al (2017), Zoph and Le (2017). It is thus possible that the mapping between cognitive states and brain activity that a DL model learns can change with these factors of the training.…”