“…Deep learning models typically function as black boxes offering very little insight into their decision making mechanics. To expose model understanding at various depths, researchers have proposed various structural probing (Tenney et al, 2018;Hewitt and Manning, 2019;Lin et al, 2019) and behavioral probing methods (McCoy et al, 2020;Goldberg, 2019;Warstadt et al, 2019;Ettinger, 2020), as well as input saliency maps to highlight the most important tokens/sentences in the input for each prediction (Serrano and Smith, 2019;Ribeiro et al, 2016;Swanson et al, 2020;Tenney et al, 2019), and input token relationships (Lamm et al, 2020). Alongside, there is work on producing textual rationales (Lei et al, 2016), which are snippets of NL to help explain model predictions.…”