The Blockformer speech recognition model has recently been proposed as a state-of-the-art (SOTA) model ontheAishell-1 Chinese speech dataset. This model exhibited significant improvements in character error rate (CER) when compared to its baseline, Conformer. The key improvement of Blockformer is the addition of the Squeeze-and-Excitation (SE) block on top of Conformer, which enables better utilization of the information contained in each Conformer block. In our study of Blockformer, we identified scope for improving its block information extraction method. To this end, we used the attention mechanism to enhance the SE block's efficacy in squeezing block information. And we enhanced the model's structure in attention inference mode to align more effectively with the training structure. Under the four inference modes, namely attention, attention rescoring, ctc greedy search, and ctc prefix beam search, the CER reaches 4.67%, 4.43%, 4.75% and 4.75%. All of these rates are at the level of Blockformer or exceed it.