All rights reserved. No part of this periodical may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying, recording, or any information storage and retrieval system, without permission in writing from the publisher. Permission for printing and for reprinting the material contained herein has been obtained by the publisher. Effects of Near Soil Surface Characteristics on Soil Detachment by Overland Flow in a Natural Succession Grassland Soil & Water Management & Conservation S oil detachment was defined by Govers et al. (1990) as the dislodgment of soil particles from the soil mass at a particular location on the soil surface by the erosive forces of rainfall or surface flow of water. Soil detachment by overland flow, such as rill erosion, is generally considered as the most important process of sediment production and contributes greatly to total soil loss (Poesen et al., 2003). It occurs when the effect of flowing water exceeds the certain threshold of soil resistance (Knapen et al., 2007), which implies that the erosive forces of overland flow and the resistance of the topsoil mass are two factors influencing soil detachment in rills. Overland flow is the external driving force for rill formation and development. Both laboratory and field experiments have found that hydraulic parameters of flowing water were closely related to the process of soil detachment by overland
This paper proposes an efficient memory transformer Emformer for low latency streaming speech recognition. In Emformer, the longrange history context is distilled into an augmented memory bank to reduce self-attention's computation complexity. A cache mechanism saves the computation for the key and value in self-attention for the left context. Emformer applies a parallelized block processing in training to support low latency models. We carry out experiments on benchmark LibriSpeech data. Under average latency of 960 ms, Emformer gets WER 2.50% on test-clean and 5.62% on test-other. Comparing with a strong baseline augmented memory transformer (AM-TRF), Emformer gets 4.6 folds training speedup and 18% relative real-time factor (RTF) reduction in decoding with relative WER reduction 17% on test-clean and 9% on test-other. For a low latency scenario with an average latency of 80 ms, Emformer achieves WER 3.01% on test-clean and 7.09% on test-other. Comparing with the LSTM baseline with the same latency and model size, Emformer gets relative WER reduction 9% and 16% on test-clean and testother, respectively.
Transformer-based acoustic modeling has achieved great success for both hybrid and sequence-to-sequence speech recognition. However, it requires access to the full sequence, and the computational cost grows quadratically with respect to the input sequence length. These factors limit its adoption for streaming applications. In this work, we proposed a novel augmented memory self-attention, which attends on a short segment of the input sequence and a bank of memories. The memory bank stores the embedding information for all the processed segments. On the librispeech benchmark, our proposed method outperforms all the existing streamable transformer methods by a large margin and achieved over 15% relative error reduction, compared with the widely used LC-BLSTM baseline. Our findings are also confirmed on some large internal datasets.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.