Next Article in Journal
The Positive Effects on Volleyball Receiving Skills When Training with Lighter Balls
Next Article in Special Issue
Packet Loss Concealment Based on Phase Correction and Deep Neural Network
Previous Article in Journal
An Efficient Reliability Method with Multiple Shape Parameters Based on Radial Basis Function
Previous Article in Special Issue
Code-Switching in Automatic Speech Recognition: The Issues and Future Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Entropy-Based Dynamic Rescoring with Language Model in E2E ASR Systems

Department of Electrical Engineering and Information Systems, Graduate School of Engineering, The University of Tokyo, Tokyo 113-8656, Japan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 9690; https://doi.org/10.3390/app12199690
Submission received: 26 August 2022 / Revised: 19 September 2022 / Accepted: 22 September 2022 / Published: 27 September 2022
(This article belongs to the Special Issue Advances in Speech and Language Processing)

Abstract

:
Language models (LM) have played crucial roles in automatic speech recognition (ASR), whether as an essential part of a conventional ASR system composed of an acoustic model and LM, or as an integrated model to enhance the performance of novel end-to-end ASR systems. With the development of machine learning and deep learning, language modeling has made great progress in natural language processing applications. In recent years, efforts have been made to leverage the advantages of novel LM to ASR. The most common way to apply an integration is still shallow fusion because it can be easily implemented by zero-overhead while obtaining significant improvement. Our method can further enhance the applicability of shallow fusion without hyperparameter tuning while maintaining similar performance.

1. Introduction

In conventional ASR systems [1] that introduce deep learning or not, LM is still an essential part of the whole system. In end-to-end models, LM is not necessary since an end-to-end model itself can decode the intermediate representations of input features into a word sequence.
However, even for an end-to-end model, there are still benefits to introducing LM into it. LM is useful for introducing extra corpora information. For example, LM can be pretrained on a huge mount of text data, such as the Wikipedia corpus, with a much higher training speed compared to an end-to-end model trained on same mount of text data with audio signal data to be processed. In another common situation, that speech corpus may not be sufficient. In this case, LM is even more useful. However, what should be noticed is that transcripts of speech corpora and pure text corpora may not come from the same data distribution. For example, speech corpora contain speech expressions, such as ah, em, etc., while text corpora contain more written expressions with words that are hardly used in speech.
With the development of machine learning and deep learning, language modeling has made great progress in natural language processing applications. LM has evolved from the conventional n-gram model to a transformer-based [2] end-to-end system [3,4,5] capable of training on tremendous corpora with much higher prediction accuracy. In recent years, efforts have been made to leverage the advantages of novel LM to ASR. Some of them focus on integrating novel LM into the ASR system by techniques such as CTC [6,7], while others focus on adapting LM to the task domain.
Because shallow fusion can be easily implemented by combining logits from LM and ASR while obtaining significant improvement, it is still the most common method for applying LM integration in practice. Our method is an extension of shallow fusion. With the estimation of model confidence, this method can automatically decide and dynamically change an LM weight of shallow fusion during decoding. Experiments on two corpora, Librispeech and TEDLIUM-2, validate that our method can reach similar performance compared to tuned LM weights.
This article will start with introducing conventional and end-to-end (E2E) [3] frameworks of ASR and how LM can be integrated into end-to-end systems, then explain our proposals and show the results of previous proposals and our proposals on LM integration. Finally, we will summarize the experiments’ results and introduce several future plans on LM integration.

2. Prerequisite

This section will briefly introduce several challenges of large-vocabulary continuous-speech recognition systems and techniques used in our implementation to cope with them.
There are two kinds of challenges of large-vocabulary continuous speech recognition systems, as the name implies: the large-vocabulary challenge, and the word segmentation challenge.
In the large-vocabulary challenge, the vocabulary for speech recognition may contain hundreds of thousands of words. So, it can be a heavy burden for a neural-network-based system to store and process such a huge amount of words. We introduce two techniques to settle this issue.

2.1. Word Embedding

Word embedding methods learn a real-valued vector representation for a predefined fixed sized vocabulary from a corpus of text. The learning process is either performed jointly with the neural network model on some task, such as document classification, or is an unsupervised process, using document statistics. Since the word embedding vector is a real-valued dense vector compared to a one-hot vector of the same word which is a sparse vector, the dimension of a word’s representation is compressed from hundreds of thousands to only hundreds. Thus, there will be much less data of the words’ representations left for storing and processing. In our experiments’ implementation, the word embedding is implemented by an embedding layer, which is a word embedding that is learned jointly with a neural network model on a specific natural language processing task. It requires that the document text be cleaned and prepared such that each word is one-hot encoded. The size of the vector space is specified as part of the model, such as 50, 100, or 300 dimensions. The vectors are initialized with small random numbers. The embedding layer is used on the front end of a neural network and is fitted in a supervised way using the backpropagation algorithm. This approach of learning an embedding layer requires a lot of training data and can be slow, but will learn an embedding targeted to both the specific text data and the NLP task.

2.2. Byte Pair Encoding

Since the word embedding models pre-trained on Wikipedia are either limited by vocabulary size or the frequency of word occurrences, rare words would never be captured, resulting in unknown <unk> tokens when occurring in the text. By using the BPE method, not only can this <unk> tokens issues be largely solved, but also the vocabulary size will greatly decrease from hundreds of thousands to hundreds. BPE is an original corpus compression technique and nowadays is used in state-of-the-art NLP systems as a subwords segmentation method. Its basic idea is to recursively finding the most frequent byte pair in the corpus, then add it to vocabulary.
  • Initialize vocabulary.
  • Represent each word in the corpus as a combination of the characters along with the special end of word token </w>.
  • Iteratively count character pairs in all tokens of the vocabulary.
  • Merge every occurrence of the most frequent pair, add the new character n-gram to the vocabulary.
  • Repeat step 3 until the desired number of merge operations are completed, or the desired vocabulary size is achieved (which is a hyperparameter).
The second challenge of large-vocabulary continuous speech recognition systems is “continuous”, which means the system should be capable of segmenting the speech into segmentations of words by itself but not relying on speech signals, such as an explicit pause in between words. In an end-to-end model, in order to cope with this challenge, CTC loss and the attention mechanism are introduced to relate each token in a output word sequence with its input features.

2.3. Attention

CTC explicitly learns the monotonic alignment between the speech feature and character transcription. CTC joint training helps the sequence-to-sequence attention to be monotonic, which is reasonable for an ASR task. First, the CTC layer receives the output sequence of the encoder X e . It then computes the probability p c t c ( Y | X e ) over an arbitrary alignment π between X e and Y, where π [ t ] is a character ID aligned to the t-th frame in X e as follows:
C = s o f t m a x ( X e W c t c + b c t c ) ,
p ( B ( π ) = Y | X e ) = t = 1 n s u b C [ t , π [ t ] ] ,
p c t c ( Y | X e ) = Σ p ( B ( π ) = Y | X e ) ,
where W c t c R d a t t × d c h a r , b c t c R d c h a r are learnable parameters, C R n s u b × d c h a r is the CTC output, and C [ t , π [ t ] ] is the probability of the alignment between the output character π [ t ] and the t-th frame in X e . The many-to-one mapping B ( π ) removes redundant symbols from the alignment π , for example, B(aa ϕ b) = ab, where ϕ is a blank symbol. One-to-many mapping B 1 projects a string into a set of strings with redundant symbols: B 1 ( Y ) = π | Y = B ( π ) .
Regarding attention [5], when creating a context for our next prediction, we should not put equal weight on the information we collect so far. This is a big waste of computing resources. We should create a context of what we are interested in. However, this focus can shift in time. Before, in the RNN, we make predictions based on the new input x and the historical output context h. For an attention-based system, we can look into the whole input x again for each step but it will be replaced with the attention. For each input feature x i , we train an FC layer with tanh output to score how important the feature i is for the current time step under the previous output context. Then, we normalize the score using a softmax function. The attention Z is the weighted output of the input features. The key point is that we introduce an extra step to mask out information about which we care less at the current step.

2.4. Connectionist Temporal Classification

Phones are the acoustic realization of phonemes. We can have many allophones for the same phonemes, for example, the phoneme /p/ in “pit” and “spit” are pronounced differently. For speech recognition, we collected corpora which are phonetically transcribed and time aligned (the start and the end times of each phone are marked). TIMIT is one popular corpus that contains utterances from 630 North American speakers. The audio clip is divided into frames. A phone occupies multiple frames. When decoding, both of these methods will be combined with beam search [8] to reduce the search time.

3. Related Works

Previous work [9,10,11,12,13,14,15,16,17,18,19,20,21,22,23] addressing the issue of utilizing unpaired text has proposed ways of integrating an external pretrained LM, trained on all of the text data, with the ASR model. The main LM integration approaches from past work have been referred to as shallow, deep, and cold fusion. The three approaches differ in two important criteria (Table 1).
Model integration: At what point in the ASR model’s computation should the LM be integrated? In deep and cold fusion, the external LM is fused directly into the ASR model by combining their hidden states, resulting in a single model with tight integration. In contrast, in shallow fusion, the LM and ASR models remain separate and only their scores are combined, similar to an ensemble. The shallow fusion score combination is also similar to the interpolation of acoustic and language models performed in traditional ASR.
Training integration (Figure 1): At what point in the ASR model’s training should the LM be integrated? Deep and shallow fusion use late integration, where both the ASR and LM models are trained separately and then combined, while cold fusion uses the external pretrained LM model from the very start of the ASR model training. An important point is that early training integration approaches are computationally costlier if either of the two models is frequently changing.
  • Shallow fusion: Rescoring candidates with LM scores.
  • Cold fusion: Ensemble learning with end-to-end systems and LM.
  • Deep fusion: Fine-tuning LM on speech transcripts then rescoring.

3.1. Shallow Fusion

Rescoring with Pretrained Language Model
The objective of ASR is to produce a token sequence based on the input feature sequence. In most of the cases, ASR systems produce output tokens (words) one by one. In a neural-network-based end-to-end system, the model would produce a probability distribution on vocabulary in each time step. So beam search can be applied to find the sequence with the highest probability.
However, in practice, beam search receives a score which is log-likelihood from the end-to-end model produced a probability distribution. Thus, introducing a language model into the beam-search process can be done by simply adding the log-likelihood of the language model output tokens to the original score.
In the equation of the rescoring method, s ( y | x ) is the final score of output tokens based on input features x, and β P e n a l t y ( | y | c ) is a penalty item which is a function of the output sequence length aiming at suppressing longer candidates since a longer one tends to produce more meaningless words, such as ah, em, etc. Additionally, α , β are hyper-parameters which are weights to determine the importance of each item in this equation.
s ( y | x ) = log ( P E 2 E ( y | x ) ) + α log ( P L M ( y ) ) + β P e n a l t y ( | y | c )

3.2. Rescoring with Pretrained LM

The objective of ASR is producing the most likely sequence of tokens based on the input feature sequence; in most of the cases, ASR systems produce output tokens (words) one by one. In a neural-network-based E2E system, after a history of recognized words, it would produce the probability distribution over the candidate tokens. To realize efficient search, beam search is often applied to find the token sequence with the highest probability.
In practice, however, beam search receives a score which is a likelihood of the E2E model’s output tokens. Thus, a LM can be integrated into the beam-search process by simply adding the likelihood of LM output tokens to the original score.
s ( y | x ) = log ( P E 2 E ( y | x ) ) + α log ( P L M ( y ) ) + β penalty ( | y | c )
In Equation (5), s ( y | x ) is the final score of output sequence y based on input features x, and β penalty ( | y | c ) is a penalty which is a function of the y length, aiming at suppressing longer candidates since longer sequences tend to include meaningless words, such as ah. α and β are hyper-parameters to control the importance of each term in this equation.

4. Our Proposal: Dynamic Rescoring Method

For simplicity, β penalty ( | y | c ) in Equation (5) can be included in log ( P E 2 E ( y | x ) ) or α log ( P L M ( y ) ) . Then, this rescoring equation becomes a simple linear combination in Equation (6). Here, λ is a hyperparameter, and score s ( y | x ) is a linear combination of the likelihood of the E2E model and that of the LM. In this approach, λ determines the importance of both models statically, which may be obtained by greedy search using development datasets.
However, it is easily assumed that the optimal value of λ depends on context and λ should be controlled dynamically and adaptively. By taking into account that any language model can be assessed by the entropy of the predicted tokens of the model, we propose a method of dynamic control of λ based on the entropies of P E 2 E ( y | x ) and P L M ( y ) .
As a model is better with smaller entropy, in Equation (7), λ is defined to quantify goodness or importance of the LM, which vary dynamically depending on x. Finaly, we summarize this method in the Algorithm 1.
s ( y | x ) = ( 1 λ ) log ( P E 2 E ( y | x ) ) + λ log ( P L M ( y ) )
λ = 1 H L M ( Y ) H E 2 E ( Y | X = x ) + H L M ( Y )
Algorithm 1: Calculate output token logits.
Data: s p e e c h , p r e v i o u s _ t o k e n s
Result: n e x t _ t o k e n
s 2 s _ n e x t _ t o k e n = s 2 s _ m o d e l ( s p e e c h , p r e v i o u s _ t o k e n s )
l m _ n e x t _ t o k e n = L M ( p r e v i o u s _ t o k e n s )
e n t r o p y _ s 2 s = e n t r o p y ( s 2 s _ n e x t _ t o k e n )
e n t r o p y _ l m = e n t r o p y ( l m _ n e x t _ t o k e n )
l m _ w e i g h t = 1.0 e n t r o p y _ l m / ( e n t r o p y _ s 2 s + e n t r o p y _ l m )
n e x t _ t o k e n = ( 1.0 l m _ w e i g h t ) * log ( s 2 s _ n e x t _ t o k e n ) + l m _ w e i g h t * log ( l m _ n e x t _ t o k e n )

Effectiveness Analysis

We conducted several experiments to analyze possible reasons for why our proposal works. Since perplexity is the common metric for language modeling, we decide to check the distributions of each utterance output’s entropy as an estimation of the perplexities of the original shallow fusion and our method. Figure 2 shows the perplexities of the two approaches. It tells that the dynamically weighting strategy indeed decreases the perplexity of the integration system.
We were also curious about how dynamic weight changes during testing, so we drew a figure of the distribution of dynamic weight during decoding. In Figure 3, there seems to be two peaks in this distribution, which implies different parts of the dev set prefer different λ s. One reason could be that the dev set is composed of different data from different domains. This phenomenon may be common for the production environment, which is complicated. By static weight in shallow fusion, this can never be captured.

5. Experiments

This section will focus on our experiments. Before showing any results of experiments, we will introduce the corpus we used in our experiments and our experiment conditions.

5.1. Corpus

LibriSpeech [24] and TED-LIUM2 [25] are the corpora we performed our experiments on. We took TED-LIUM2 corpus as the example to explain the structure of those corpora, while LibriSpeech is similar. It is built on TED talk videos. It contains four datasets, Train, Dev, Test, and LM, and a dictionary that contains 152,000 words. The Train, Dev, and Test datasets each includes an audio file and a transcription file. They collectively contain 1495 audio talks and 1495 aligned automatic transcripts. LM dataset is a text corpus which selects monolingual data for LM from WMT12 publicly available corpora.
Figure 4 shows the format of transcript files from TED-LIUM2. Each line in the transcripts corresponds to a certain segmentation of audio file. So, we care about the start time, end time, and text columns of the transcripts.

5.2. Metrics

There are several metrics used in our experiments: word error rate (WER), word accuracy (WAcc), substitution error, deletion error and insertion error.
WER is built based on edit distance, which means the fewest number of modifications to make two sequences the same. One sequence is the reference sequence, and the other one is hypothesis sequence. By substituting, deleting, and inserting certain amounts of tokens in the hypothesis sequence, the hypothesis sequence can always be the same as the reference one. So, WER basically means how few token corrections are necessary by those operations. W A c c = 1 W E R .
W E R = # S u b s t i t u t i o n + # D e l e t i o n + # I n s e r t i o n # S u b s t i t u t i o n + # D e l e t i o n + # C o r r e c t n e s s

5.3. Experimental Settings

We adopt a Transformer-based ASR system comprised of 6 encoder blocks and 6 decoder blocks with the feed-forward inner dimension of 2048, the model dimension of 256, and the attention head number 4 (Figure 5), which are unchanged in all experiments. The input features were 240-dimensional log Mel-filterbank energy features (80-dim static, + Δ , and + Δ Δ ). The feature is extracted with a 10 ms frameshift of a 25 ms window. Each feature was mean- and variance-normalized per speaker, and every four frames were spliced (three left, one current, and zero right). The low and high cutoff frequencies were set to 20 Hz and 8000 Hz, respectively. Speed perturbation was not used in the fine-tuning stage. We then subsampled the input features every three frames. The model was jointly trained with CTC (weight α = 0.2 ). Both the CTC target and S2S target are byte pair encoding (BPE) tokens trained on train set. The “noam” optimizer was used with 25,000 warm-up steps and an initial learning rate of 5. The model was trained with the ESPnet toolkit [3] using a batch size of 32 for 30 epochs on an 11 GB GTX1080 TI GPU.

5.4. Results

There are three tasks in our experiments. Task one is for validation of effectiveness of the rescoring method and building a baseline for following tasks, which is conducted on LibriSpeech. Task two shows the result of proposed rescoring method, which is conducted on LibriSpeech. Task three shows the robustness of our proposal on different corpora while considering different development sets, which is conducted on LibriSpeech and TED-LIUM2.

5.5. Task One (Baseline)

Objective: Comparing results of E2E model with or without LM while using rescoring.
Implementation details: The E2E model is composed of a transformer encoder network and a transformer decoder network. The loss function of the E2E model is composed of attention loss and CTC loss. LM is composed of a transformer encoder network and a full connection network as decoder. Its loss function is attention loss. The ASR system decodes with the rescoring method described in Equation (5).
As the Result shown in Table 2:

5.6. Task Two (Our Proposal)

Objective: Improving rescoring method.
Proposal: To balance E2E model and LM by their entropy. In the original rescoring method, lambda is a hyper-parameter, and score s ( s | x ) in Equation (5) is a linear combination of log-likelihood of E2E model and LM. The proposal is to introduce the entropy of the model prediction’s distribution into λ in Equation (7).
As the Result shown in Table 3:
The above result shows that our rescoring slightly exceeds the original rescoring method. However, one important advantage of our method is that it does not require tuning the weight of language model in the rescoring function. Additionally, it brings few overhead.
The result shows that our proposal is better than the original rescoring method with the same LM.

5.7. Task Three (Robustness)

In this task, we ran experiments on two corpora of TED-LIUM2 [25] and LibriSpeech [24]. By these experiments, we tried to prove (1) the effectiveness of our proposal separately for each corpus and (2) the robustness of this method across different development datasets, compared to static control of λ . Ref. [3] was used as the baseline E2E ASR system.
Figure 6 and Figure 7 show the word accuracies with TED-LIUM2 and LibriSpeech, respectively. In both figures, the blue curve shows the word accuracies when λ is controlled statically in Equation (6). In Figure 6, the optimal value is 0.2, and in Figure 7, it is 0.4. The orange line is the word accuracy with our proposal. In both figures, we can claim that the performance with our proposal can approximate the highest performance with static weights, the optimal values of which are different between the two corpora. It should be noted that in our proposal based on the model entropy, very good performance is realized with no effort for searching for the optimal value of λ and with only little overhead, which is the entropy calculation of O (vocabulary size). A good enough LM weight can be automatically obtained.

5.8. Comparison with Original Shallow Fusion

The most severe deficit of static weights is that the development dataset used to optimize the weights may have a different distribution than the testing data or real-world data. To simulate this situation, we ran another set of experiments with different development datasets for fixing static weights and those with dynamic weights.
We split the development set from TED-LIUM2 into two, denoted as Dev 1 and Dev 2. Table 4 shows that the optimal static weight depends on the development data used, and it is 0.30 for Dev 1 and 0.10 for Dev 2. If the static weight is determined with Dev 2 as 0.10, according to Figure 6, the ASR performance will be slightly lower than our proposal, indicated by the orange line.

6. Conclusions

We introduced the idea of dynamically rescoring in LM integration and proposed a method based on entropy of models. This method calculates the model’s entropy on the fly and gives the model more importance when it has lower entropy so that the overall system entropy can be improved. This method is easy to implement with little overhead. Most importantly, it shows the feasibility of dynamically rescoring because it can reach a similar performance as what a tuned hyperparameter provides. In future work, we will try to introduce more delicate mechanisms into this approach to enhance it so that it can suppress all of the static weights. We would like to apply it to multilingual ASR since tuning static weights for each language in multilingual ASR is much more expensive, and our method could be more advantageous.

Author Contributions

Conceptualization, Z.G.; Data curation, Z.G.; Formal analysis, Z.G.; Funding acquisition, N.M.; Investigation, Z.G.; Methodology, Z.G.; Project administration, Z.G.; Resources, Z.G. and N.M.; Software, Z.G.; Supervision, D.S. and N.M.; Validation, Z.G., D.S. and N.M.; Visualization, Z.G.; Writing—original draft, Z.G.; Writing—review & editing, Z.G. and N.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

LibriSpeech ASR corpus: https://www.openslr.org/12; TED-LIUMv2: https://openslr.magicdatatech.com/19/ (accessed on 21 September 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Juang, B.H.; Rabiner, L.R. Automatic Speech Recognition—A Brief History of the Technology Development; Georgia Institute of Technology; Atlanta Rutgers University and the University of California: Santa Barbara, CA, USA, 2005. [Google Scholar]
  2. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. (NerIPS) 2017, 30, 5998–6008. [Google Scholar]
  3. Watanabe, S.; Hori, T.; Karita, S.; Hayashi, T.; Nishitoba, J.; Unno, Y.; Soplin, N.E.Y.; Heymann, J.; Wiesner, M.; Chen, N.; et al. ESPnet: End-to-end speech processing toolkit. arXiv 2018, arXiv:1804.00015. [Google Scholar]
  4. Chan, W.; Jaitly, N.; Le, Q.; Vinyals, O. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 4960–4964. [Google Scholar]
  5. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  6. Graves, A.; Fernández, S.; Gomez, F.; Schmidhuber, J. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006. [Google Scholar]
  7. Kim, S.; Hori, T.; Watanabe, S. Joint CTC-attention based end-to-end speech recognition using multi-task learning. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017. [Google Scholar]
  8. Drexler, J.; Glass, J. Subword regularization and beam search decoding for end-to-end automatic speech recognition. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019. [Google Scholar]
  9. Zeineldeen, M.; Glushko, A.; Michel, W.; Zeyer, A.; Schlüter, R.; Ney, H. Investigating methods to improve language model integration for attention-based encoder-decoder asr models. arXiv 2021, arXiv:2104.05544. [Google Scholar]
  10. Narisetty, C.P.; Hayashi, T.; Ishizaki, R.; Watanabe, S.; Takeda, K. Leveraging State-of-the-art ASR Techniques to Audio Captioning. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2021, Online, 15–19 November 2021. [Google Scholar]
  11. McDermott, E.; Sak, H.; Variani, E. A Density Ratio Approach to Language Model Fusion in End-to-End Automatic Speech Recognition. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017. [Google Scholar]
  12. Andrés-Ferrer, J.; Albesano, D.; Zhan, P.; Vozila, P. Contextual density ratio for language model biasing of sequence to sequence ASR systems. arXiv 2022, arXiv:2206.14623. [Google Scholar]
  13. Toshniwal, S.; Kannan, A.; Chiu, C.-C.; Wu, Y.; Sainath, T.N.; Livescu, K. A comparison of techniques for language model integration in encoder-decoder speech recognition. In Proceedings of the 2018 IEEE Spoken Language Technology Workshop (SLT), Athens, Greece, 18–21 December 2018. [Google Scholar]
  14. Zhao, D.; Sainath, T.N.; Rybach, D.; Rondon, P.; Bhatia, D.; Li, B.; Pang, R. Shallow-Fusion End-to-End Contextual Biasing. In Proceedings of the Interspeech, Graz, Austria, 15–19 September 2019. [Google Scholar]
  15. Shan, C.; Weng, C.; Wang, G.; Su, D.; Luo, M.; Yu, D.; Xie, L. Component fusion: Learning replaceable language model component for end-to-end speech recognition system. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019. [Google Scholar]
  16. Sriram, A.; Jun, H.; Satheesh, S.; Coates, A. Cold fusion: Training seq2seq models together with language models. arXiv 2017, arXiv:1708.06426. [Google Scholar]
  17. Battenberg, E.; Chen, J.; Child, R.; Coates, A.; Li, Y.G.Y.; Liu, H.; Satheesh, S.; Sriram, A.; Zhu, Z. Exploring neural transducers for end-to-end speech recognition. In Proceedings of the 2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), Okinawa, Japan, 16–20 December 2017. [Google Scholar]
  18. Wang, D.; Zheng, T.F. Transfer learning for speech and language processing. In Proceedings of the 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Hong Kong, China, 16–19 December 2015. [Google Scholar]
  19. Zhao, W.X.; Jiang, J.; Weng, J.; He, J.; Lim, E.-P.; Yan, H.; Li, X. Comparing twitter and traditional media using topic models. In European Conference on Information Retrieval; Springer: Berlin/Heidelberg, Germanny, 2011. [Google Scholar]
  20. Xing, C.; Wu, W.; Wu, Y.; Liu, J.; Huang, Y.; Zhou, M.; Ma, W.Y. Topic aware neural response generation. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  21. Kuhn, R.; De Mori, R. A cache-based natural language model for speech recognition. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 570–583. [Google Scholar] [CrossRef]
  22. Sennrich, R.; Haddow, B.; Birch, A. Improving Neural Machine Translation Models with Monolingual Data. arXiv 2016, arXiv:1511.06709. [Google Scholar]
  23. Renduchintala, A.; Ding, S.; Wiesner, M.; Watanabe, S. Multi-Modal Data Augmentation for End-to-End ASR. arXiv 2018, arXiv:1803.10299. [Google Scholar]
  24. Panayotov, V.; Chen, G.; Povey, D.; Khudanpur, S. Librispeech: An asr corpus based on public domain audio books. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 5206–5210. [Google Scholar]
  25. Rousseau, A.; Deléglise, P.; Esteve, Y. TED-LIUM: An Automatic Speech Recognition dedicated corpus. In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC’12), Istanbul, Turkey, 21–27 May 2012; pp. 125–129. [Google Scholar]
Figure 1. LM integration approach structure comparison. The left side is shallow fusion, while the right side is deep and cold fusion.
Figure 1. LM integration approach structure comparison. The left side is shallow fusion, while the right side is deep and cold fusion.
Applsci 12 09690 g001
Figure 2. Perplexity of decoding output from the original shallow fusion and dynamic rescoring.
Figure 2. Perplexity of decoding output from the original shallow fusion and dynamic rescoring.
Applsci 12 09690 g002
Figure 3. Dynamic weight distribution on dev set.
Figure 3. Dynamic weight distribution on dev set.
Applsci 12 09690 g003
Figure 4. TED-LIUM2 transcript format.
Figure 4. TED-LIUM2 transcript format.
Applsci 12 09690 g004
Figure 5. Transformer-based E2E ASR model jointly trained with CTC loss.
Figure 5. Transformer-based E2E ASR model jointly trained with CTC loss.
Applsci 12 09690 g005
Figure 6. Results on TED-LIUM2.
Figure 6. Results on TED-LIUM2.
Applsci 12 09690 g006
Figure 7. Results on LibriSpeech.
Figure 7. Results on LibriSpeech.
Applsci 12 09690 g007
Table 1. LM integration approach comparison.
Table 1. LM integration approach comparison.
Integration ApproachIntegration StructureIntegration Time
Shallow FusionScore InterpolationDecoding Stage
Deep FusionModel CombinationDecoding Stage
Cold FusionModel CombinationTraining Stage
Table 2. Task One.
Table 2. Task One.
Integration MethodWAccSub ErrorDel ErrorIns Error
No LM88.39.81.91.4
Ori. Rescoring90.56.72.81.0
Table 3. Task Two.
Table 3. Task Two.
Integration MethodWAccSub ErrorDel ErrorIns Error
No LM88.39.81.91.4
Ori. Rescoring90.56.72.81.0
Our Rescoring90.77.51.81.0
Table 4. Results with different dev sets.
Table 4. Results with different dev sets.
LM WeightDev 1Dev 2
0.0088.390.2
0.1089.190.9
0.3089.790.5
0.3389.090.4
0.4088.789.5
0.5082.486.2
0.6761.065.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gong, Z.; Saito, D.; Minematsu, N. Entropy-Based Dynamic Rescoring with Language Model in E2E ASR Systems. Appl. Sci. 2022, 12, 9690. https://doi.org/10.3390/app12199690

AMA Style

Gong Z, Saito D, Minematsu N. Entropy-Based Dynamic Rescoring with Language Model in E2E ASR Systems. Applied Sciences. 2022; 12(19):9690. https://doi.org/10.3390/app12199690

Chicago/Turabian Style

Gong, Zhuo, Daisuke Saito, and Nobuaki Minematsu. 2022. "Entropy-Based Dynamic Rescoring with Language Model in E2E ASR Systems" Applied Sciences 12, no. 19: 9690. https://doi.org/10.3390/app12199690

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop