Loading [MathJax]/extensions/MathZoom.js
Improving the robustness of LSTMs for word classification using stressed word endings in dual-state word-beam search | IEEE Conference Publication | IEEE Xplore

Improving the robustness of LSTMs for word classification using stressed word endings in dual-state word-beam search


Abstract:

In recent years, long short-term memory neural networks (LSTMs) followed by a connectionist temporal classification (CTC) have shown strength in solving handwritten text ...Show More

Abstract:

In recent years, long short-term memory neural networks (LSTMs) followed by a connectionist temporal classification (CTC) have shown strength in solving handwritten text recognition problems. Such networks can handle not only sequence variability but also geometric variation by using a convolutional front end, at the input side. Although different approaches have been introduced for decoding activations in the CTC output layer, only limited consideration is given to the use of proper label-coding schemes. In this paper, we use a limited-size ensemble of end-to-end convolutional LSTM Neural Networks to evaluate four label-coding schemes. Additionally, we evaluate two CTC search techniques: Best-path search vs dual-state word-beam search (DSWBS). The classifiers in the ensemble have comparable architectures but variable numbers of hidden units. We tested the coding and search approaches on three datasets: A standard benchmark IAM dataset (English) and two more difficult historical handwritten datasets (diaries and field notes, highly multilingual). Results show that stressing the word endings in the label-coding scheme yields a higher performance, especially for DSWBS. However, stressing the start-of-word shapes with a token appears to be disadvantageous.
Date of Conference: 08-10 September 2020
Date Added to IEEE Xplore: 25 November 2020
ISBN Information:
Conference Location: Dortmund, Germany

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.