Abstract:
Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate...Show MoreMetadata
Abstract:
Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate network architecture, gradient-based learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional neural networks, which are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation recognition, and language modeling. A new learning paradigm, called graph transformer networks (GTN), allows such multimodule systems to be trained globally using gradient-based methods so as to minimize an overall performance measure. Two systems for online handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of graph transformer networks. A graph transformer network for reading a bank cheque is also described. It uses convolutional neural network character recognizers combined with global training techniques to provide record accuracy on business and personal cheques. It is deployed commercially and reads several million cheques per day.
Published in: Proceedings of the IEEE ( Volume: 86, Issue: 11, November 1998)
DOI: 10.1109/5.726791
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Gradient-based Learning ,
- Transformer ,
- Convolutional Network ,
- Recognizable ,
- Backpropagation ,
- Object Detection ,
- Speech Recognition ,
- Word Level ,
- Linear Graph ,
- Gradient-based Methods ,
- Graphical Output ,
- Sequence Labeling ,
- End Nodes ,
- Symbol Sequence ,
- Input Field ,
- Input Graph ,
- Dollar Amount ,
- Discrimination Training ,
- Discriminant Criterion ,
- Beam Search ,
- Input Symbols ,
- Recognition System ,
- Whole Word ,
- Third Set Of Experiments ,
- Identical Output ,
- Graph Structure ,
- Pair Of Nodes ,
- Second Derivative ,
- Loss Function ,
- Composition Operator ,
- Handwritten Digit Recognition ,
- Training Set ,
- Error Rate ,
- Convolutional Layers ,
- Feature Maps ,
- Parameter Vector ,
- Trainable Parameters ,
- Linear Classifier ,
- Test Error ,
- Maximum A Posteriori ,
- Output Of Module ,
- Units In Layer ,
- Character Images ,
- Forward Algorithm ,
- Training Error ,
- Input Patterns ,
- Database Version ,
- Input Image
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Gradient-based Learning ,
- Transformer ,
- Convolutional Network ,
- Recognizable ,
- Backpropagation ,
- Object Detection ,
- Speech Recognition ,
- Word Level ,
- Linear Graph ,
- Gradient-based Methods ,
- Graphical Output ,
- Sequence Labeling ,
- End Nodes ,
- Symbol Sequence ,
- Input Field ,
- Input Graph ,
- Dollar Amount ,
- Discrimination Training ,
- Discriminant Criterion ,
- Beam Search ,
- Input Symbols ,
- Recognition System ,
- Whole Word ,
- Third Set Of Experiments ,
- Identical Output ,
- Graph Structure ,
- Pair Of Nodes ,
- Second Derivative ,
- Loss Function ,
- Composition Operator ,
- Handwritten Digit Recognition ,
- Training Set ,
- Error Rate ,
- Convolutional Layers ,
- Feature Maps ,
- Parameter Vector ,
- Trainable Parameters ,
- Linear Classifier ,
- Test Error ,
- Maximum A Posteriori ,
- Output Of Module ,
- Units In Layer ,
- Character Images ,
- Forward Algorithm ,
- Training Error ,
- Input Patterns ,
- Database Version ,
- Input Image