The model is divided into two parts: an encoder and a decoder. The encoder consists of a CNN and a Transformer. We use Res2Net and Multi-spectral Channel Attention as the...
Abstract:
To address the issue of insufficient feature extraction in handwritten font generation models concerning stroke position, tilt angle, and stroke thickness, which leads to...Show MoreMetadata
Abstract:
To address the issue of insufficient feature extraction in handwritten font generation models concerning stroke position, tilt angle, and stroke thickness, which leads to significant discrepancies in details between generated fonts and real handwritten fonts, we propose a handwritten font generation model, MCA2Net-HFG. The model is based on Residual Resolution Network(Res2Net) and Multi-spectral Channel Attention (MCA) as the main structure of the encoder. We use Res2Net as the backbone network of the Convolutional Neural Networks(CNN) encoder to capture font features at different scales, such as stroke thickness, curvature, and position, as well as other fine-grained characteristics. Additionally, incorporating Multi-spectral Channel Attention into Res2Net captures key feature information located at different frequencies within the channels, enabling the model to better learn important characteristics of fonts during the training phase. This ensures the authenticity of the generated handwritten fonts. Experimental results on the IAM dataset show that compared to the current best methods, MCA2Net-HFG reduces the Fréchet Inception Distance(FID) score by 18.60% and the Geometry Score(GS) score by 50.59%. This indicates that MCA2Net-HFG generates higher quality handwritten fonts and has greater practical value. Additionally, results on different test sets validate the model’s good generalization performance.
The model is divided into two parts: an encoder and a decoder. The encoder consists of a CNN and a Transformer. We use Res2Net and Multi-spectral Channel Attention as the...
Published in: IEEE Access ( Volume: 12)
Funding Agency:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Channel Attention ,
- Convolutional Neural Network ,
- Feature Information ,
- Generalization Performance ,
- Tilt Angle ,
- Fine-grained Features ,
- Inception Distance ,
- Fréchet Inception Distance ,
- Good Generalization Performance ,
- Training Set ,
- Decrease In Values ,
- Feature Maps ,
- Long Short-term Memory ,
- Recurrent Neural Network ,
- Sequence Features ,
- Receptive Field ,
- Generative Adversarial Networks ,
- Image Generation ,
- Multi-scale Features ,
- Content Features ,
- Discrete Cosine Transform ,
- Transformer Encoder ,
- Word Error Rate ,
- Global Average Pooling ,
- Channel Attention Mechanism ,
- Style Features ,
- Multi-scale Feature Extraction ,
- Attention Vector ,
- One-hot Encoding ,
- Frequency Components
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Channel Attention ,
- Convolutional Neural Network ,
- Feature Information ,
- Generalization Performance ,
- Tilt Angle ,
- Fine-grained Features ,
- Inception Distance ,
- Fréchet Inception Distance ,
- Good Generalization Performance ,
- Training Set ,
- Decrease In Values ,
- Feature Maps ,
- Long Short-term Memory ,
- Recurrent Neural Network ,
- Sequence Features ,
- Receptive Field ,
- Generative Adversarial Networks ,
- Image Generation ,
- Multi-scale Features ,
- Content Features ,
- Discrete Cosine Transform ,
- Transformer Encoder ,
- Word Error Rate ,
- Global Average Pooling ,
- Channel Attention Mechanism ,
- Style Features ,
- Multi-scale Feature Extraction ,
- Attention Vector ,
- One-hot Encoding ,
- Frequency Components
- Author Keywords