Enhancing realism in handwritten text images with generative adversarial networks

Parul Dubey, Manjushree Nayak, Hitesh Gehani, Ashwini Kukade, Vinay Keswani, Pushkar Dubey

Abstract


Image synthesis is particularly important for applications that want to create realistic handwritten documents, which is why handwritten text generation is a critical area within its domain. Even with today's highly advanced technology, generating diverse and accurate representations of human handwriting is still a tough problem because of the variability in style. In this study, we tackle the problem of instability during the training phase of generative adversarial networks (GANs) for generating handwritten text images. Using the MNIST dataset, which includes 60,000 training and 10,000 test images of handwritten digits, we trained a GAN model to generate synthetic handwritten images. The methodology involves optimizing both the generator and discriminator using adversarial training, binary cross-entropy loss, as well as the optimizer Adam. A brand-new decaying learning rate schedule was introduced to speed up convergence. Performance was evaluated using the Fréchet inception distance (FID) metric. The results show that this model effectively generated high-quality synthetic images of handwritten digits, which resembled real data closely in the face of it all and also that there was a steady reduction in FID scores across epochs indicating improved performance.

Keywords


Discriminator; Generative adversarial network; Generator; Handwritten text; Image generation; Neural network

Full Text:

PDF


DOI: https://doi.org/10.11591/eei.v14i3.9190

Refbacks

  • There are currently no refbacks.


Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

Bulletin of EEI Stats

Bulletin of Electrical Engineering and Informatics (BEEI)
ISSN: 2089-3191, e-ISSN: 2302-9285
This journal is published by the Institute of Advanced Engineering and Science (IAES) in collaboration with Intelektual Pustaka Media Utama (IPMU).