Text-to-image synthesis with self-supervised bi-stage generative adversarial network

Yong Xuan Tan*, Chin Poo Lee, Mai Neo, Kian Ming Lim, Jit Yan Lim

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

5 Citations (Scopus)

Abstract

Text-to-image synthesis is challenging as generating images that are visually realistic and semantically consistent with the given text description involves multi-modal learning with text and image. To address the challenges, this paper presents a text-to-image synthesis model that utilizes self-supervision and bi-stage image distribution architecture, referred to as the Self-Supervised Bi-Stage Generative Adversarial Network (SSBi-GAN). The self-supervision diversifies the learned representation thus improving the quality of the synthesized images. Besides that, the bi-stage architecture with Residual network enables the generation of larger images with finer visual contents. Not only that, some enhancements including L1 distance, one-sided smoothing and feature matching are incorporated to enhance the visual realism and semantic consistency of the images as well as the training stability of the model. The empirical results on Oxford-102 and CUB datasets corroborate the ability of the proposed SSBi-GAN in generating visually realistic and semantically consistent images.

Original languageEnglish
Pages (from-to)43-49
Number of pages7
JournalPattern Recognition Letters
Volume169
DOIs
Publication statusPublished - May 2023
Externally publishedYes

Keywords

  • GAN
  • Generative adversarial network
  • Self-supervised learning
  • Text-to-image-synthesis

Fingerprint

Dive into the research topics of 'Text-to-image synthesis with self-supervised bi-stage generative adversarial network'. Together they form a unique fingerprint.

Cite this