TY - JOUR
T1 - Enhanced Text-to-Image Synthesis With Self-Supervision
AU - Tan, Yong Xuan
AU - Lee, Chin Poo
AU - Neo, Mai
AU - Lim, Kian Ming
AU - Lim, Jit Yan
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2023
Y1 - 2023
N2 - The task of Text-to-Image synthesis is a difficult challenge, especially when dealing with low-data regimes, where the number of training samples is limited. In order to address this challenge, the Self-Supervision Text-to-Image Generative Adversarial Networks (SS-TiGAN) has been proposed. The method employs a bi-level architecture, which allows for the use of self-supervision to increase the number of training samples by generating rotation variants. This, in turn, maximizes the diversity of the model representation and enables the exploration of high-level object information for more detailed image construction. In addition to the use of self-supervision, SS-TiGAN also investigates various techniques to address the stability issues that arise in Generative Adversarial Networks. By implementing these techniques, the proposed SS-TiGAN has achieved a new state-of-the-art performance on two benchmark datasets, Oxford-102 and CUB. These results demonstrate the effectiveness of the SS-TiGAN method in synthesizing high-quality, realistic images from text descriptions under low-data regimes.
AB - The task of Text-to-Image synthesis is a difficult challenge, especially when dealing with low-data regimes, where the number of training samples is limited. In order to address this challenge, the Self-Supervision Text-to-Image Generative Adversarial Networks (SS-TiGAN) has been proposed. The method employs a bi-level architecture, which allows for the use of self-supervision to increase the number of training samples by generating rotation variants. This, in turn, maximizes the diversity of the model representation and enables the exploration of high-level object information for more detailed image construction. In addition to the use of self-supervision, SS-TiGAN also investigates various techniques to address the stability issues that arise in Generative Adversarial Networks. By implementing these techniques, the proposed SS-TiGAN has achieved a new state-of-the-art performance on two benchmark datasets, Oxford-102 and CUB. These results demonstrate the effectiveness of the SS-TiGAN method in synthesizing high-quality, realistic images from text descriptions under low-data regimes.
KW - GAN
KW - generative adversarial networks
KW - generative model
KW - self-supervised learning
KW - Text-to-image synthesis
UR - http://www.scopus.com/inward/record.url?scp=85153801523&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2023.3268869
DO - 10.1109/ACCESS.2023.3268869
M3 - Article
AN - SCOPUS:85153801523
SN - 2169-3536
VL - 11
SP - 39508
EP - 39519
JO - IEEE Access
JF - IEEE Access
ER -