Digital Repository

Z-GAN For Enhancing Generative Adversarial Network Based T2I Synthesis Using Transformer Model

Show simple item record

dc.contributor.author Boyagoda, Yenuli
dc.date.accessioned 2024-03-29T05:44:19Z
dc.date.available 2024-03-29T05:44:19Z
dc.date.issued 2023
dc.identifier.citation Boyagoda, Yenuli (2023) Z-GAN For Enhancing Generative Adversarial Network Based T2I Synthesis Using Transformer Model. BSc. Dissertation, Informatics Institute of Technology en_US
dc.identifier.issn 2019794
dc.identifier.uri http://dlib.iit.ac.lk/xmlui/handle/123456789/1942
dc.description.abstract "This research thesis focuses on the task of using Generative Adversarial Networks (GANs) to produce realistic images from text descriptions. The primary goal of this work is to address the research gap in the field of text-to-image synthesis, which is the inability of existing models to generate intact objects that are consistent with the given textual description, which causes the synthetic images to deviate from reality. This objective is accomplished by the suggested method, which combines a pre-trained SBERT (Sentence-Bidirectional Encoder Representations from Transformers) model with a conditional GAN architecture to produce high-quality images that are semantically accurate to the input text description. The proposed method utilizes a GAN model, which uses supervised learning techniques to give useful feedback to train the generator and improve the consistency of the generated images. The thesis presents an in-depth review of the proposed approach utilizing a variety of evaluation metrics. Both quantitative and qualitative evaluations are used in the evaluation to show that the proposed methods effectively contribute to text-to-image synthesis models in terms of improvements in text information utilization and images with rich semantic consistency that are compatible with the input text descriptions. Additionally, the thesis explores the drawbacks and difficulties of the proposed approach and demonstrates potential future research pathways to get around this issue. The findings of this study have significance for many applications, such as generative art, computer vision, enhancing human communication, and many more. Overall, this thesis advances the field of text-to-image synthesis by presenting a novel architecture to generate images with rich semantic consistency from textual descriptions while resolving the problems associated with producing synthetic images and text utilization in GAN models." en_US
dc.language.iso en en_US
dc.publisher en_US
dc.subject Text-to-Image Generation en_US
dc.subject Generative Adversarial Networks en_US
dc.subject Computer Vision en_US
dc.title Z-GAN For Enhancing Generative Adversarial Network Based T2I Synthesis Using Transformer Model en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search


Advanced Search

Browse

My Account