UCSY's Research Repository

Building Annotated Image Dataset for Myanmar Text to Image Synthesis

Show simple item record

dc.contributor.author Htwe, Nan Kham
dc.contributor.author Pa, Win Pa
dc.date.accessioned 2022-06-20T08:10:46Z
dc.date.available 2022-06-20T08:10:46Z
dc.date.issued 2021-02-25
dc.identifier.uri https://onlineresource.ucsy.edu.mm/handle/123456789/2610
dc.description.abstract Text to image synthesis is the translation of images from the input language text. The learning process can become easier when the spoken words can visualize with the images. It is one of the popular research field in combination of NLP and computer vision. Generative Adversarial Networks (GAN) have growth in the generation of images from text descriptions. We build the baseline system of Myanmar text to image synthesis and a type of annotated images dataset because there is not efficient annotated image dataset to be used in this implementation. It was created by using partial part of Oxford-102 flowers dataset. Word2Vec algorithms is used to convert word to vectors for the input sentence to GAN. GAN is applied for generation of images from Myanmar language text. This is the first text to image generation using GAN in Myanmar. The two-evaluation metrics are used to measure the quality of images. The quality of the generated images is evaluated using Inception score. The Fréchet Inception Distance (FID) is used to measure the distance between the real images (images from original dataset) and the generated images from the model. en_US
dc.publisher ICCA en_US
dc.subject GAN, Word2vec, Inception Score, FID en_US
dc.title Building Annotated Image Dataset for Myanmar Text to Image Synthesis en_US
dc.type Presentation en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Repository



Browse

My Account

Statistics