dc.contributor.author |
Turarova, Marzhan |
|
dc.contributor.author |
Bekbayeva, Roza |
|
dc.contributor.author |
Abdykerimova, Lazzat |
|
dc.contributor.author |
Aitimov, Murat |
|
dc.contributor.author |
Bayegizova, Aigulim |
|
dc.contributor.author |
Smailova, Ulmeken |
|
dc.contributor.author |
Kassenova, Leila |
|
dc.contributor.author |
Glazyrina, Natalya |
|
dc.date.accessioned |
2024-12-11T04:23:24Z |
|
dc.date.available |
2024-12-11T04:23:24Z |
|
dc.date.issued |
2024 |
|
dc.identifier.issn |
2088-8708 |
|
dc.identifier.other |
DOI: 10.11591/ijece.v14i2.pp2014-2023 |
|
dc.identifier.uri |
http://rep.enu.kz/handle/enu/20041 |
|
dc.description.abstract |
Modern developments in the fields of natural language processing (NLP)
and computer vision (CV) emphasize the increasing importance of
generating images from text descriptions. The presented article analyzes and
compares two key methods in this area: generative adversarial network with
conditional latent semantic analysis (GAN-CLS) and ultra-long transformer
network (XLNet). The main components of GAN-CLS, including the
generator, discriminator, and text encoder, are discussed in the context of
their functional tasks—generating images from text inputs, assessing the
realism of generated images, and converting text descriptions into latent
spaces, respectively. A detailed comparative analysis of the performance of
GAN-CLS and XLNet, the latter of which is widely used in the organic
light-emitting diode (OEL) field, is carried out. The purpose of the study is
to determine the effectiveness of each method in different scenarios and then
provide valuable recommendations for selecting the best method for
generating images from text descriptions, taking into account specific tasks
and resources. Ultimately, our paper aims to be a valuable research resource
by providing scientific guidance for NLP and CV experts. |
ru |
dc.language.iso |
en |
ru |
dc.publisher |
International Journal of Electrical and Computer Engineering |
ru |
dc.relation.ispartofseries |
Vol. 14, No. 2; |
|
dc.subject |
Discriminator |
ru |
dc.subject |
Extra-long transformer network |
ru |
dc.subject |
Generative adversarial network |
ru |
dc.subject |
with conditional latent semantic |
ru |
dc.subject |
Generator |
ru |
dc.subject |
Machine learning |
ru |
dc.subject |
Natural language processing |
ru |
dc.title |
Generating images using generative adversarial networks based on text descriptions |
ru |
dc.type |
Article |
ru |