ODD-VGAN: Optimised Dual Discriminator Video Generative Adversarial Network for Text-to-Video Generation with Heuristic Strategy
Rayeesa Mehmood (),
Rumaan Bashir and
Kaiser J. Giri ()
Additional contact information
Rayeesa Mehmood: Department of Computer Science, Islamic University of Science & Technology, Kashmir, J&K 192122, India
Rumaan Bashir: Department of Computer Science, Islamic University of Science & Technology, Kashmir, J&K 192122, India
Kaiser J. Giri: Department of Computer Science, Islamic University of Science & Technology, Kashmir, J&K 192122, India
Journal of Information & Knowledge Management (JIKM), 2025, vol. 24, issue 03, 1-23
Abstract:
Generating videos is a novel area of computer vision research that is still far from being addressed. The reason for the same being that videos are very complex in nature where both spatial and temporal coherence need to be taken care of. Compared to the unconditional video generation, an automated video generation from the text description is an even more difficult task, in which maintaining semantic consistency and visual quality are very crucial. The video generation from the text description seems to be non-trivial owing to the intrinsic complexity that occurs in the frames and video framework. The conditional generative models are required to be implemented for this challenging task of text-to-video generation. “Generative adversarial networks (GANs)†have had a lot of success in producing images conditioned over the natural language description. But, it is yet to be employed for producing realistic videos from text that are temporally and spatially coherent and semantically consistent with the text descriptions. Thus, a new Optimised Dual Discriminator Video Generative Adversarial Network (ODD-VGAN) for text-to-video generation is suggested in this paper. The hyper-parameters of ODD-VGAN are optimised using the improved reptile search algorithm (IRSA). The efficiency of the proposed approach is demonstrated by both qualitative and quantitative experimental results.
Keywords: Text-to-video generation; Optimised Dual Discriminator Video Generative Adversarial Network; improved reptile search algorithm; optimised deep learning model; conditional GAN (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
http://www.worldscientific.com/doi/abs/10.1142/S0219649223500417
Access to full text is restricted to subscribers
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:wsi:jikmxx:v:24:y:2025:i:03:n:s0219649223500417
Ordering information: This journal article can be ordered from
DOI: 10.1142/S0219649223500417
Access Statistics for this article
Journal of Information & Knowledge Management (JIKM) is currently edited by Professor Suliman Hawamdeh
More articles in Journal of Information & Knowledge Management (JIKM) from World Scientific Publishing Co. Pte. Ltd.
Bibliographic data for series maintained by Tai Tone Lim ().