EconPapers    
Economics at your fingertips  
 

ODD-VGAN: Optimised Dual Discriminator Video Generative Adversarial Network for Text-to-Video Generation with Heuristic Strategy

Rayeesa Mehmood (), Rumaan Bashir and Kaiser J. Giri ()
Additional contact information
Rayeesa Mehmood: Department of Computer Science, Islamic University of Science & Technology, Kashmir, J&K 192122, India
Rumaan Bashir: Department of Computer Science, Islamic University of Science & Technology, Kashmir, J&K 192122, India
Kaiser J. Giri: Department of Computer Science, Islamic University of Science & Technology, Kashmir, J&K 192122, India

Journal of Information & Knowledge Management (JIKM), 2025, vol. 24, issue 03, 1-23

Abstract: Generating videos is a novel area of computer vision research that is still far from being addressed. The reason for the same being that videos are very complex in nature where both spatial and temporal coherence need to be taken care of. Compared to the unconditional video generation, an automated video generation from the text description is an even more difficult task, in which maintaining semantic consistency and visual quality are very crucial. The video generation from the text description seems to be non-trivial owing to the intrinsic complexity that occurs in the frames and video framework. The conditional generative models are required to be implemented for this challenging task of text-to-video generation. “Generative adversarial networks (GANs)†have had a lot of success in producing images conditioned over the natural language description. But, it is yet to be employed for producing realistic videos from text that are temporally and spatially coherent and semantically consistent with the text descriptions. Thus, a new Optimised Dual Discriminator Video Generative Adversarial Network (ODD-VGAN) for text-to-video generation is suggested in this paper. The hyper-parameters of ODD-VGAN are optimised using the improved reptile search algorithm (IRSA). The efficiency of the proposed approach is demonstrated by both qualitative and quantitative experimental results.

Keywords: Text-to-video generation; Optimised Dual Discriminator Video Generative Adversarial Network; improved reptile search algorithm; optimised deep learning model; conditional GAN (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:

Downloads: (external link)
http://www.worldscientific.com/doi/abs/10.1142/S0219649223500417
Access to full text is restricted to subscribers

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:wsi:jikmxx:v:24:y:2025:i:03:n:s0219649223500417

Ordering information: This journal article can be ordered from

DOI: 10.1142/S0219649223500417

Access Statistics for this article

Journal of Information & Knowledge Management (JIKM) is currently edited by Professor Suliman Hawamdeh

More articles in Journal of Information & Knowledge Management (JIKM) from World Scientific Publishing Co. Pte. Ltd.
Bibliographic data for series maintained by Tai Tone Lim ().

 
Page updated 2025-06-28
Handle: RePEc:wsi:jikmxx:v:24:y:2025:i:03:n:s0219649223500417