Generating videos from text has proven to be a significant challenge for existing generative models. We tackle this problem by training a conditional generative model to extract both static and dynamic information from text. This is manifested in a hybrid framework, employing a Variational Autoencoder (VAE) and a Generative Adversarial Network (GAN). The static features, called "gist," are used to sketch text-conditioned background color and object layout structure. Dynamic features are considered by transforming input text into an image filter. To obtain a large amount of data for training the deep-learning model, we develop a method to automatically create a matched text-video corpus from publicly available online videos. Experimental res...
Learning to represent and generate videos from unlabeled data is a very challenging problem. To gene...
This paper presents a new framework, Knowledge-Transfer Generative Adversarial Network (KT-GAN), for...
Recent years have witnessed various types of generative models for natural language generation (NLG)...
We live in a world made up of different objects, people, and environments interacting with each othe...
This paper proposes two network architectures to perform video generation from captions using Variat...
This electronic version was submitted by the student author. The certified thesis is available in th...
The challenges of training generative adversarial network (GAN) to produce discrete tokens, have see...
Content creation aims to make the information (i.e., ideas, thoughts) accessible to the audience thr...
Creation of images using generative adversarial networks has been widely adapted into multi-modal re...
Cross-modal generation is playing an important role in translating information between different dat...
In the context of generative models, text-to-image generation achieved impressive results in recent ...
Content creation can be broadly defined as a way of conveying thoughts and expressing ideas through ...
Given two video frames X0 and Xn+1, we aim to generate a series of intermediate frames Y1, Y2, . . ....
Automatically generating images based on a natural language description is a challenging problem wit...
Generative models have shown impressive results in generating synthetic images. However, video synth...
Learning to represent and generate videos from unlabeled data is a very challenging problem. To gene...
This paper presents a new framework, Knowledge-Transfer Generative Adversarial Network (KT-GAN), for...
Recent years have witnessed various types of generative models for natural language generation (NLG)...
We live in a world made up of different objects, people, and environments interacting with each othe...
This paper proposes two network architectures to perform video generation from captions using Variat...
This electronic version was submitted by the student author. The certified thesis is available in th...
The challenges of training generative adversarial network (GAN) to produce discrete tokens, have see...
Content creation aims to make the information (i.e., ideas, thoughts) accessible to the audience thr...
Creation of images using generative adversarial networks has been widely adapted into multi-modal re...
Cross-modal generation is playing an important role in translating information between different dat...
In the context of generative models, text-to-image generation achieved impressive results in recent ...
Content creation can be broadly defined as a way of conveying thoughts and expressing ideas through ...
Given two video frames X0 and Xn+1, we aim to generate a series of intermediate frames Y1, Y2, . . ....
Automatically generating images based on a natural language description is a challenging problem wit...
Generative models have shown impressive results in generating synthetic images. However, video synth...
Learning to represent and generate videos from unlabeled data is a very challenging problem. To gene...
This paper presents a new framework, Knowledge-Transfer Generative Adversarial Network (KT-GAN), for...
Recent years have witnessed various types of generative models for natural language generation (NLG)...