中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Deconvolutional Generative Adversarial Networks with Application to Video Generation

文献类型:会议论文

作者Yu HY(俞宏远); Huang Y(黄岩); Pi, Lihong; Wang L(王亮)
出版日期2019
会议日期2019年
会议地点西安
英文摘要

This paper proposes a novel model for video generation and especially makes the attempt to deal with the problem of video generation from text descriptions, i.e., synthesizing realistic videos conditioned on given texts. Existing video generation methods cannot be easily adapted to handle this task well, due to the frame discontinuity issue and their text-free generation schemes. To address these problems, we propose a recurrent deconvolutional generative adversarial network (RD-GAN), which includes a recurrent deconvolutional network (RDN) as the generator and a 3D convolutional neural network (3D-CNN) as the discriminator. The RDN is a deconvolutional version of conventional recurrent neural network, which can well model the long-range temporal dependency of generated video frames and make good use of conditional information. The proposed model can be jointly trained by pushing the RDN to generate realistic videos so that the 3D-CNN cannot distinguish them from real ones. We apply the proposed RD-GAN to a series of tasks including conventional video generation, conditional video generation, video prediction and video classification, and demonstrate its effectiveness by achieving well performance.
 

语种英语
源URL[http://ir.ia.ac.cn/handle/173211/48520]  
专题自动化研究所_智能感知与计算研究中心
通讯作者Wang L(王亮)
作者单位1.The Institute of Microelectronics, Tsinghua University (THU)
2.中国科学院大学
3.自动化研究所,NLPR,CRIPAC
推荐引用方式
GB/T 7714
Yu HY,Huang Y,Pi, Lihong,et al. Deconvolutional Generative Adversarial Networks with Application to Video Generation[C]. 见:. 西安. 2019年.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。