资源算法TecoGAN

TecoGAN

2020-03-03 | |  108 |   0 |   0

TecoGAN

This repository contains source code and materials for the TecoGAN project, i.e. code for a TEmporally COherent GAN for video super-resolution. Authors: Mengyu Chu, You Xie, Laura Leal-Taixe, Nils Thuerey. Technical University of Munich.

This repository so far contains the code for the TecoGAN inference and training. Data generation, i.e., download, will follow soon. Pre-trained models are also available below, you can find links for downloading and instructions below. The video and pre-print of our paper can be found here:

Video: https://www.youtube.com/watch?v=pZXFXtfd-Ak Preprint: https://arxiv.org/pdf/1811.09393.pdf

图片.png

Additional Generated Outputs

Our method generates fine details that persist over the course of long generated video sequences. E.g., the mesh structures of the armor, the scale patterns of the lizard, and the dots on the back of the spider highlight the capabilities of our method. Our spatio-temporal discriminator plays a key role to guide the generator network towards producing coherent detail.

tecoGAN-lizard.gif

tecoGAN-armour.gif

tecoGAN-spider.gif

Running the TecoGAN Model

Below you can find a quick start guide for running a trained TecoGAN model. For further explanations of the parameters take a look at the runGan.py file.
Note: evaluation (test case 2) currently requires an Nvidia GPU with CUDAtkinter is also required and may be installed via the python3-tk package.

# Install tensorflow1.8+,pip3 install --ignore-installed --upgrade tensorflow-gpu # or tensorflow# Install PyTorch (only necessary for the metric evaluations) and other things...pip3 install -r requirements.txt# Download our TecoGAN model, the _Vid4_ and _TOS_ scenes shown in our paper and video.python3 runGan.py 0# Run the inference mode on the calendar scene.# You can take a look of the parameter explanations in the runGan.py, feel free to try other scenes!python3 runGan.py 1 

# Evaluate the results with 4 metrics, PSNR, LPIPS[1], and our temporal metrics tOF and tLP with pytorch.# Take a look at the paper for more details! python3 runGan.py 2

Train the TecoGAN Model

1. Prepare the Training Data

The training and validation dataset can be downloaded with the following commands into a chosen directory TrainingDataPath. Note: online video downloading requires youtube-dl.

# Install youtube-dl for online video downloadingpip install --user --upgrade youtube-dl# take a look of the parameters first:python3 dataPrepare.py --help# To be on the safe side, if you just want to see what will happen, the following line won't download anything,# and will only save information into log file.# TrainingDataPath is still important, it the directory where logs are saved: TrainingDataPath/log/logfile_mmddHHMM.txtpython3 dataPrepare.py --start_id 2000 --duration 120 --disk_path TrainingDataPath --TEST# This will create 308 subfolders under TrainingDataPath, each with 120 frames, from 28 online videos.# It takes a long time.python3 dataPrepare.py --start_id 2000 --duration 120 --REMOVE --disk_path TrainingDataPath

Once ready, please update the parameter TrainingDataPath in runGAN.py (for case 3 and case 4), and then you can start training with the downloaded data!

Note: most of the data (272 out of 308 sequences) are the same as the ones we used for the published models, but some (36 out of 308) are not online anymore. Hence the script downloads suitable replacements.

2. Train the Model

This section gives command to train a new TecoGAN model. Detail and additional parameters can be found in the runGan.py file. Note: the tensorboard gif summary requires ffmpeg.

# Install ffmpeg for the  gif summarysudo apt-get install ffmpeg # or conda install ffmpeg# Train the TecoGAN model, based on our FRVSR model# Please check and update the following parameters: # - VGGPath, it uses ./model/ by default. The VGG model is ca. 500MB# - TrainingDataPath (see above)# - in main.py you can also adjust the output directory of the  testWhileTrain() function if you like (it will write into a train/ sub directory by default)python3 runGan.py 3# Train without Dst, (i.e. a FRVSR model)python3 runGan.py 4# View log via tensorboardtensorboard --logdir='ex_TecoGANmm-dd-hh/log' --port=8008

Tensorboard GIF Summary Example

图片.png

Acknowledgements

This work was funded by the ERC Starting Grant realFlow (ERC StG-2015-637014).
Part of the code is based on LPIPS[1], Photo-Realistic SISR[2] and gif_summary[3].

Reference

[1] The Unreasonable Effectiveness of Deep Features as a Perceptual Metric (LPIPS)
[2] Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network
[3] gif_summary

TUM I15 https://ge.in.tum.de/ , TUM https://www.tum.de/


上一篇:Locally-Consistent-Deformable-Convolution

下一篇:tecoGAN_docker

用户评价
全部评价

热门资源

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • capsnet-with-caps...

    CapsNet with capsule-wise convolution Project ...

  • inferno-boilerplate

    This is a very basic boilerplate example for pe...