资源算法TF-deformable-conv

TF-deformable-conv

2020-03-03 | |  83 |   0 |   0

TF-deformable-conv

This is a repository for a Deformable Convolution operation in Tensorflow. This repo largely borrows cuda codes from original implementation.

Check here for a inplementation of Deformable net in tensorflow.

Prerequisite

Tensorflow 1.2 (with GPU configured)

Cuda 8.0

g++ 4.9.2

Note on all version problem : Only tested on platform where corresponding version of g++ and cuda , and Tensorflow installed.

  • g++5 might encounter undefined symbol problem, It's suggested to reinstall g++4.9 to solve this problem, as pointed out by @cotrane in this issue.

  • If you want to use Tensorflow 1.4+ and CuDNN 6, @John1231983 has an modified script in this issue.

Usage

  1. Set up TF_INC and CUDA_HOME, where TF_INC can be set up as TF_INC=$(python -c 'import tensorflow as tf; print(tf.sysconfig.get_include())'). Make sure CUDA_HOME be the path where cuda is installed, such as default: /usr/local/cuda.

  2. Build the op. If you have Tensorflow source installed, you could copy all cpp files contained in ./lib and BUILD to $(Tensorflow_source_dir)/tensoflow/core/user_ops, then run bazel build --config=opt --config=cuda //tensorflow/core/user_ops:deform_conv.so in $(Tensorflow_source_dir). If not, run ./lib/nvcc_complie.shand ./lib/g++_complie.sh in sequence to build deform_conv.so. (If cuda_config.h is reported to be missed, check here)

  3. import lib.deform_conv_op as deform_conv_op in your python script (make sure PYTHON_PATH was set currectly).

Demo

A simple WGAN script trained on MNIST, to validated the backpropagation.

Since offset mostly stays between -1 and 1 there is no need to visualize it. Considering the simplicity of discriminator task, I'm not suprised about it. Might considering bring scaled MNIST in and pretrain regular conv part or change the initializer of offset conv to random normal to make deform matters.

TODO

  •  Basic test with original implementation.

  •  --Make sure gradient work.(weird bug happened, data grad used to be correct except for first time calculated, now in my test it works normal, but if you find any bug just open an issue)

  •  Simple benchmark.

  •  Some demo and visualization.

  •  Backward time costs too much.

  •  Other ops.

Benchmark

Benchmark script is borrowed from here. The forward time is fine, for 100x3x224x224 data, it runs about in 0.077s. But backward time is generaly undesired, it costs 0.558s to run a batch of same data. Note I write all backward of three inputs(data, offset, kernels) together, rather than like many tensorflow conv ops spliting input_backwards and kernel_backwards to two ops, so this might be one of the reason. In addition, because sometimes I find it hard to manipulate tensorflow::Tensor , I write a simple cuda kernel that does nothing but add one tensor to another, for accumulating gradients along batch in kernel gradient implementation, don't know whether it affects performance.


上一篇:deformableConvolution_3D

下一篇:tf-deformable-conv-layer

用户评价
全部评价

热门资源

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • capsnet-with-caps...

    CapsNet with capsule-wise convolution Project ...

  • inferno-boilerplate

    This is a very basic boilerplate example for pe...