资源算法C3D-tensorflow

C3D-tensorflow

2020-01-09 | |  32 |   0 |   0

C3D-tensorflow

This is a repository trying to implement C3D-caffe on tensorflow,useing models directly converted from original C3D-caffe.
Be aware that there are about 5% video-level accuracy margin on UCF101 split1  between our implement in tensorflow and  the original C3D-caffe.

Requirements:

  1. Have installed the tensorflow >= 1.2 version

  2. You must have installed the following two python libs: a) tensorflowb) Pillow

  3. You must have downloaded the UCF101 (Action Recognition Data Set)

  4. Each single avi file is decoded with 5FPS (it's depend your decision) in a single directory.

    • you can use the ./list/convert_video_to_images.sh script to decode the ucf101 video files

    • run ./list/convert_video_to_images.sh .../UCF101 5

  5. Generate {train,test}.list files in list directory. Each line corresponds to "image directory" and a class (zero-based). For example:

    • you can use the ./list/convert_images_to_list.sh script to generate the {train,test}.list for the dataset

    • run ./list/convert_images_to_list.sh .../dataset_images 4, this will generate test.list and train.list files by a factor 4 inside the root folder

database/ucf101/train/ApplyEyeMakeup/v_ApplyEyeMakeup_g01_c01 0
database/ucf101/train/ApplyEyeMakeup/v_ApplyEyeMakeup_g01_c02 0
database/ucf101/train/ApplyEyeMakeup/v_ApplyEyeMakeup_g01_c03 0
database/ucf101/train/ApplyLipstick/v_ApplyLipstick_g01_c01 1
database/ucf101/train/ApplyLipstick/v_ApplyLipstick_g01_c02 1
database/ucf101/train/ApplyLipstick/v_ApplyLipstick_g01_c03 1
database/ucf101/train/Archery/v_Archery_g01_c01 2
database/ucf101/train/Archery/v_Archery_g01_c02 2
database/ucf101/train/Archery/v_Archery_g01_c03 2
database/ucf101/train/Archery/v_Archery_g01_c04 2
database/ucf101/train/BabyCrawling/v_BabyCrawling_g01_c01 3
database/ucf101/train/BabyCrawling/v_BabyCrawling_g01_c02 3
database/ucf101/train/BabyCrawling/v_BabyCrawling_g01_c03 3
database/ucf101/train/BabyCrawling/v_BabyCrawling_g01_c04 3
database/ucf101/train/BalanceBeam/v_BalanceBeam_g01_c01 4
database/ucf101/train/BalanceBeam/v_BalanceBeam_g01_c02 4
database/ucf101/train/BalanceBeam/v_BalanceBeam_g01_c03 4
database/ucf101/train/BalanceBeam/v_BalanceBeam_g01_c04 4
...

Usage

  1. python train_c3d_ucf101.py will train C3D model. The trained model will saved in models directory.

  2. python predict_c3d_ucf101.py will test C3D model on a validation data set.

  3. cd ./C3D-tensorflow-1.0 &&python Random_clip_valid.py will get the random-clip accuracy on UCF101 test set with provided sports1m_finetuning_ucf101.model.

  4. C3D-tensorflow-1.0/Random_clip_valid.py code is compatible with tensorflow 1.0+ , with a little bit different with the old repository

  5. IMPORTANT NOTE: when you load the sports1m_finetuning_ucf101.model,you should use the tranpose operation like: pool5 = tf.transpose(pool5, perm=[0,1,4,2,3]),or in Random_clip_valid.py looks like:["transpose", [0, 1, 4, 2, 3]], but if you load conv3d_deepnetA_sport1m_iter_1900000_TF.model or c3d_ucf101_finetune_whole_iter_20000_TF.model,you don't need tranpose operation,just comment that line code.

Experiment result:

  • Note:
    1.All report results are done specific on UCF101 split1 (train videos:9537,test videos:3783).
    2.ALL the results are video-level accuracy,unless stated otherwise.
    3.We follow the same way to extract clips from video as the C3D paper saying:'To extract C3D feature, a video is split into 16 frame long clips with a 8-frame overlap between two consecutive clips.These clips are passed to the C3D network to extract fc6 activations. These clip fc6 activations are averaged to form a 4096-dim video descriptor which is then followed by an L2-normalization'

  • C3D as feature extractor:

platformfeature extractor modelfc6+SVMfc6+SVM+L2 norm
caffeconv3d_deepnetA_sport1m_iter_1900000.caffemodel81.99%83.39%
tensorflowconv3d_deepnetA_sport1m_iter_1900000_TF.model79.38%81.44%
tensorflowc3d_ucf101_finetune_whole_iter_20000_TF.model79.67%81.33%
tensorflowsports1m_finetuning_ucf101.model82.73%85.35%
  • finetune C3D network on UCF101 split1 with  pre-trained model:

platformpre-trained modeltrain-strategyvideo-accuracyclip-accuracyrandom-clip
caffec3d_ucf101_finetune_whole_iter_20000.caffemodeldirectly test-79.87%-
tensorflowc3d_ucf101_finetune_whole_iter_20000_TF.modeldirectly test78.35%72.77%57.15%
tensorflow-Aconv3d_deepnetA_sport1m_iter_1900000_TF.caffemodelwhole finetuning76.0%71%69.8%
tensorflow-Bsports1m_finetuning_ucf101.modelfreeze conv,only finetune fc layers79.93%74.65%76.6%
  • Note:
    1.the tensorflow-A model corresponding to the original C3D model pre-trained on UCF101 provided by @ hx173149 .
    2.the tensorflow-B model is just freeze the conv layers in tensorflow-A and finetuning  four more epochs on fc layers with learning rate=1e-3.
    3.the random-clip column means random choose one clip from each video in UCF101 test split 1 ,so the result are not so robust.But according to the Law of Large Numbers,we may assume this items is positive correlated to your video-level accuracy.
    4.it's obvious that it if you do more finetuning work based on c3d_ucf101_finetune_whole_iter_20000_TF.model,and you may achieve better performance,i didn't do it because of time limit.
    5.with no doubt that you can get better result by appropriately finetuning the network

Trained models:

ModelDescriptionCloudsDownload
C3D sports1M  TFC3D sports1M converted from caffe C3DDropboxC3D sports1M
C3D UCF101 TFC3D UCF101 trained model converted from caffe C3DDropboxC3D UCF101
C3D UCF101  TF trainfinetuning on UCF101 split1 use C3D sports1M model by  @ hx173149DropboxC3D UCF101 split1
split1 meanfile  TFUCF101 split1 meanfile converted from caffe C3DDropboxUCF101 split1 meanfile
everything aboveall four files abovebaiduyunbaiduyun

Other verions

References:


上一篇:matchnet_keras

下一篇:c3d-keras

用户评价
全部评价

热门资源

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • capsnet-with-caps...

    CapsNet with capsule-wise convolution Project ...

  • inferno-boilerplate

    This is a very basic boilerplate example for pe...