Giter Club home page Giter Club logo

e3d_lstm's Introduction

E3D-LSTM

It contains a Tensorflow implementation of the following paper:

Please note that this is not an officially supported Google product. This codebase was reproduced after the first author left Google in accordance with company policy

If you find this code useful in your research then please cite

@inproceedings{wang2019eidetic,
  title={Eidetic 3D LSTM: A Model for Video Prediction and Beyond},
  author={Wang, Yunbo and Jiang, Lu and Yang, Ming-Hsuan and Li, Li-Jia and Long, Mingsheng and Fei-Fei, Li.},
  booktitle={ICLR},
  year={2019}
}

We present a new model, Eidetic 3D LSTM (E3D-LSTM), that integrates 3D convolutions into RNNs. The encapsulated 3D-Conv makes local perceptrons of RNNs motion-aware and enables the memory cell to store better short-term features. We evaluate the E3D-LSTM network on (a) future video prediction (for unsupervised video representation learning) and early activity recognition to infer what is happening or what will happen after observing only limited frames of video.

Method

Setup

All code was developed and tested on Nvidia V100 the following environment.

  • Python 2.7
  • opencv3
  • scikit-image
  • numpy
  • tensorflow>=1.0
  • cuda>=8.0
  • cudnn>=5.0

Please download the data via the following external links.

  • Moving MNIST is a dataset with two moving digits bouncing in a 64 by 64 area.
  • KTH Actions is a human action dataset. This dataset contains frames from original videos. It selects the reasonable, predictable ones and resize them.

Quick Start

To train our model on the Moving NIST dataset using:

python -u run.py \
    --is_training True \
    --dataset_name mnist \
    --train_data_paths ~/data/moving-mnist-example/moving-mnist-train.npz \
    --valid_data_paths ~/data/moving-mnist-example/moving-mnist-valid.npz \
    --pretrained_model pretrain_model/moving_mnist_e3d_lstm/model.ckpt-80000 \
    --save_dir checkpoints/_mnist_e3d_lstm \
    --gen_frm_dir results/_mnist_e3d_lstm \
    --model_name e3d_lstm \
    --allow_gpu_growth True \
    --img_channel 1 \
    --img_width 64 \
    --input_length 10 \
    --total_length 20 \
    --filter_size 5 \
    --num_hidden 64,64,64,64 \
    --patch_size 4 \
    --layer_norm True \
    --sampling_stop_iter 50000 \
    --sampling_start_value 1.0 \
    --sampling_delta_per_iter 0.00002 \
    --lr 0.001 \
    --batch_size 4 \
    --max_iterations 1 \
    --display_interval 1 \
    --test_interval 1 \
    --snapshot_interval 10000

A full list of commands can be found in the script folder. The training script has a number of command-line flags that you can use to configure the model architecture, hyperparameters, and input / output settings. Below are the parameters about our model:

  • --model_name: The model name. Default value is e3d_lstm.
  • --pretrained_model: Directory to find our pretrained models. See below for the download instruction.
  • --num_hidden: Comma separated number of units of e3d lstms
  • --filter_size: Filter of a single e3d-lstm layer.
  • --layer_norm: Whether to apply tensor layer norm.

scheduled_sampling, sampling_stop_iter, sampling_start_value and sampling_changing_rate are hyperparameters used for scheduled sampling in training. The standard parameters for training and testing are:

  • --is_training: Is it training or testing.
  • --train_data_paths, --valid_data_paths: Training and validation dataset path.
  • --gen_frm_dir: Directory to store the prediction results.
  • --allow_gpu_growth: Whether allows GPU to grow.
  • --input_length 10: Input sequence length.
  • --total_length 20: Input and output sequence length in total.

To test a model, set --is_training False.

Pretrained Models

First download our pretrained models. You can test it on the dataset:

We noticed that there is a bug in the current code about "global_memory" which may be the cause for the mismatched pretrained models on the KTH dataset. As this code repo was reproduced after the first author left Google, this issue did not exist in our original experiments and the results reported in the paper are good. We are working on fixing this issue and refreshing our pre-trained KTH models. We apologize for the inconvenience and thank you for your patience.

e3d_lstm's People

Contributors

roadjiang avatar wyb15 avatar yunbo426 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

e3d_lstm's Issues

can't get performance result on kth

Hello, thank you for the paper and releasing the code. I run the test on the kth action dataset uising the downloaded the pre-trained weights, but I can't get the results of the paper which report on the kth task.
Maybe it's my parameters or I am not running it the right way?

pd11
pd12
pd13
pd14
pd15

can't get performance result on kth 10->20 prediction task

Hi ,
Thanks for your codes, I have a question about the second experiment on kth action dataset, the results I get are very different from yours.
I train the model from scrach and get:
itr: 200000
training loss: 154.842010
mse per seq: 2290.635904 64.156601 74.118351 81.989762 89.040916 95.132853 100.411171 105.039668 109.337186 113.253830 117.086717 120.570082 124.080130 127.446181 130.632763 133.742619 136.622508 139.171452 141.407400 143.135595 144.260118
psnr per frame: 23.274492 26.751829 25.718252 25.055241 24.532928 24.130718 23.816675 23.556755 23.334009 23.138975 22.954979 22.791883 22.640432 22.499092 22.369574 22.250683 22.139042 22.043234 21.966614 21.913063 21.885820
fmae per frame: 876.817078 700.009094 747.653564 776.031311 800.656433 820.194458 836.558167 850.919067 864.060059 876.001282 887.723083 898.347839 908.817383 918.579346 927.850098 936.855286 945.207214 952.337158 958.406006 963.335754 966.798523
ssim per frame: 0.731547 0.796917 0.778144 0.767673 0.759210 0.751992 0.745644 0.740227 0.735251 0.730660 0.726327 0.722522 0.718926 0.715599 0.712442 0.709682 0.707075 0.705135 0.703583 0.702254 0.701677

And i find that your paper refer to " In the end, we split the database into a training set of 108,717 sequences and a test set of 4,086 sequences."
but when i run the code, it shows
begin load datadata/kth_action
there are 127271 pictures
there are 5200 sequences
begin load datadata/kth_action
there are 74833 pictures
there are 3167 sequences

Why the dataset sequences is different with the paper. Do you know what's wrong ?

About the src/layers/rnn_cell.py

In the call() function of EideticLSTMCell class, on line 210 and line 215 there are:
output_gate = tf.tanh(o_x + o_h + o_c + o_m)
output = tf.tanh(memory) * tf.sigmoid(output_gate)
However, according to the paper, I think these 2 lines should be:
output_gate = tf.sigmoid(o_x + o_h + o_c + o_m)
output = tf.tanh(memory) * output_gate
Are there any explanations that there's and additional tanh activation on the (o_x + o_h + o_c + o_m)?

pretrained model

i couldn't find the pretrained model , After I opened the URL in Readme , it was an empty URL.

Question about speed and memory?

Thanks to your great week on this repo, I have some questions about speed and memory of this model,
1, Do you use a single V100 to train this model? How long does it take to complete on the two dataset, moving mnist and KTH actions?
2, In your default settings, batch_size = 4, why use such a small batch_size? Is it because that the model is too large? What's the memory of your V100, 32G or 16G?

Model performance on KTH 10->20 task

Hello, thank you for the paper and releasing the code. I'm having difficulty reproducing the results for the KTH Action task in section 4.2. I've downloaded the pre-trained weights for KTH Actions (200,000 ckpt) and used it to test the model.

System Info
python 2.7
opencv 4.1.0.25
tensorflow-gpu 1.9.0
CUDA 9.0
GPU:
name: TITAN X (Pascal) major: 6 minor: 1 memoryClockRate(GHz): 1.531
pciBusID: 0000:03:00.0
totalMemory: 11.91GiB freeMemory: 11.75GiB

script
#!/usr/bin/env bash cd .. python -u run.py \ --is_training False \ --dataset_name action \ --train_data_paths data/kth \ --valid_data_paths data/kth \ --pretrained_model kth_e3d_lstm_pretrain/model.ckpt-200000 \ --save_dir checkpoints/_kth_e3d_lstm \ --gen_frm_dir results/_kth_e3d_lstm \ --model_name e3d_lstm \ --allow_gpu_growth True \ --img_channel 1 \ --img_width 128 \ --input_length 10 \ --total_length 30 \ --filter_size 5 \ --num_hidden 64,64,64,64 \ --patch_size 8 \ --layer_norm True \ --reverse_input False \ --sampling_stop_iter 100000 \ --sampling_start_value 1.0 \ --sampling_delta_per_iter 0.00001 \ --lr 0.001 \ --batch_size 2 \ --max_iterations 1 \ --display_interval 1 \ --test_interval 1 \ --snapshot_interval 5000
output
(e3d_lstm_official) kstan@yixing:~/e3d_lstm/scripts$ ./e3d_lstm_kth_test.sh
Initializing models
2019-05-15 14:37:16.852811: I tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports instructions that this TensorFlow binary was not compiled to use: SSE4.1 SSE4.2 AVX AVX2 FMA
2019-05-15 14:37:19.055412: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1392] Found device 0 with properties:
name: TITAN X (Pascal) major: 6 minor: 1 memoryClockRate(GHz): 1.531
pciBusID: 0000:03:00.0
totalMemory: 11.91GiB freeMemory: 11.75GiB
2019-05-15 14:37:19.055439: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1471] Adding visible gpu devices: 0
2019-05-15 14:37:19.262277: I tensorflow/core/common_runtime/gpu/gpu_device.cc:952] Device interconnect StreamExecutor with strength 1 edge matrix:
2019-05-15 14:37:19.262310: I tensorflow/core/common_runtime/gpu/gpu_device.cc:958] 0
2019-05-15 14:37:19.262318: I tensorflow/core/common_runtime/gpu/gpu_device.cc:971] 0: N
2019-05-15 14:37:19.262531: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1084] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 11376 MB memory) -> physical GPU (device: 0, name: TITAN X (Pascal), pci bus id: 0000:03:00.0, compute capability: 6.1)
load model: kth_e3d_lstm_pretrain/model.ckpt-200000
begin load datadata/kth
there are 127271 pictures
there are 5200 sequences
begin load datadata/kth
there are 74833 pictures
there are 3167 sequences
2019-05-15 14:39:52 itr: 1
training loss: 16082.05078125
2019-05-15 14:39:52 test...
mse per seq: 1853.1817014023088
96.02373807308271
80.29797137965903
84.68072711946989
83.75463825016179
84.48666421838448
84.61139482557209
85.35639578890967
86.27750272624341
87.66025201745674
89.2119170410002
90.84818150224523
92.64167446828084
94.38503250199183
96.13222195449993
98.02904253614453
99.92525694480216
101.83609684253146
103.8342688265889
105.73710226033657
107.45162212494725
psnr per frame: 23.111416
23.2865
23.752821
23.5958
23.57663
23.51337
23.477915
23.422129
23.364187
23.28756
23.209711
23.131495
23.047438
22.969624
22.893667
22.811342
22.732689
22.653484
22.571104
22.496899
22.43397
ssim per frame: 0.6098243
0.63740635
0.62530535
0.6226238
0.61893517
0.6169444
0.6149846
0.61348057
0.61197215
0.61037815
0.60889727
0.60745543
0.6060252
0.6047545
0.60347193
0.6020237
0.6007725
0.59954363
0.59822935
0.5971006
0.59618074

visual results
gt11:gt11 gt12:gt12 gt13:gt13 gt14:gt14 gt15:gt15

pd11: pd11pd12: pd12pd13: pd13pd14: pd14pd15: pd15

...

It seems like the results are very different than what's presented in the paper -- what might I be doing wrong here?

Note: I've successfully reproduced the results and achieved the same SSIM and MSE on moving mnist task in section 4.1, so I don't think it's a system/hardware issue. So I think it could be possible that there is a mistake in the downloaded pretrained KTH action model.

Best,
Kevin

How could we download the dataset?

Dear professor
now this linkage for KTH dataset on dropbox is not avaliable. We have tried thousands times but still could not download them. How could we make it ?

have a valueerror

i have an error when i run the run.py :
ValueError: Dimensions must be equal, but are 64 and 8 for 'generator/e3d-lstm_10/mul_1' (op: 'Mul') with input shapes: [8,64,64,1], [8,64,64].

a question about downloading MMNIST dataset

I want to get the MMNIST dataset through the link in README, but I can't make it. And when I used MMNIST dataset I downloaded, the code showed the error. Can anyone please share the MMNIST dataset (BAIDU YUN or email) in this project? My mail is [email protected].

Cannot get your mse score when I train 80k epoches?

I train the movingmnist dataset with 80k epoches, but at last when I test the checkpoint which I save, I cannot get the perfect score.
I get the score as below:

2019-12-07 14:34:19 test...
mse per seq: 775.0606184674647
31.52545323113879
44.8356685294417
56.46880273685188
67.7944408103316
77.83094204308274
85.80292131045539
93.0013996116623
100.49140779575508
106.04341360944545
111.2661687892998
psnr per frame: 17.902477
21.60128
20.07348
19.040869
18.222927
17.590435
17.141094
16.781181
16.430117
16.174725
15.968664

I just set the params in the run.py as below:

FLAGS.DEFINE_string('train_data_paths', 'data/moving-mnist-example/moving-mnist-train.npz', 'train data paths.')
FLAGS.DEFINE_string('valid_data_paths', 'data/moving-mnist-example/moving-mnist-valid.npz', 'validation data paths.')
FLAGS.DEFINE_string('save_dir', 'checkpoints/_mnist_e3d_lstm', 'dir to store trained net.')
FLAGS.DEFINE_string('gen_frm_dir', 'results/_mnist_e3d_lstm', 'dir to store result.')
FLAGS.DEFINE_string('logdir', './Summary', 'dir to store summary.')

FLAGS.DEFINE_boolean('is_Training', True, 'training or testing')
FLAGS.DEFINE_string('dataset_name', 'mnist', 'The name of dataset.')
FLAGS.DEFINE_integer('input_length', 10, 'input length.')
FLAGS.DEFINE_integer('total_length', 20, 'total input and output length.')
FLAGS.DEFINE_integer('img_width', 64, 'input image width.')
FLAGS.DEFINE_integer('img_channel', 1, 'number of image channel.')
FLAGS.DEFINE_integer('patch_size', 4, 'patch size on one dimension.')
FLAGS.DEFINE_boolean('reverse_input', False,
                     'reverse the input/outputs during training.')

FLAGS.DEFINE_string('model_name', 'e3d_lstm', 'The name of the architecture.')
FLAGS.DEFINE_string('pretrained_model', '', '.ckpt file to initialize from.')
FLAGS.DEFINE_string('num_hidden', '64,64,64,64',
                    'COMMA separated number of units of e3d lstms.')
FLAGS.DEFINE_integer('filter_size', 5, 'filter of a e3d lstm layer.')
FLAGS.DEFINE_boolean('layer_norm', True, 'whether to apply tensor layer norm.')

FLAGS.DEFINE_boolean('scheduled_sampling', True, 'for scheduled sampling')
FLAGS.DEFINE_integer('sampling_stop_iter', 50000, 'for scheduled sampling.')
FLAGS.DEFINE_float('sampling_start_value', 1.0, 'for scheduled sampling.')
FLAGS.DEFINE_float('sampling_changing_rate', 0.00002, 'for scheduled sampling.')

FLAGS.DEFINE_float('lr', 0.001, 'learning rate.')
FLAGS.DEFINE_integer('batch_size', 4, 'batch size for training.')
FLAGS.DEFINE_integer('max_iterations', 80000, 'max num of steps.')
FLAGS.DEFINE_integer('display_interval', 1,
                     'number of iters showing training loss.')
FLAGS.DEFINE_integer('test_interval', 1000, 'number of iters for test.')
FLAGS.DEFINE_integer('snapshot_interval', 1000,
                     'number of iters saving models.')
FLAGS.DEFINE_integer('num_save_samples', 10, 'number of sequences to be saved.')
FLAGS.DEFINE_integer('n_gpu', 1,
                     'how many GPUs to distribute the training across.')
FLAGS.DEFINE_boolean('allow_gpu_growth', True, 'allow gpu growth')

I test your pretrain model, I got the same mse in your paper:

2019-12-08 19:01:27 itr: 1
training loss: 1674.2725
2019-12-08 19:01:27 test...
mse per seq: 414.5043303178164
22.084744220267318
27.563711177848862
32.04450662723763
36.26114293711936
39.79543918693711
43.05908385832945
46.98905318724607
51.02225838013307
55.17859578562643
60.505794957071124
psnr per frame: 20.58732
23.137892
22.231098
21.548489
20.990143
20.562159
20.216097
19.843668
19.465357
19.1416
18.736696

I show some tensorboard result in my training e3d_lstm process:
image
I get the lowest train_loss is 1970 when its step on 20.25k.

And I find that I cannot get the loss as your pretrain model (loss_train:1674.2725).
So can I ask what the params you set?

I Can't get the results of the paper which report on the kth action task

Hi yunbo,
I'm very sorry for interrupting you.

I run the test on the kth action dataset according to your script, but the results I get are very different from those in your paper.

Maybe it's my parameters or I am not running it the right way?

Below is my running command.

python -u run.py \
    --is_training False\
    --dataset_name action \
    --train_data_paths ~/codes/e3d_lstm/data/kth_action \
    --valid_data_paths ~/codes/e3d_lstm/data/kth_action \
    --pretrained_model pretrain_model/kth_e3d_lstm_pretrain/model.ckpt-200000 \
    --save_dir checkpoints/_kth_e3d_lstm \
    --gen_frm_dir results/_kth_e3d_lstm \
    --model_name e3d_lstm \
    --allow_gpu_growth True \
    --img_channel 1 \
    --img_width 128 \
    --input_length 10 \
    --total_length 30 \
    --filter_size 5 \
    --num_hidden 64,64,64,64 \
    --patch_size 8 \
    --layer_norm True \
    --reverse_input False \
    --sampling_stop_iter 100000 \
    --sampling_start_value 1.0 \
    --sampling_delta_per_iter 0.00001 \
    --lr 0.001 \
    --n_gpu 1 \
    --batch_size 2 \
    --max_iterations 1 \
    --display_interval 1 \
    --test_interval 1 \
    --snapshot_interval 5000

Below is my running log.

WARNING:tensorflow:From run.py:186: The name tf.app.run is deprecated. Please use tf.compat.v1.app.run instead.

WARNING:tensorflow:From run.py:77: The name tf.gfile.Exists is deprecated. Please use tf.io.gfile.exists instead.

W1222 21:52:35.771336 140271583331904 deprecation_wrapper.py:119] From run.py:77: The name tf.gfile.Exists is deprecated. Please use tf.io.gfile.exists instead.

WARNING:tensorflow:From run.py:78: The name tf.gfile.DeleteRecursively is deprecated. Please use tf.io.gfile.rmtree instead.

W1222 21:52:35.771526 140271583331904 deprecation_wrapper.py:119] From run.py:78: The name tf.gfile.DeleteRecursively is deprecated. Please use tf.io.gfile.rmtree instead.

WARNING:tensorflow:From run.py:79: The name tf.gfile.MakeDirs is deprecated. Please use tf.io.gfile.makedirs instead.

W1222 21:52:35.771702 140271583331904 deprecation_wrapper.py:119] From run.py:79: The name tf.gfile.MakeDirs is deprecated. Please use tf.io.gfile.makedirs instead.

Initializing models
WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:65: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead.

W1222 21:52:35.778021 140271583331904 deprecation_wrapper.py:119] From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:65: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead.

WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:87: The name tf.variable_scope is deprecated. Please use tf.compat.v1.variable_scope instead.

W1222 21:52:35.780277 140271583331904 deprecation_wrapper.py:119] From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:87: The name tf.variable_scope is deprecated. Please use tf.compat.v1.variable_scope instead.

WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:88: The name tf.get_variable_scope is deprecated. Please use tf.compat.v1.get_variable_scope instead.

W1222 21:52:35.780394 140271583331904 deprecation_wrapper.py:119] From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:88: The name tf.get_variable_scope is deprecated. Please use tf.compat.v1.get_variable_scope instead.

WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/layers/rnn_cell.py:168: conv3d (from tensorflow.python.layers.convolutional) is deprecated and will be removed in a future version.
Instructions for updating:
Use `tf.keras.layers.Conv3D` instead.
W1222 21:52:35.789141 140271583331904 deprecation.py:323] From /home/dmyan/codes/e3d_lstm/src/layers/rnn_cell.py:168: conv3d (from tensorflow.python.layers.convolutional) is deprecated and will be removed in a future version.
Instructions for updating:
Use `tf.keras.layers.Conv3D` instead.
WARNING:tensorflow:From /home/dmyan/.conda/envs/py_27/lib/python2.7/site-packages/tensorflow/python/ops/init_ops.py:1251: calling __init__ (from tensorflow.python.ops.init_ops) with dtype is deprecated and will be removed in a future version.
Instructions for updating:
Call initializer instance with the dtype argument instead of passing it to the constructor
W1222 21:52:35.791179 140271583331904 deprecation.py:506] From /home/dmyan/.conda/envs/py_27/lib/python2.7/site-packages/tensorflow/python/ops/init_ops.py:1251: calling __init__ (from tensorflow.python.ops.init_ops) with dtype is deprecated and will be removed in a future version.
Instructions for updating:
Call initializer instance with the dtype argument instead of passing it to the constructor
WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/layers/rnn_cell.py:96: The name tf.get_variable is deprecated. Please use tf.compat.v1.get_variable instead.

W1222 21:52:35.942549 140271583331904 deprecation_wrapper.py:119] From /home/dmyan/codes/e3d_lstm/src/layers/rnn_cell.py:96: The name tf.get_variable is deprecated. Please use tf.compat.v1.get_variable instead.

WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:97: The name tf.trainable_variables is deprecated. Please use tf.compat.v1.trainable_variables instead.

W1222 21:52:44.186022 140271583331904 deprecation_wrapper.py:119] From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:97: The name tf.trainable_variables is deprecated. Please use tf.compat.v1.trainable_variables instead.

WARNING:tensorflow:From /home/dmyan/.conda/envs/py_27/lib/python2.7/site-packages/tensorflow/python/training/moving_averages.py:433: initialized_value (from tensorflow.python.ops.variables) is deprecated and will be removed in a future version.
Instructions for updating:
Use Variable.read_value. Variables in 2.X are initialized automatically both in eager and graph (inside tf.defun) contexts.
W1222 21:53:13.805632 140271583331904 deprecation.py:323] From /home/dmyan/.conda/envs/py_27/lib/python2.7/site-packages/tensorflow/python/training/moving_averages.py:433: initialized_value (from tensorflow.python.ops.variables) is deprecated and will be removed in a future version.
Instructions for updating:
Use Variable.read_value. Variables in 2.X are initialized automatically both in eager and graph (inside tf.defun) contexts.
WARNING:tensorflow:From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:122: The name tf.train.Saver is deprecated. Please use tf.compat.v1.train.Saver instead.

W1222 21:53:16.815418 140271583331904 deprecation_wrapper.py:119] From /home/dmyan/codes/e3d_lstm/src/models/model_factory.py:122: The name tf.train.Saver is deprecated. Please use tf.compat.v1.train.Saver instead.

2019-12-22 21:53:17.006586: I tensorflow/core/platform/cpu_feature_guard.cc:142] Your CPU supports instructions that this TensorFlow binary was not compiled to use: SSE4.1 SSE4.2 AVX AVX2 AVX512F FMA
2019-12-22 21:53:17.031369: I tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency: 3800000000 Hz
2019-12-22 21:53:17.032462: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x563521ce8aa0 executing computations on platform Host. Devices:
2019-12-22 21:53:17.032494: I tensorflow/compiler/xla/service/service.cc:175]   StreamExecutor device (0): <undefined>, <undefined>
2019-12-22 21:53:17.033555: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcuda.so.1
2019-12-22 21:53:17.048990: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1640] Found device 0 with properties: 
name: TITAN RTX major: 7 minor: 5 memoryClockRate(GHz): 1.77
pciBusID: 0000:1d:00.0
2019-12-22 21:53:17.049160: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.0
2019-12-22 21:53:17.050196: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10.0
2019-12-22 21:53:17.051319: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcufft.so.10.0
2019-12-22 21:53:17.051549: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcurand.so.10.0
2019-12-22 21:53:17.052788: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusolver.so.10.0
2019-12-22 21:53:17.053794: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusparse.so.10.0
2019-12-22 21:53:17.056837: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
2019-12-22 21:53:17.059303: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1763] Adding visible gpu devices: 0
2019-12-22 21:53:17.059342: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.0
2019-12-22 21:53:17.158900: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1181] Device interconnect StreamExecutor with strength 1 edge matrix:
2019-12-22 21:53:17.158936: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1187]      0 
2019-12-22 21:53:17.158945: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1200] 0:   N 
2019-12-22 21:53:17.162652: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1326] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 22845 MB memory) -> physical GPU (device: 0, name: TITAN RTX, pci bus id: 0000:1d:00.0, compute capability: 7.5)
2019-12-22 21:53:17.164135: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x563544d5aca0 executing computations on platform CUDA. Devices:
2019-12-22 21:53:17.164153: I tensorflow/compiler/xla/service/service.cc:175]   StreamExecutor device (0): TITAN RTX, Compute Capability 7.5
WARNING:tensorflow:From /home/dmyan/.conda/envs/py_27/lib/python2.7/site-packages/tensorflow/python/training/saver.py:1276: checkpoint_exists (from tensorflow.python.training.checkpoint_management) is deprecated and will be removed in a future version.
Instructions for updating:
Use standard file APIs to check for files with this prefix.
W1222 21:53:20.888210 140271583331904 deprecation.py:323] From /home/dmyan/.conda/envs/py_27/lib/python2.7/site-packages/tensorflow/python/training/saver.py:1276: checkpoint_exists (from tensorflow.python.training.checkpoint_management) is deprecated and will be removed in a future version.
Instructions for updating:
Use standard file APIs to check for files with this prefix.
INFO:tensorflow:Restoring parameters from pretrain_model/kth_e3d_lstm_pretrain/model.ckpt-200000
I1222 21:53:20.888901 140271583331904 saver.py:1280] Restoring parameters from pretrain_model/kth_e3d_lstm_pretrain/model.ckpt-200000
load model: pretrain_model/kth_e3d_lstm_pretrain/model.ckpt-200000
INFO:tensorflow:Restoring parameters from pretrain_model/kth_e3d_lstm_pretrain/model.ckpt-200000
I1222 21:53:22.366884 140271583331904 saver.py:1280] Restoring parameters from pretrain_model/kth_e3d_lstm_pretrain/model.ckpt-200000
begin load data/home/dmyan/codes/e3d_lstm/data/kth_action
there are 127271 pictures
there are 5200 sequences
I1222 21:54:30.102876 140271583331904 kth_action.py:49] Initialization for read data 
begin load data/home/dmyan/codes/e3d_lstm/data/kth_action
there are 74833 pictures
there are 3167 sequences
I1222 21:55:07.155915 140271583331904 kth_action.py:49] Initialization for read data 
2019-12-22 21:56:00.419773: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
2019-12-22 21:56:01.487206: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10.0
2019-12-22 21:56:03 itr: 1
training loss: 13579.5546875
2019-12-22 21:56:03 test...
I1222 21:56:03.320033 140271583331904 kth_action.py:49] Initialization for read data 
mse per seq: 1693.4428418943269
79.28188222763437
70.02298803082465
72.65927987692105
73.55713736220244
75.19940400018167
76.14778310711897
77.5233913718035
78.97197559522965
80.70456069493248
82.5432857120632
84.30587631678627
86.22168846636427
88.00625679994488
89.80391043435903
91.70583107828416
93.60850595991143
95.47444786365772
97.46097950263635
99.30191913828847
100.94173835518235
psnr per frame: 23.059969
23.26729
23.797897
23.652412
23.607574
23.519243
23.468027
23.394089
23.319853
23.23082
23.140533
23.055035
22.963602
22.880823
22.799498
22.714308
22.632303
22.553648
22.470747
22.396952
22.334736

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.