Giter Club home page Giter Club logo

siamdw's Introduction

Deeper and Wider Siamese Networks for Real-Time Visual Tracking

We are hiring research interns for visual tracking and neural architecture search projects: [email protected]

News

  • 🏆 We are the Winner of VOT-19 RGB-D challenge [codes and models]
  • 🏆 We won the Runner-ups in VOT-19 Long-term and RGB-T challenges [codes and models]
  • ☀️☀️ We add the results on VOT-18, VOT-19, GOT10K, VISDRONE19, and LaSOT datasets.
  • ☀️☀️ The training and testing code of SiamFC+ and SiamRPN+ have been released.
  • ☀️☀️ Our paper has been accepted by CVPR2019 (Oral).
  • ☀️☀️ We provide a parameter tuning toolkit for siamese tracking framework.

Introduction

Siamese networks have drawn great attention in visual tracking because of their balanced accuracy and speed. However, the backbone network utilized in these trackers is still the classical AlexNet, which does not fully take advantage of the capability of modern deep neural networks.

Our proposals improve the performances of fully convolutional siamese trackers by,

  1. introducing CIR and CIR-D units to unveil the power of deeper and wider networks like ResNet and Inceptipon;
  2. designing backbone networks according to the analysis on internal network factors (e.g. receptive field, stride, output feature size), which affect tracking performances.

Main Results

Main results on VOT and OTB

Models OTB13 OTB15 VOT15 VOT16 VOT17
Alex-FC 0.608 0.579 0.289 0.235 0.188
Alex-RPN - 0.637 0.349 0.344 0.244
CIResNet22-FC 0.663 0.644 0.318 0.303 0.234
CIResIncep22-FC 0.662 0.642 0.310 0.295 0.236
CIResNext23-FC 0.659 0.633 0.297 0.278 0.229
CIResNet22-RPN 0.674 0.666 0.381 0.376 0.294

Main results trained with GOT-10k (SiamFC)

Models OTB13 OTB15 VOT15 VOT16 VOT17
Alex-FC - - - - 0.188
CIResNet22-FC 0.664 0.654 0.361 0.335 0.266
CIResNet22W-FC 0.689 0.674 0.368 0.352 0.269
CIResIncep22-FC 0.673 0.650 0.332 0.305 0.251
CIResNext22-FC 0.668 0.651 0.336 0.304 0.246
Raw Results 📎 OTB2013 📎 OTB2015 📎 VOT15 📎 VOT16 📎 VOT17
  • Some reproduced results listed above are slightly better than the ones in the paper.
  • Recently we found that training on GOT10K dataset can achieve better performance for SiamFC. So we provide the results being trained on GOT10K.

New added results

Benchmark VOT18 VOT19 GOT10K VISDRONE19 LaSOT
Performance 0.270 0.242 0.416 0.383 0.384
Raw Results 📎 VOT18 📎 VOT19 📎 GOT10K 📎 VISDRONE 📎 LaSOT
  • We add resutls of SiamFCRes22W on recent benchmarks.
  • Download pretrained on GOT10K model and hyper-parameters.

Environment

The code is developed with Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz GPU: NVIDIA .GTX1080

Quick Start

Test

See details in test.md

Train

See details in train.md

☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️☁️

Citation

If any part of our paper and code is helpful to your work, please generously cite with:

@InProceedings{SiamDW_2019_CVPR,
author = {Zhang, Zhipeng and Peng, Houwen},
title = {Deeper and Wider Siamese Networks for Real-Time Visual Tracking},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2019}
} 

License

Licensed under an MIT license.

siamdw's People

Contributors

hongyuanyu avatar hwpengms avatar judasdie avatar penghouwen avatar zhangliliang avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

siamdw's Issues

The scale of response map.

Thanks for your great work! It helps me a lot.

The code of convolution layer of Siamfc model is:
def forward(self, z_f, x_f): if not self.training: return 0.1 * F.conv2d(x_f, z_f) else: return 0.1 * self._conv2d_group(x_f, z_f)

Why change the scale of the results of cross-correlation? How to determine parameter 0.1?

Will the authors provide the ImageNet pretrained models?

Hi, thanks for the wonderful work.
We'd like to re-train your trackers from the ImageNet pretrained weights, as described in the paper of SiamDW. However, we cannot find the initial model links in the README. Will you provide your initial weights of ResNet22, Incep22, etc. to facilitate training? Thanks.

Error occurred when testing on benchmark OTB 2013

I used this to run tracker:
CUDA_VISIBLE_DEVICES=0 python ./siamese_tracking/test_siamfc.py --arch SiamFCRes22 --resume ./snapshot/CIResNet22.pth --dataset OTB2013

Error occurred:
load pretrained model from ./snapshot/CIResNet22.pth
remove prefix 'module.'
missing keys:{'features.features.layer2.0.bn2.num_batches_tracked', 'features.features.layer1.1.bn3.num_batches_tracked', 'features.features.layer2.2.bn2.num_batches_tracked', 'features.features.layer2.2.bn1.num_batches_tracked', 'features.features.layer1.1.bn1.num_batches_tracked', 'features.features.layer2.4.bn3.num_batches_tracked', 'features.features.bn1.num_batches_tracked', 'features.features.layer1.2.bn3.num_batches_tracked', 'features.features.layer1.0.bn3.num_batches_tracked', 'features.features.layer2.3.bn2.num_batches_tracked', 'features.features.layer1.0.bn1.num_batches_tracked', 'features.features.layer2.3.bn3.num_batches_tracked', 'features.features.layer1.0.bn2.num_batches_tracked', 'features.features.layer2.0.downsample.1.num_batches_tracked', 'features.features.layer1.0.downsample.1.num_batches_tracked', 'features.features.layer2.4.bn2.num_batches_tracked', 'features.features.layer1.2.bn2.num_batches_tracked', 'features.features.layer2.3.bn1.num_batches_tracked', 'features.features.layer2.0.bn1.num_batches_tracked', 'features.features.layer1.1.bn2.num_batches_tracked', 'features.features.layer2.2.bn3.num_batches_tracked', 'features.features.layer2.4.bn1.num_batches_tracked', 'features.features.layer1.2.bn1.num_batches_tracked', 'features.features.layer2.0.bn3.num_batches_tracked'}
unused checkpoint keys:{'connect_model.loc_adjust.weight', 'connect_model.loc_adjust.bias'}
Traceback (most recent call last):
File "./siamese_tracking/test_siamfc.py", line 248, in
main()
File "./siamese_tracking/test_siamfc.py", line 141, in main
track(tracker, net, dataset[video], args)
File "./siamese_tracking/test_siamfc.py", line 75, in track
if len(im.shape) == 2:
AttributeError: 'NoneType' object has no attribute 'shape'

when I trained SiamRPNRes22 on dataset ['GOT10K', 'LASOT'], errors occurred. But others datasets ['YTB', 'VID', 'COCO', 'DET'] are correct.

@JudasDie Thank you for your excellent research work,I have some questions to ask you.

  1. What is the problem about ['GOT10K', 'LASOT']?
    when I trained SiamRPNRes22 on datasets ['GOT10K', 'LASOT'], errors occurred. But others datasets ['YTB', 'VID', 'COCO', 'DET'] are correct.
    detailed errors are : AttributeError: 'NoneType' object has no attribute 'shape'.
    if I trained SiamFC on datasets ['GOT10K'] or ['VID'], it is all right.

  2. Why did it only get the best result VOT2015 EAO=0.332, when i trained SiamRPNRes22 on datasets ['YTB', 'VID', 'COCO', 'DET'] in GPU1080, it was lower than your paper result about 0.38.

where is script & model for CIResNet-16/19/43?

Thanks for your great works and I found this repo to be really helpful.

Yet, other than the best-performing model CIResNet-22, I am also interested in several variants that's discussed in your paper, including the light-weight CIResNet-16 and the deeper CIResNet-43. However, I did not find the pre-trained model nor the code in the repo.

Could you please also share those models?

Thanks in advance.

Submission to vot-challenge

How are you planning to submit your results to vot-challenge?

Is it the same way you are doing it now by copying handcrafted script results to vot-workspace?

because vot-toolkit works only with pytorch=0.3.1

any related suggestion would be helpful :)
Thank you

The question for training.

"2>&1 | tee logs/siamfc_train.log" what is it:?

And when i running the training code. There is a mistake ---"BrokenPipeError: [Errno 32] Broken pipe".

Questions with mapping labels and images

Hi Team Researchmm
First, thanks for your brilliant work!
I encountered problems with the data you provided(both GOT-10k and VID), although the images are cropped to the size 255 times 255, probably the labels still stay the same as that of the original GOT-10k and VID. Can you please explain how to map labels with those cropped images and if it is possible could you please upload the processing file?
Thanks a lot!

get_eao.m

出错 get_eao (line 32)
practical = get_frame_value(sequence, 'practical');
未定义与 'struct' 类型的输入参数相对应的函数 'get_frame_value'
.....

Results question

OTB2015 Best: F:\python\tracking\SiamDWmaster\siamese_tracking\result\OTB2015\SiamFCIncep22(0.1734)

The result when i run eval_otb.py .what is the 0.1734?

permission!

Thanks for your job!
Could you give me permission? I can't download training pairs by GoogleDrive.

About PAIRS: 600000

Hi, Thank you for your excellent work. I have a quetion about PAIRS. Why is it set to 600000?

Share your break-through

Some researers have shared their break-through with me. For example, Baojie told me that after replacing backbone to DenceNet in my code, the performance can still get improvement (as in the figure). So glad you people implement ideas and get improvement with our code rather than stop at the original things. This is the exact purpose for us to build a complete train-test-tuning system for siamese tracking.
git

Share your break-through with other researchers and get a comment!

The principal of selecting parameters.

I notice that different parameters are used when test proposed models with different benchmarks.
For example, the parameters of SiamFCRes22 of OTB2015 are
scale_step: 1.1897 scale_lr: 0.2226 scale_penalty: 0.9370 w_influence: 0.2897
Could you please detail the selection strategy?

Parameter tuning

I only run the parameter tuning process of pretrained SiamFCRes22 model with OTB2015 and got the error message:
Traceback (most recent call last): File "siamese_tracking/onekey_fc.py", line 101, in <module> main() File "siamese_tracking/onekey_fc.py", line 92, in main 2>&1 | tee logs/tpe_tune_fc.log'.format(trainINFO['MODEL'], 'snapshot/'+ resume, tuneINFO['DATA'], (len(info['GPUS']) + 1) // 2)) UnboundLocalError: local variable 'resume' referenced before assignment.
Could you please to tell me what can I do to avoid this error?

Some errors occured when unzipped SiamDW_DATA in Baidu driver

Thank you for your excellent job in object tracking area. I downloaded the SiamDW_DATA datas in Baidu driver,but some errors occured when I tried to unzip the datasets.The errors are shown below.
image
I tried several times,but there are same problems, I don't know how to solve these problem.Could you help me? or someone could help me?

About get_eao.m

what can i modify the pwd in the get_eao.m to make the evaluation work? i am a stranger to VOT dataset. And should i need to run the workspace_create.m in the toolkit?

REG_Loss is nan when training SiamRPN

When trained SiamRPN on the dataset VID and GOT10K, i get a problem that REG_Loss is nan. But it was ok at the beginning of the epoch.

The problems are as follows:

PROGRESS: 0.18%

Epoch: [1][180/3125] lr : 0.0010000 Batch Time: 0.283 Data Time:0.012 CLS_Loss:0.45886 REG_Loss:2.81824 Loss:3.27710
Progress: 180 / 93750 [0%], Speed: 0.283 s/iter, ETA 0:07:21 (D:H:M)

PROGRESS: 0.19%

Epoch: [1][190/3125] lr : 0.0010000 Batch Time: 0.282 Data Time:0.012 CLS_Loss:0.45153 REG_Loss:2.76308 Loss:3.21461
Progress: 190 / 93750 [0%], Speed: 0.282 s/iter, ETA 0:07:19 (D:H:M)

PROGRESS: 0.20%

Epoch: [1][200/3125] lr : 0.0010000 Batch Time: 0.281 Data Time:0.011 CLS_Loss:0.44463 REG_Loss:nan Loss:nan
Progress: 200 / 93750 [0%], Speed: 0.281 s/iter, ETA 0:07:17 (D:H:M)

PROGRESS: 0.21%

Epoch: [1][210/3125] lr : 0.0010000 Batch Time: 0.280 Data Time:0.011 CLS_Loss:0.43784 REG_Loss:nan Loss:nan
Progress: 210 / 93750 [0%], Speed: 0.280 s/iter, ETA 0:07:17 (D:H:M)

/home/xgd/sunyang/SiamDW/siamese_tracking/../lib/dataset/siamrpn.py:291: RuntimeWarning: invalid value encountered in log
delta[2] = np.log(tw / (w + eps) + eps)
/home/xgd/sunyang/SiamDW/siamese_tracking/../lib/dataset/siamrpn.py:292: RuntimeWarning: invalid value encountered in log
delta[3] = np.log(th / (h + eps) + eps)
/home/xgd/sunyang/SiamDW/siamese_tracking/../lib/dataset/siamrpn.py:291: RuntimeWarning: invalid value encountered in log
delta[2] = np.log(tw / (w + eps) + eps)
/home/xgd/sunyang/SiamDW/siamese_tracking/../lib/dataset/siamrpn.py:292: RuntimeWarning: invalid value encountered in log
delta[3] = np.log(th / (h + eps) + eps)
Warning: NaN or Inf found in input tensor.
Warning: NaN or Inf found in input tensor.

It seems that the gradient exploded and i can not figured it out ,can you help me?

issues about datasets

who can share the VID, YTB, GOT10K, COCO, DET and LASOT according to baiduyun? they are so huge and difficult to download in domestic. can share a baiduyun link ?thanks so much!

The results of the test model did not meet expectations?

Thanks for your code fisrt, it help so much.
According to the configuration process you said, I successfully ran the test code.
I then converted the txt files into a mat files and put it into the OTB framework to calculate the result. However, I found that the SiamFCRes22 model only reached 0.615 on the OTB2013. So I am very confused about what is the reason.

About Toolkit preparation in SiamRPN training

Thanks for sharing your code first. I am sorry to bother you as a deep learning beginner.
I did modified path_to/toolkit in lib/core/get_eao.m to my vot-toolkit path as

addpath('/home/ltp/vot-toolkit/'); toolkit_path; % Make sure that VOT toolkit is in the path
pwd = ['/home/ltp/vot-toolkit/', 'vot', year]  % year is a str (can not be a number)

but I still got warning before training like this:

File /home/ltp/桌面/Workspace/SiamDW-master/lib/core/get_eao.m, line 12, in get_eao
unrecognized argument /home/ltp/vot-toolkit/vot2017

This problem has bothered me for weeks,hope you can help,thanks.

20190830151133

about receptive field

I recently read your paper. It is not clear how to calculate the RF of table3 in the paper. Can you introduce it? Thank you.

The results of the training model did not meet expectations?

Thanks for your code fisrt, it help so much.
According to the configuration process you said, I successfully ran the training code of SiamFCRes22 network with CIResNet22_PRETRAIN.model as pretrained model.
In the SamFC. yaml file, the only thing I changed was the path of data. And because of the limited conditions, the number of my GPU is one.
The training code I run is as follows:
python ./siamese_tracking/train_siamfc.py --cfg experiments/train/SiamFC.yaml --gpus 0,1,2,3 --workers 32 2>&1 | tee logs/siamfc_train.log
But the best result of training 50 epochs is only about 0.62. I don't know what the problem is. I would like to ask if there is any inconsistency or neglect that caused this problem.

Thank you.

LaSOT only contains the results for 279 videos

Hi, Zhipeng,

A nice work! I noticed that for LaSOT, you only provide the results for 279 videos. Did you want us to guess which one was missing (I am kidding :))? Could you please share the complete version. Thanks.

About training result

Hi there!
I train the SiamFCRes22 following the instruction, but the performance is much lower than yours (see below)

Model OTB2013(AUC)
SiamFCRes22checkpoint_e30 0.5981
SiamFCRes22checkpoint_e50 0.5770
CIResNet22-FC 0.663

And my questions are:

  • According the training details in the paper, you trained 50 epoches, so is X-checkpoint_50.pth the final model?
  • I use the default parameters(SiamFCRes22.yaml) to test, do i need to perform Param-Tune to tune my parameters for my X-checkpoint_e50.pth?

Thanks a lot, and here is my training log:

2019-05-07 20:25:27,742 Namespace(cfg='../experiments/train/SiamFC.yaml', gpus='0', workers=32)
2019-05-07 20:25:27,743 {'CHECKPOINT_DIR': 'snapshot',
'GPUS': '0',
'OUTPUT_DIR': 'logs',
'PRINT_FREQ': 10,
'SIAMFC': {'DATASET': {'BLUR': 0,
'COLOR': 1,
'FLIP': 0,
'GOT10K': {'ANNOTATION': '/home/tjcv/dataset/SiamDW_trainset/GOT10K/train.json',
'PATH': '/home/tjcv/dataset/SiamDW_trainset/GOT10K/crop255'},
'ROTATION': 0,
'SCALE': 0.05,
'SHIFT': 4,
'VID': {'ANNOTATION': '/home/tjcv/dataset/SiamDW_trainset/VID/train.json',
'PATH': '/home/tjcv/dataset/SiamDW_trainset/VID/crop255'}},
'TEST': {'DATA': 'OTB2015',
'END_EPOCH': 50,
'MODEL': 'SiamFCIncep22',
'START_EPOCH': 30},
'TRAIN': {'BATCH': 32,
'END_EPOCH': 50,
'LR': 0.001,
'LR_END': 1e-07,
'LR_POLICY': 'log',
'MODEL': 'SiamFCRes22',
'MOMENTUM': 0.9,
'PAIRS': 600000,
'PRETRAIN': '../pretrain/CIResNet22_PRETRAIN.model',
'RESUME': False,
'SEARCH_SIZE': 255,
'START_EPOCH': 0,
'STRIDE': 8,
'TEMPLATE_SIZE': 127,
'WEIGHT_DECAY': 0.0001,
'WHICH_USE': 'VID'},
'TUNE': {'DATA': 'OTB2015',
'METHOD': 'GENE',
'MODEL': 'SiamFCIncep22'}},
'WORKERS': 32}
2019-05-07 20:25:30,937 trainable params:
2019-05-07 20:25:30,937 features.features.conv1.weight
2019-05-07 20:25:30,937 features.features.bn1.weight
2019-05-07 20:25:30,937 features.features.bn1.bias
2019-05-07 20:25:30,938 features.features.layer1.0.conv1.weight
2019-05-07 20:25:30,938 features.features.layer1.0.bn1.weight
2019-05-07 20:25:30,938 features.features.layer1.0.bn1.bias
2019-05-07 20:25:30,938 features.features.layer1.0.conv2.weight
2019-05-07 20:25:30,938 features.features.layer1.0.bn2.weight
2019-05-07 20:25:30,938 features.features.layer1.0.bn2.bias
2019-05-07 20:25:30,938 features.features.layer1.0.conv3.weight
2019-05-07 20:25:30,938 features.features.layer1.0.bn3.weight
2019-05-07 20:25:30,938 features.features.layer1.0.bn3.bias
2019-05-07 20:25:30,938 features.features.layer1.0.downsample.0.weight
2019-05-07 20:25:30,938 features.features.layer1.0.downsample.1.weight
2019-05-07 20:25:30,938 features.features.layer1.0.downsample.1.bias
2019-05-07 20:25:30,938 features.features.layer1.1.conv1.weight
2019-05-07 20:25:30,938 features.features.layer1.1.bn1.weight
2019-05-07 20:25:30,938 features.features.layer1.1.bn1.bias
2019-05-07 20:25:30,938 features.features.layer1.1.conv2.weight
2019-05-07 20:25:30,938 features.features.layer1.1.bn2.weight
2019-05-07 20:25:30,938 features.features.layer1.1.bn2.bias
2019-05-07 20:25:30,938 features.features.layer1.1.conv3.weight
2019-05-07 20:25:30,938 features.features.layer1.1.bn3.weight
2019-05-07 20:25:30,938 features.features.layer1.1.bn3.bias
2019-05-07 20:25:30,938 features.features.layer1.2.conv1.weight
2019-05-07 20:25:30,939 features.features.layer1.2.bn1.weight
2019-05-07 20:25:30,939 features.features.layer1.2.bn1.bias
2019-05-07 20:25:30,939 features.features.layer1.2.conv2.weight
2019-05-07 20:25:30,939 features.features.layer1.2.bn2.weight
2019-05-07 20:25:30,939 features.features.layer1.2.bn2.bias
2019-05-07 20:25:30,939 features.features.layer1.2.conv3.weight
2019-05-07 20:25:30,939 features.features.layer1.2.bn3.weight
2019-05-07 20:25:30,939 features.features.layer1.2.bn3.bias
2019-05-07 20:25:30,939 features.features.layer2.0.conv1.weight
2019-05-07 20:25:30,939 features.features.layer2.0.bn1.weight
2019-05-07 20:25:30,939 features.features.layer2.0.bn1.bias
2019-05-07 20:25:30,939 features.features.layer2.0.conv2.weight
2019-05-07 20:25:30,939 features.features.layer2.0.bn2.weight
2019-05-07 20:25:30,939 features.features.layer2.0.bn2.bias
2019-05-07 20:25:30,939 features.features.layer2.0.conv3.weight
2019-05-07 20:25:30,939 features.features.layer2.0.bn3.weight
2019-05-07 20:25:30,939 features.features.layer2.0.bn3.bias
2019-05-07 20:25:30,939 features.features.layer2.0.downsample.0.weight
2019-05-07 20:25:30,939 features.features.layer2.0.downsample.1.weight
2019-05-07 20:25:30,939 features.features.layer2.0.downsample.1.bias
2019-05-07 20:25:30,939 features.features.layer2.2.conv1.weight
2019-05-07 20:25:30,939 features.features.layer2.2.bn1.weight
2019-05-07 20:25:30,940 features.features.layer2.2.bn1.bias
2019-05-07 20:25:30,940 features.features.layer2.2.conv2.weight
2019-05-07 20:25:30,940 features.features.layer2.2.bn2.weight
2019-05-07 20:25:30,940 features.features.layer2.2.bn2.bias
2019-05-07 20:25:30,940 features.features.layer2.2.conv3.weight
2019-05-07 20:25:30,940 features.features.layer2.2.bn3.weight
2019-05-07 20:25:30,940 features.features.layer2.2.bn3.bias
2019-05-07 20:25:30,940 features.features.layer2.3.conv1.weight
2019-05-07 20:25:30,940 features.features.layer2.3.bn1.weight
2019-05-07 20:25:30,940 features.features.layer2.3.bn1.bias
2019-05-07 20:25:30,940 features.features.layer2.3.conv2.weight
2019-05-07 20:25:30,940 features.features.layer2.3.bn2.weight
2019-05-07 20:25:30,940 features.features.layer2.3.bn2.bias
2019-05-07 20:25:30,940 features.features.layer2.3.conv3.weight
2019-05-07 20:25:30,940 features.features.layer2.3.bn3.weight
2019-05-07 20:25:30,940 features.features.layer2.3.bn3.bias
2019-05-07 20:25:30,940 features.features.layer2.4.conv1.weight
2019-05-07 20:25:30,940 features.features.layer2.4.bn1.weight
2019-05-07 20:25:30,940 features.features.layer2.4.bn1.bias
2019-05-07 20:25:30,940 features.features.layer2.4.conv2.weight
2019-05-07 20:25:30,940 features.features.layer2.4.bn2.weight
2019-05-07 20:25:30,940 features.features.layer2.4.bn2.bias
2019-05-07 20:25:30,940 features.features.layer2.4.conv3.weight
2019-05-07 20:25:30,940 features.features.layer2.4.bn3.weight
2019-05-07 20:25:30,941 features.features.layer2.4.bn3.bias
2019-05-07 20:25:30,941 GPU NUM: 1
2019-05-07 20:25:30,945 model prepare done
2019-05-07 20:25:40,947 Epoch: [1][10/18750] lr: 0.0010000 Batch Time: 0.642s Data Time:0.334s Loss:11.23705
2019-05-07 20:25:40,947 Progress: 10 / 937500 [0%], Speed: 0.642 s/iter, ETA 6:23:12 (D:H:M)

2019-05-07 20:25:40,947
PROGRESS: 0.00%

2019-05-07 20:25:43,821 Epoch: [1][20/18750] lr: 0.0010000 Batch Time: 0.465s Data Time:0.167s Loss:8.01549
2019-05-07 20:25:43,821 Progress: 20 / 937500 [0%], Speed: 0.465 s/iter, ETA 5:01:01 (D:H:M)

2019-05-07 20:25:43,821
PROGRESS: 0.00%

2019-05-07 20:25:46,703 Epoch: [1][30/18750] lr: 0.0010000 Batch Time: 0.406s Data Time:0.112s Loss:5.91713
2019-05-07 20:25:46,703 Progress: 30 / 937500 [0%], Speed: 0.406 s/iter, ETA 4:09:41 (D:H:M)

2019-05-07 20:25:46,703
PROGRESS: 0.00%

2019-05-07 20:25:49,628 Epoch: [1][40/18750] lr: 0.0010000 Batch Time: 0.378s Data Time:0.084s Loss:4.73139
2019-05-07 20:25:49,629 Progress: 40 / 937500 [0%], Speed: 0.378 s/iter, ETA 4:02:19 (D:H:M)

2019-05-07 20:25:49,629
PROGRESS: 0.00%

2019-05-07 20:25:52,498 Epoch: [1][50/18750] lr: 0.0010000 Batch Time: 0.359s Data Time:0.067s Loss:3.99356
2019-05-07 20:25:52,498 Progress: 50 / 937500 [0%], Speed: 0.359 s/iter, ETA 3:21:35 (D:H:M)

2019-05-07 20:25:52,498
PROGRESS: 0.01%
...

last step of test

when I try to analysis testing results i meet some errors that is

image

and i don't know how to modify my get_eao.m ,can you show me how to modify it.especially what meaning of 'year‘ in pwd = ['/home/han/project/vot-toolkit/', 'vot-workspace/', 'year'] % year is a str (can not be a number)

Bad results with pretrained SiamRPN models???

Hi, Zhipeng, good job! Thanks for your kind release of this code. When I test the siamRPN with pre-trained model, however, I find the results is always lost even at the beginning. I wonder if you released the right version of the pretrained model? I see the following on my terminal:

python test_siamrpn.py
load pretrained model from ../snapshot/CIResNet22.pth
remove prefix 'module.'
missing keys:set(['connect_model.adjust.weight', 'connect_model.search_reg.bias', 'connect_model.search_cls.bias', 'connect_model.search_reg.weight', 'connect_model.template_reg.weight', 'connect_model.template_reg.bias', 'connect_model.search_cls.weight', 'connect_model.template_cls.bias', 'connect_model.template_cls.weight', 'connect_model.adjust.bias'])
unused checkpoint keys:set([u'connect_model.loc_adjust.weight', u'connect_model.loc_adjust.bias'])

Would you please kindly help to run the code succesfully? the code of SiamFC is OK (I tried). I used this setting for testing:

parser.add_argument('--arch', dest='arch', default='SiamRPNRes22', help='backbone architecture')
parser.add_argument('--resume', default='../snapshot/CIResNet22.pth', type=str, help='pretrained model')

Screenshot from 2019-07-01 19-32-03
Screenshot from 2019-07-01 19-33-03

About the groundtruth in GOT10K's json.

Thanks for sharing your code. I noticed that the groundtruth in GOT10K's json file is much bigger than the resolution of GOT10K's train pictures. The resolution is 127127 and 255255, but the gt is more like [344, 223, 776, 1002]. I have no idea how it comes, can you please explain it?

opencv-python=3.1.0.5 cannot be installed

ERROR: Could not find a version that satisfies the requirement opencv-python==3.1.0.5 (from versions: 3.4.2.17, 3.4.3.18, 3.4.4.19, 3.4.5.20, 3.4.6.27, 3.4.7.28, 4.0.0.21, 4.0.1.23, 4.0.1.24, 4.1.0.25, 4.1.1.26)
ERROR: No matching distribution found for opencv-python==3.1.0.5
Could you help me to fix it?

About training detaills

Hi,

During training for SiamFC+, do you freeze the weights of first 7*7 conv as described in the paper? If yes, why I cannot find the corresponding operations in this code? And does the same operation applies for the SiamRPN+?

Thanks for your time.

Population and group size of GA.

I run the tuning process of SiamFC with OTB-2015 on the cloud by 8 GPUs, but it takes a long time to get the results.

The population and group size of GA in tune_gune are both set to 100, the population size seems quite bigger, so I am thinking to reduce it, is that possible? Do you test the tuning process of SiamFC using GA
with small population size?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.