Comments (6)
Hi, thanks for following our work! When using the tool NNI, we did not use any other special search strategy, you can see our settings for hyperparameters in "search_space.json". Anyway, everyone's training process may be different, you can try more combinations of parameters, in our experience, batchsize at [24,32,40, 48, 56, 64, 72] and other multiples of 8, and init_lr at [0.0001,0.00001,0.000001] might be better, we will adjust the range of these thresholds according to the training situation. Whenever the model converges, you can save the model, adjust the search space, read in the model weights and continue the training,there is no special time-saving ways.
from tanet.
Can I reproduce the results of the paper by using the default options?
https://github.com/woshidandan/TANet/blob/d911967eede3fc8a526d5a43bf5359aece187a1c/code/AVA/option.py#L3-L28
I apologize for not knowing much about NNI, my previous understanding of it was: "find a better set of parameters, such as batchsize, lr, epoch number, etc., through multiple experimental results. And then others can use this set of parameters for training to get the same good results."
But from your answer (and I also skimmed the NNI documentation), it seems that NNI is actually: "the hyperparameters are constantly adjusted during training (if it is specified in the search space), and the model converges from the initial state to the best (assuming it is best) state in only one experiment. That means if I want to reproduce your results I should use the same from batchsize and lr at each epoch as you do. But if I use the same NNI configuration as you do, I will theoretically also get the same hyperparameters as you do at each epoch and get the same results. And it might not be very time consuming because the whole process is only one experiment instead of multiple experiments to take the best."
Is that so? If there is anything wrong with what I said please let me know.
One more question, train_nni.py does not anywhere read search_space.json, so how does NNI know which hyperparameters need to be adjusted?
from tanet.
Q: Can I reproduce the results of the paper by using the default options?
A: No, these configurations in "options.py" will be overwritten by NNI.
Q: But if I use the same NNI configuration as you do, I will theoretically also get the same hyperparameters as you do at each epoch and get the same results
A: No, the training process will be somewhat random, for example, the initialization parameters of the network.
Q: How does NNI know which hyperparameters need to be adjusted?
A: Read this and you will understand everything: https://github.com/microsoft/nni
from tanet.
I see, thank you!
from tanet.
Hi, I think I really understand nni now, it's just normal hyperparameter search. nni uses different hyperparameters to conduct experiments, and supervises the experimental process and results to get better parameters for the next experiment.
So if that's the case, what hyperparameters did you end up using to get the results in the paper? I think that searching for hyperparameters from scratch is pointless and may lead to different results than yours.
Specifically, what are the epoch, batch size and learing rate?
from tanet.
Hi, I think I really understand nni now, it's just normal hyperparameter search. nni uses different hyperparameters to conduct experiments, and supervises the experimental process and results to get better parameters for the next experiment.
So if that's the case, what hyperparameters did you end up using to get the results in the paper? I think that searching for hyperparameters from scratch is pointless and may lead to different results than yours. Specifically, what are the epoch, batch size and learing rate?
I have the same question. I think it's best to provide the hyperparameter settings directly. ;)
from tanet.
Related Issues (20)
- 【请求】数据集上传到平台 HOT 1
- 不同batch size时inference结果不同 HOT 1
- 关于nni训练 HOT 1
- Is it possible to share the trained weights? Thanks HOT 3
- Error on single image inference, thanks HOT 7
- python的版本 HOT 1
- How to get a pretrained backbone? HOT 2
- Which hyperparameter should be set by individually HOT 1
- 关于nni训练nni.report_intermediate_result在trial中只输出一次信息的问题 HOT 1
- Some questions about reproducibility
- Inference for images? HOT 1
- 载入预训练好的权重的推理结果 HOT 1
- 请问文中提出的包含AVA, FLICKR- AES和TAD66的测试集可以提供吗? HOT 1
- 请问我该如何产生您展示的视频demo
- Results interpretation
- Apply for supplementary material
- 您好,抱歉打扰您,请问我该怎么做才能对单张图像测试,相关代码可以提供吗?万分感谢
- LCC and SRCC values got smaller and smaller as the number of training epoch increased.
- The inference effect is very poor
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from tanet.