hybrid-feature-fusion's People
hybrid-feature-fusion's Issues
Checkpoints Issues
/root/anaconda3/envs/hff/lib/python3.7/site-packages/torchvision/models/_utils.py:193: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
f"The parameter '{pretrained_param}' is deprecated since 0.13 and may be removed in the future, "
/root/anaconda3/envs/hff/lib/python3.7/site-packages/torchvision/models/_utils.py:207: UserWarning: Arguments other than a weight enum or None
for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing weights=ResNet18_Weights.IMAGENET1K_V1
. You can also use weights=ResNet18_Weights.DEFAULT
to get the most up-to-date weights.
warnings.warn(msg)
Traceback (most recent call last):
File "eval_hybrid.py", line 144, in
main()
File "eval_hybrid.py", line 124, in main
_ = load_from_checkpoint(decoder_path, decoder_lstm, optimizer=None, scheduler=None, verbose=True)
File "/media/ubuntu/新增磁碟區/hybrid-feature-fusion/src/utils.py", line 33, in load_from_checkpoint
checkpoint = torch.load(checkpoint_path,map_location='cpu')
File "/root/anaconda3/envs/hff/lib/python3.7/site-packages/torch/serialization.py", line 699, in load
with _open_file_like(f, 'rb') as opened_file:
File "/root/anaconda3/envs/hff/lib/python3.7/site-packages/torch/serialization.py", line 230, in _open_file_like
return _open_file(name_or_buffer, mode)
File "/root/anaconda3/envs/hff/lib/python3.7/site-packages/torch/serialization.py", line 211, in init
super(_open_file, self).init(open(name, mode))
IsADirectoryError: [Errno 21] Is a directory: '.'
Sorry, I am a beginner at this topic., there are two things that I want to ask.
- Can you provide more tutorials to save the checkpoints?
- Can you upload the checkpoints for the encoder & decoder files?
Thank you so much.
Pretrained Models
Hi,
First of all, thanks for sharing this code and for introducing this new benchmark . The paper is very interesting.
Is it possible to have the pretrained models that generate the results of the paper?
Thank you very much.
How to visualize the qualitative result?
Hello, excellent work! I am interested in this paper. Finally, I can get the average precision for jaad and titan datasets. But I got in trouble here. Can you provide the command or how I can visualize the qualitative result of your paper from the draw.py file? Because when I run python3 draw.py, there has nothing happens. I appreciate any help you can provide.
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.