Comments (4)
Hi,
Just to clarify something, figure 3 is only used to explain how the optimization of our loss function works.
The three ground truths are not available at the same time. You will see only one ground truth at every iteration, but during training you will see all of them.
The figure 3 explains only the sampling framework where we have EWTA. For simplicity, you can assume the simpler version of our approach where in the sampling network you generate multiple hypotheses (a set of points (x,y)) and then during fitting, you fit those hypotheses into your final mixture model.
In practice, we train the sampling network to generate 20 hypotheses and then fit them into 4 modes (as mentioned in section 6.1).
To get an idea about the EWTA loss implementation, we have already provided the code for the loss function.
Multimodal-Future-Prediction/net.py
Line 66 in d0a5d0f
We also provided the loss function used in the fitting network (nll) at:
Multimodal-Future-Prediction/net.py
Line 138 in d0a5d0f
Feel free to raise more questions if you still need help.
from multimodal-future-prediction.
Thank you for your explanation!
To confirm, can you please elaborate this sentence?
"The three ground truths are not available at the same time. You will see only one ground truth at every iteration, but during training you will see all of them."
Does it mean, we use 3 ground truth labels(3 future trajectories(x,y position data on image)) paired with one image when we are training???
from multimodal-future-prediction.
No, we use only one ground truth. Every training sample has an input (e.g, image) and a single ground truth. We generate multiple hypotheses (e.g, 8 or 20) and use the EWTA loss function (make_sampling_loss() in our repository) which takes a set of hypotheses (hyps) and a single ground truth (gt).
What we mean by figure 3 is that during training, for some iteration we see an image with its single ground truth and for another iteration (maybe after a long time) the network sees a similar input image with a different ground truth. The EWTA loss function will encourage the network to use one head in the first case while using another head in the latter case.
from multimodal-future-prediction.
Thank you for the explanation!
from multimodal-future-prediction.
Related Issues (12)
- OSError: dlopen(wemd/lib/libwemd.so, 6): image not found HOT 5
- Question related to make_sampling_loss function HOT 3
- Question about training the fitting stage HOT 3
- Training scripts for SDD dataset
- What is the `tb` package? HOT 4
- EMD on CPI dataset HOT 2
- Training scripts for CPI dataset HOT 14
- Questions about sampling network training process HOT 7
- Questions about dataset creation HOT 3
- Feasibility for the training script HOT 1
- Questions about the optimizers used for training the sampling and fitting neural networks HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from multimodal-future-prediction.