Giter Club home page Giter Club logo

rel_pose's People

Contributors

crockwell avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

rel_pose's Issues

Generating Epipolar lines

Hi @crockwell, this is a really cool work!

I am trying to generate the epipolar lines given the relative pose predicted by your model. However, I am not able to obtain correct epipolar lines -- my hunch is that I have misunderstood either the coordinate system (X,Y,Z axes) convention used by the model or maybe the rotation matrix is transposed?

This is what I am getting with my own script (which obviously looks wrong):
6ed97d8b-407d-40e4-9ba4-9f0fe90ab6cb

I was wondering if you can share the script you used to generate the epipolar line figures in the paper? That would be immensely helpful!

Thanks,
Yash

What is the input image size?

On skimming through the code it looks like it is 384 x 384, but in the supplemental paper it mentions it is 256 x 256.

Can you please confirm the input image size?

Several Questions about the paper

Hello! Thanks for open-sourcing this amazing work! Here I have some questions about the paper.

  1. I wonder how to get the actual U^T@U given the ground-truth rotation and translation as shown in Figure 9 in the paper? In my understanding, constructing actual U^T@U requires ground truth pixel correspondences, did you use some sort of descriptor->matching->critical-correspondences-filtering pipeline to get the actual pixel matches?
  2. I still couldn't understand why the network could learn to predict translation in actual scale, in my understanding it is unlikely to predict relative translation in actual scale by just giving 2D-pixel information as the network' input, could you kindly elaborate more on that? thanks!

Looking forward to your reply!

Regarding the depth scale

Hi,
thanks for the nice work.

I have one question regarding the depth scale in matterport. What I understood is that because matterport has rather larger translations than other datasets, the loss scale of translation error will be much bigger, which might affect negatively to the learning process.
That's why in here:

rel_pose[:3] /= Matterport.DEPTH_SCALE

you divided by 5 to adjust the scale?

If this is right, then my other question is, don't we have to change the intrinsics similarly?

Thanks!

Quick question regarding Equation 2 from paper

Thank you for the interesting work. I just had a quick question regarding Equation 2 of the paper:
image
Previously, it is explained that $Q_1$ and $K_2$ are both $P \times D$ matrices, where $P$ is the # of patches and $D$ is the feature dimension. Then, $\text{norm}(Q_1K_2^T)$ should be of dimension $P \times P$, with the first $P$ indicating image 1 patches and the latter indicating image 2 patches. However, it seems odd to me that the $V_2$ is also matrix-multiplied on the left side, effectively matching the image 2 patches' features of $V_2$ with the image 1's attention rows of $\text{norm}(Q_1K_2^T)$. Further, if the two images had different # of patches, this operation would not work either. Could I get some insight as to why this is done? I apologize if I have misinterpreted something.

Intrinsic parameters necessary?

Hi, thanks for sharing your work, really appreciate it! I have some questions about intrinsic parameters. Is the camera intrinsic parameters necessary to this algorithm? Can this algorithm work without intrinsic and extrinsic camera parameters? How is the performance (with and without intrinsic parameters)? Because I might have some dataset and want to test, but I have no idea about the intrinsic and extrinsic camera parameters.

Thank you very much!

7scenes dataset

Was there a speicfic reason to not evalaute on 7scenes and the cambridge dataset as they are often used in related works?

EMM’s ∼ U⊤U is valid only when `self.cross_features` = True ?

Enjoyed reading your paper. Nice work!

Nevertheless, I had a doubt as I went through the code and noticed the hyper parameter cross_features which I think governs equation 2's V_i, whether it is V_i.T A V_i or V_j.T A V_i. It looks to me when self.cross_features is False, the Essential Matrix module does not really capture the matrix U⊤U. Correct? If so, why would self.cross_features=False still be helpful?

Thank you.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.