Giter Club home page Giter Club logo

Comments (7)

pjckoch avatar pjckoch commented on September 3, 2024 1

Hi @brade31919 ,

thanks a lot for getting back to me so quickly.

I understand that space for a paper publication is limited and not every detail can be included.
Thanks for all the clarifications!

Regarding 3, the problem is simply that I have the regular Nuscenes dataset already downloaded and currently don't have enough space left to download yours. Sorry for the misunderstanding, I didn't want to imply that there was something wrong with your command.

Best,
Patrick

from radar_depth.

brade31919 avatar brade31919 commented on September 3, 2024 1

Hi @isht7
The projected depth values in the original depth maps are float32 and the unit is meter (m). However, we don't want to save float32 because it took too much storage. A common technique is that we can convert it to uint16 by int(depth * 256). This keeps certain degree of accuracy but takes less storage. That's why we need to divide the value by 256. after reading the depth map from h5 files. I remember it's also used in kitti dataset?

from radar_depth.

brade31919 avatar brade31919 commented on September 3, 2024

Hi @pjckoch,

  1. Regarding the learning rate issue, I think the number on the paper is wrong (sorry for that). I didn't modify the default argument for learning rate since the submission, so I think it's 0.01. Regarding the batch size, as I mentioned in the README.md, I used batch size = 8 to train the current released model because the original checkpoint was deleted (trained by batch size 16). I don't want to list all the detailed reasons because I think the users might not even be interested in them. The reason why I used batch size 8 is that I can't use the cluster I used during the paper submission anymore (so no V100). On the machine I can access currently, batch size = 8 is the biggest one I can use, and it reached the similar performance.

  2. Yeah I used weight decay during the training and I didn't mention it in my paper. I don't remember whether the reason is the page limit or that I simply forgot it, but I think it's hard for one to include all the details in the paper. I didn't spend lots of time tuning the hyper-parameters, and I believe you can find something through tuning them, but that was not the main objective of my project at that time.

  3. I tried the command I provided just now and it worked. Can you specify the situation you encountered? It's hard to explain. I didn't merely change the data structure. The depth maps are not provided in the nuScenes dataset (both LiDAR and Radar), so I did the projection, discarded some unrelated info, and saved them. And the processed dataset did make the code release easier because I don't have the storage to verify all the code on the raw nuScenes dataset.

  4. That's just some old code fragments that I failed to remove thoroughly during the release. I tried multi-task model before. In that setting, we'll have (1) RGB images, (2) projected Radar depth maps, and (3) Radar points in vector format. We did depth estimation and point cloud classification simultaneously to see whether we can remove the noisy measurements and improve the depth estimation.

Sincerely,
Juan-Ting Lin

from radar_depth.

pjckoch avatar pjckoch commented on September 3, 2024

One more question:
are your pretrained models trained with sensor samples from all directions (i.e. front, front_right, back_right, back, back_left and front_left)? I've trained your model myself and results look a tad blurrier, but perhaps that's because I only loaded front and back view.

from radar_depth.

brade31919 avatar brade31919 commented on September 3, 2024

No, I only used the front and back view. There are not many Radar points on other directions. What do you mean by a tad blurrier?

  1. Can you show the metric results? Like RMSE, MAE, Delta1, etc.
  2. Did you use the processed data?

from radar_depth.

isht7 avatar isht7 commented on September 3, 2024

@brade31919 thanks for replying to the queries above. I had a question - in the .h5 files in the folder ver2_lidar1_radar3_radar_only, what is the unit of lidar_depth maps? The values in these maps are very large and I see that you divided them by 256. during data processing here.

from radar_depth.

isht7 avatar isht7 commented on September 3, 2024

Thank you very much @brade31919 for the prompt reply. I noted that you set a numpy seed to split into train and val.
As noted here, the behavior of np.random.choice may change over different python / numpy versions.
Could you share the train / val scenes used by you?

This script which I borrowed from your code should find the splits. If you could share the variables train_scenes and val_scenes, that would be great! Instead of this we could also check that the last line prints the same value on both your and my computer. On my computer, the output of

print (np.sum(train_scenes), np.sum(val_scenes))

is

322700 38125

Could you please check and tell if you also get the same output for this print statement? If the outputs are not same, it would be great if you could share the variables train_scenes and val_scenes in the code snippet.

from radar_depth.

Related Issues (18)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.