Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implementation details for RT Stereo #10

Open
nakul3112 opened this issue Sep 22, 2021 · 5 comments
Open

Implementation details for RT Stereo #10

nakul3112 opened this issue Sep 22, 2021 · 5 comments

Comments

@nakul3112
Copy link

Hi @JiaRenChang,

 I was working on AnyNet (https://github.com/mileyan/AnyNet)  project, but haven't got appropriate results because I was not able to compile SPN module  successfully (Error while running make.sh: Not able to find setup.py file). Meanwhile, I read your paper on Real-Time Stereo and was  curious to implement the same, since it has been trained on datasets like Sceneflow & KITTI just like AnyNet.

 I would appreciate your time and efforts if you could clarify few queries I have:
  1. As I read in your paper, the outcome of this model is 10-35fps on Nvidia GPU, correct?
  2. Does this model needs SPN?
  3. Is it recommneded to finetune the model trained on Sceneflow previously , with one of the KITTI datasets (Kitti2015 or Kitti2012)?
  4. Also, does Test_img.py file output the colored disparity map as shown in your paper.

Regards,
Nakul

@JiaRenChang
Copy link
Owner

Hi, @nakul3112 :
1.the proposed model is 10-35 fps on Nvidia TX2 which is a emebdding system (https://www.seeedstudio.com/NVIDIA-Jetson-TX2-4GB-Module-p-4414.html).

  1. This model does NOT need SPN.
  2. Yes.
  3. The output disparity in a single channel image (gray image).

@nakul3112
Copy link
Author

Hi @JiaRenChang,

Thanks for your quick response. Also:

  1. What strategy would you suggest to get the colored output disparity? Do we change the datatype of image? or something else?

  2. Also, the conclusion and benchmark results from the paper show that the output is even sharper than AnyNet, even though both the model give speed of 12-35 FPS, am I correct?

Regards,
Nakul

@JiaRenChang
Copy link
Owner

Hi, @nakul3112:
1. The KITTI website offers the codes for colorizing disparity. You can check it.
2. Yes, the proposed method can give superior results than AnyNet in my experiments.
Jia-Ren

@nakul3112
Copy link
Author

Hi @JiaRenChang ,

Thanks.

  1. what is the difference between Test_img.py and submission.py?
  2. So while training, the default model is RTStereoNet,but I see in finetune.py and test_img.py, the default model is StackHourGlass , could u explain if this is by purpose? I started training the model with RTStereoNet as default, and wanted to clarify those doubts before I start finetuning.

Thanks again for helping me out,
Regards,
Nakul

@nakul3112
Copy link
Author

Hi @JiaRenChang ,

Would appreciate your help in following queries.
  1. What is the difference between Test_img.py and submission.py?
    2) while training, the default model is RTStereoNet, but I see in finetune.py and test_img.py, the default model is StackHourGlass , could you explain if this is by any purpose? I started training the model with RTStereoNet as default, and wanted to clarify those doubts before I start finetuning with KITTI.

Thanks again for your time
Regards,
Nakul

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants