Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

what kind of depth image is expected to have best performance with current checkpoint? #14

Open
lyc233333 opened this issue Jan 2, 2025 · 2 comments

Comments

@lyc233333
Copy link

thanks for your great jobs in this project!

i am trying to use it to refine the depth image got from rgbd/stereo cameras, but it works not so great. could you please tell something more about the depth image used in this project? like the format, range, precision etc, and i can add more preprocessing on it.

@TychoBomer
Copy link

Searching for the same thing!

@charisoudis
Copy link

Hi, I am also exploring its application to RGBD data. It seems that the model does not generalize well to different prompt depth domains, in contrast with DepthAnything v2 (metric checkpoint). In particular, the fusion modules are expecting low res depth maps to draw sparse scale-shifts from and perform (sprarse) local refinement of the mono depths. When prompted with higher-res depth maps, whose values contain larger local variances, fusion performs worse.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants