Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Confusion about the inverse-depth-smooth-loss #23

Open
SJoJoK opened this issue Dec 21, 2022 · 2 comments
Open

Confusion about the inverse-depth-smooth-loss #23

SJoJoK opened this issue Dec 21, 2022 · 2 comments

Comments

@SJoJoK
Copy link

SJoJoK commented Dec 21, 2022

Hello! I'm currently doing some research on NeRF, and I found your work SinNeRF.
It's an awesome work, and I download the code (in this repository).
While reading the code, I found that you use kornia.losses.inverse_depth_smoothness_loss to calculate the "self-supervised inverse depth smoothness loss" which you describe in the paper (equation 4) .
The conflict is that, while kornia.losses.inverse_depth_smoothness_loss using the first-order gradient of the RGB Image according to the official document and its source code, the equation in paper use the second-order gradient of RGB Image.
Thanks for any advice and help.

@Ir1d
Copy link
Member

Ir1d commented Dec 22, 2022

Hi @SJoJoK, Good catch! Equation 4 is indeed confusing. I will update the draft to align with the code.
Previously we used some more complex depth smoothness loss, and later found it easier to use the kornia implementation.
Also note that for the latest implementation, we observe that sometimes disabling the depth smoothness loss can lead to a better result.
Thank you again!

@SJoJoK
Copy link
Author

SJoJoK commented Dec 22, 2022

Got it, thanks for the information : )

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants