Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About resampling #4

Open
ahustr opened this issue Mar 15, 2024 · 2 comments
Open

About resampling #4

ahustr opened this issue Mar 15, 2024 · 2 comments

Comments

@ahustr
Copy link

ahustr commented Mar 15, 2024

image
Hello! While studying your code, I found that when resampling is used here, it seems to be calculated directly using the variance. In my understanding, resampling seems to be using the standard deviation, so the equation logsigma.unsqueeze(1) should be divided by 2. I wonder if I missed something.

@SanghyukChun
Copy link
Collaborator

Hi, thanks for your question.

First, I would like to clarify that this function is not used for PCME++, but only for sampling-based one, PCME. Hence it does not change the results of PCME++.

Second, your question looks correct, and that means I made a mistake when I implemented PCME (https://github.com/naver-ai/pcme). In general, I think it is not a very critical issue because, during training, it only affects the Monte-Carlo sampling-based pairwise distance and VIB loss (=KL divergence). It would not predict the correct PCME loss as the original design, but as the standard deviation value is usually much smaller than the mean distances, I think the corrected version and the wrong implementation will show almost similar empirical performances.

@ahustr
Copy link
Author

ahustr commented Mar 16, 2024

Yes, it is more like a weighting parameter, thanks for your wonderful paper!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants