-
Notifications
You must be signed in to change notification settings - Fork 80
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
question about the result #12
Comments
The sampling strategy! you should use popular and you should reimplemented one according to original paper |
Where are the negative samples even used for Bert? I don't see it referenced anywhere in the code other than being initialized. |
You may want to have a look the paper BERT4Rec and SASRec, where two different negative sample strategies are listed. The implementation of this library is not corresponding to neither of them.
…________________________________
From: Zach Nussbaum ***@***.***>
Sent: 06 October 2021 23:14
To: jaywonchung/BERT4Rec-VAE-Pytorch ***@***.***>
Cc: ZHAO Zibo ***@***.***>; Comment ***@***.***>
Subject: Re: [jaywonchung/BERT4Rec-VAE-Pytorch] question about the result (#12)
This email was sent to you by someone outside the University.
You should only click on links or attachments if you are certain that the email is genuine and the content is safe.
Where are the negative samples even used for Bert? I don't see it referenced anywhere in the code other than being initialized.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub<#12 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AREKLLEZK4KWXXTLCEVS74TUFTC4HANCNFSM5A3XGPKQ>.
Triage notifications on the go with GitHub Mobile for iOS<https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675> or Android<https://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub>.
The University of Edinburgh is a charitable body, registered in Scotland, with registration number SC005336. Is e buidheann carthannais a th’ ann an Oilthigh Dhùn Èideann, clàraichte an Alba, àireamh clàraidh SC005336.
|
How is the sampling so much different?
is the sampling described in the paper, and from what it looks like, is the sampling strategy here. The only difference I can see here is that random popular sampling isn't used and instead the most popular samples are used. I'm surprised this would yield a 2x improvement over the original BERT4Rec paper. |
The implementation of this library is not correct (I believe), because most of the negative samples will not change with the negative sample seed, which will result in overfitting if you tune hyperparameters on validation set. For results, random sampling > popular sampling > top popular sampling. And for why this happen, I consider that the model itself may have popularity bias that learnt from statistics of training set, as a result is harder for the model to distinguish ground truth label from popular candidates set.
…________________________________
From: Zach Nussbaum ***@***.***>
Sent: 07 October 2021 20:01
To: jaywonchung/BERT4Rec-VAE-Pytorch ***@***.***>
Cc: ZHAO Zibo ***@***.***>; Comment ***@***.***>
Subject: Re: [jaywonchung/BERT4Rec-VAE-Pytorch] question about the result (#12)
This email was sent to you by someone outside the University.
You should only click on links or attachments if you are certain that the email is genuine and the content is safe.
How is the sampling so much different?
the common strategy in [12, 22, 49], pairing each ground truth item
in the test set with 100 randomly sampled negative items that the
user has not interacted with. To make the sampling reliable and
representative [19], these 100 negative items are sampled according
to their popularity. Hence, the task becomes to rank these negative
items with the ground truth item for each user```
is the sampling described in the paper, and from what it looks like, is the sampling strategy [here](https://github.com/jaywonchung/BERT4Rec-VAE-Pytorch/blob/master/dataloaders/negative_samplers/popular.py#L13).
The only difference I can see here is that random popular sampling isn't used and instead the most popular samples are used. I'm surprised this would yield a 2x improvement over the original BERT4Rec paper.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub<#12 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AREKLLHEII7WANECYEEKMJTUFXU7FANCNFSM5A3XGPKQ>.
Triage notifications on the go with GitHub Mobile for iOS<https://apps.apple.com/app/apple-store/id1477376905?ct=notification-email&mt=8&pt=524675> or Android<https://play.google.com/store/apps/details?id=com.github.android&referrer=utm_campaign%3Dnotification-email%26utm_medium%3Demail%26utm_source%3Dgithub>.
The University of Edinburgh is a charitable body, registered in Scotland, with registration number SC005336. Is e buidheann carthannais a th’ ann an Oilthigh Dhùn Èideann, clàraichte an Alba, àireamh clàraidh SC005336.
|
Recsys.pdf |
Why did I use popular sampling and the results are much less than in the paper |
Does both the training set and the test set use popular sampling? |
Why your results are so much better than the original paper? is there anything different?
The text was updated successfully, but these errors were encountered: