Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

QuIP results on Llama-2 models not reproducible. #15

Open
seannz opened this issue Oct 8, 2024 · 0 comments
Open

QuIP results on Llama-2 models not reproducible. #15

seannz opened this issue Oct 8, 2024 · 0 comments

Comments

@seannz
Copy link

seannz commented Oct 8, 2024

Hello Jerry,

Thank you for sharing your QuIP code with us! I'm running into a bit of trouble with correctly running QuIP on the Llama-2 (7b/14b/70b) models:

  1. The latest llama.py seems slightly broken, would you be able to verify for us whether it is working as intended for you?
  2. So I put together my own llama.py using GPTQ's version of the code + your opt.py and got it to run without errors for Llama-2-7b-hf. However, the evaluation PPLs seem a bit high (wikitext2: 15.611, ptr-new: 353.37), suggesting that I must have done something wrong...

Any help/suggestions would be appreciated!

Sean,

@seannz seannz changed the title Some trouble with correctly running QuIP on Llama-2 models QuIP results on Llama-2 models not reproducible. Oct 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant