You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for sharing your QuIP code with us! I'm running into a bit of trouble with correctly running QuIP on the Llama-2 (7b/14b/70b) models:
The latest llama.py seems slightly broken, would you be able to verify for us whether it is working as intended for you?
So I put together my own llama.py using GPTQ's version of the code + your opt.py and got it to run without errors for Llama-2-7b-hf. However, the evaluation PPLs seem a bit high (wikitext2: 15.611, ptr-new: 353.37), suggesting that I must have done something wrong...
Any help/suggestions would be appreciated!
Sean,
The text was updated successfully, but these errors were encountered:
seannz
changed the title
Some trouble with correctly running QuIP on Llama-2 models
QuIP results on Llama-2 models not reproducible.
Oct 14, 2024
Hello Jerry,
Thank you for sharing your QuIP code with us! I'm running into a bit of trouble with correctly running QuIP on the Llama-2 (7b/14b/70b) models:
llama.py
seems slightly broken, would you be able to verify for us whether it is working as intended for you?llama.py
using GPTQ's version of the code + youropt.py
and got it to run without errors for Llama-2-7b-hf. However, the evaluation PPLs seem a bit high (wikitext2: 15.611, ptr-new: 353.37), suggesting that I must have done something wrong...Any help/suggestions would be appreciated!
Sean,
The text was updated successfully, but these errors were encountered: