Skip to content

Energy Estimation Accuracy with Input Tokens #86

Answered by adrienbanse
RameezI asked this question in Q&A
Discussion options

You must be logged in to vote

Hey @RameezI, sorry for the late answer, and many thanks for your interest in this project!
Following how LLMs work, we make the assumption that the typical request involves much more decoding than encoding, and therefore that the energy consumption of the encoding is negligible.
It is however true that we don’t handle the edge cases such as giving a huge text to a model, and ask for a very precise answer that would involve only a few tokens (for example, give a book to a model and ask for the age of a particular character).
Could you send us the references that state that the input context length significantly affects the compute load? If it turns out that it has more impact than what we…

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by samuelrince
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants