Replies: 1 comment
-
Currently, the token probability is simply the soft-max of the logits produced by the decoder: Lines 1825 to 1844 in 1d716d6 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I try to understand the probability mechanism of a token.
in the python whisper project, the probability is an average of the whole segment and it should be less then -1.0 in order to be good enough.
in the whisper.cpp project the probability is per token that is usually one word and it range between 0.0 and 1.0 in order to be good enough. the values of whisper.cpp looks much more compliant with the audio.
how is it calculated, and why are the differ ?
Thanks
Ofer
Beta Was this translation helpful? Give feedback.
All reactions