-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
对话内容超限 Query or prefix is too long #1705
Labels
🐞 bug
Something isn't working
Comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Dify version
0.3.32
Cloud or Self Hosted
Self Hosted (Docker)
Steps to reproduce
Dify框架报Query or prefix prompt is too long, you can reduce the prefix prompt, or shrink the max token, or switch to a llm with a larger token limit size。
后端模型基于Xinference进行本地部署的ChatGLM2,max_tokens 为8192(8K),直接在Xinference输入一段内容2000多字的中文文本,可以正常回答。
基于Dify构建的对话应用,max_tokens 设置为450,输入同一段内容就报上述错误。
✔️ Expected Behavior
正常回复
❌ Actual Behavior
Query or prefix prompt is too long, you can reduce the prefix prompt, or shrink the max token, or switch to a llm with a larger token limit size。
The text was updated successfully, but these errors were encountered: