You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
A clear and concise description of what the bug is.
The issue is "API call didn't return a message", error is repeating very frequently with letta v5.5. I am using openai's gpt 4o. The problem is that the inner monologue is displayed everytime on the ui but there is no actual user response given. I have been able to debug the issue. It is because in the normal response, when everything is correct, the message.choices contains tool calls for send message. But whenever this issue arises the message.choices doesn't contain any function call. It should have been fixed according to a previous closed bug. But it is still there and is very frustating. I am attaching the logs as well.
correct normal response:
original response choices in unpack_all_inner_thoughts_from_kwargs: {
"id": "message-fd4674bd-0943-4372-a1bd-55f84da2175d",
"choices": [
{
"finish_reason": "tool_calls",
"index": 0,
"message": {
"content": null,
"tool_calls": [
{
"id": "6b6fe6f1-3edb-4a3a-b775-e0edd",
"type": "function",
"function": {
"arguments": "{\"inner_thoughts\":\"Keeping the interaction engaging. Ensuring consistency in responding to Harshit.\",\"message\":\"Hey there, Harshit! Let me know if there's anything specific you need or if you'd just like to chat.\"}",
"name": "send_message"
}
}
],
"role": "assistant",
"function_call": null
},
"logprobs": null,
"seed": null
}
],
"created": "2024-12-15T19:25:37.356338Z",
"model": "gpt-4o-2024-08-06",
"system_fingerprint": "fp_a79d8dac1f",
"object": "chat.completion",
"usage": {
"completion_tokens": 48,
"prompt_tokens": 15582,
"total_tokens": 15630
}
}
_get_ai_reply response: {
"id": "message-fd4674bd-0943-4372-a1bd-55f84da2175d",
"choices": [
{
"finish_reason": "tool_calls",
"index": 0,
"message": {
"content": "Keeping the interaction engaging. Ensuring consistency in responding to Harshit.",
"tool_calls": [
{
"id": "6b6fe6f1-3edb-4a3a-b775-e0edd",
"type": "function",
"function": {
"arguments": "{\n \"message\": \"Hey there, Harshit! Let me know if there's anything specific you need or if you'd just like to chat.\"\n}",
"name": "send_message"
}
}
],
"role": "assistant",
"function_call": null
},
"logprobs": null,
"seed": null
}
],
"created": "2024-12-15T19:25:37.356338Z",
"model": "gpt-4o-2024-08-06",
"system_fingerprint": "fp_a79d8dac1f",
"object": "chat.completion",
"usage": {
"completion_tokens": 48,
"prompt_tokens": 15582,
"total_tokens": 15630
}
}
If you're able to reproduce with the latest version that would be great and help us debug the ticket faster (if the bug is still happening), whereas we're not able to actively debug issues for older version (eg v0.5.5)
But whenever this issue arises the message.choices doesn't contain any function call
For example I don't think this should be happening when we use structured outputs, which should be on by default for gpt-4o-2024-08-06 in the latest version (but I'm not sure about v0.5.5).
Describe the bug
A clear and concise description of what the bug is.
The issue is "API call didn't return a message", error is repeating very frequently with letta v5.5. I am using openai's gpt 4o. The problem is that the inner monologue is displayed everytime on the ui but there is no actual user response given. I have been able to debug the issue. It is because in the normal response, when everything is correct, the message.choices contains tool calls for send message. But whenever this issue arises the message.choices doesn't contain any function call. It should have been fixed according to a previous closed bug. But it is still there and is very frustating. I am attaching the logs as well.
Please describe your setup
pip install letta
?pip install letta-nightly
?git clone
?letta
? (cmd.exe
/Powershell/Anaconda Shell/Terminal)Screenshots
If applicable, add screenshots to help explain your problem.
correct response:
error response:
Additional context
Add any other context about the problem here.
I have also modified the parameters of _get_ai_reply, a little:
Letta Config
Please attach your
~/.letta/config
file or copy past it below.If you're not using OpenAI, please provide additional information on your local LLM setup:
Local LLM details
If you are trying to run Letta with local LLMs, please provide the following information:
dolphin-2.1-mistral-7b.Q6_K.gguf
)The text was updated successfully, but these errors were encountered: