You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently, we can upload an image to get answers about it using multimodal models. But this feature doesn't support RAG workflow.
It would be good if the model first understand the image and then answers the question about it using the documents stored in the vector storage. This way, users would be able to get answers by uploading images of errors.
The text was updated successfully, but these errors were encountered:
I agree with you, it would mean moving the multimodal logic (at least for bedrock) to the Langchain interface. (Or add Rag support toe the idefics one)
I will keep this issue updated if it's being added.
Currently, we can upload an image to get answers about it using multimodal models. But this feature doesn't support RAG workflow.
It would be good if the model first understand the image and then answers the question about it using the documents stored in the vector storage. This way, users would be able to get answers by uploading images of errors.
The text was updated successfully, but these errors were encountered: