Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature Request: RAG support for multimodal models #612

Open
azaylamba opened this issue Dec 2, 2024 · 1 comment
Open

Feature Request: RAG support for multimodal models #612

azaylamba opened this issue Dec 2, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@azaylamba
Copy link
Contributor

Currently, we can upload an image to get answers about it using multimodal models. But this feature doesn't support RAG workflow.
It would be good if the model first understand the image and then answers the question about it using the documents stored in the vector storage. This way, users would be able to get answers by uploading images of errors.

@azaylamba azaylamba changed the title Feature Request: RAG support multimodal models Feature Request: RAG support for multimodal models Dec 2, 2024
@charles-marion charles-marion added bug Something isn't working enhancement New feature or request and removed bug Something isn't working labels Dec 4, 2024
@charles-marion
Copy link
Collaborator

I agree with you, it would mean moving the multimodal logic (at least for bedrock) to the Langchain interface. (Or add Rag support toe the idefics one)

I will keep this issue updated if it's being added.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
Status: No status
Development

No branches or pull requests

2 participants