Replies: 2 comments 1 reply
-
Sure, this performance significant recent Visual RAG came up. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Langchain should add support for colpali, colqwen retrievers for multimodal rag along with vector store used to index these matrix embeddings. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Checked
Feature request
I predict there will be performance-significant multimodal (I suggest prioritizing text and image modality first) local embedding models starting with this model: https://huggingface.co/blog/paligemma. I have not yet found any open discussions or issues to support the proposed features like multimodal base document model and multi-embedding vector stores.
Motivation
Until reliable huge/infinite context LLM and cheap context usage costs are released, the RAG method will remain valuable. The increase in information (modalities in this case) likely leads to more effective and less risky use cases. LangChain here will make developers more efficient when developing the RAG system.
Proposal (If applicable)
Keeping the ideas afloat to related libraries likely incentivizes the initial effort.
Beta Was this translation helpful? Give feedback.
All reactions