-
Notifications
You must be signed in to change notification settings - Fork 146
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Text Chunking processor with Embedding doesn't work for nVIDIA model nvidia/nv-embedqa-mistral-7b-v2 #3383
Comments
Hi @ylwu-amzn , this is actually a bug related to model and connector preprocess function. Can you provide more context? |
@layavadi, suggest format the issue description to make it more readable. I helped format it Are you using OpenAI Embedding model or nvidia/nv-embedqa-mistral-7b-v2? |
I was able to reproduce the error. It seems that there is a potential bug within text chunking/embedding processor for remote model, I will look deeper into this issue. |
Thanks @nathaliellenaa, assign this issue to you |
Hi @layavadi, I did some debugging on my side and found that decreasing the token limit fixes the error. When the token limit is decreased, the text chunking and embedding work properly. Can you try changing the |
@layavadi did you get a chance to test this out? Please let us know so we can take action on the issue! Thanks! |
What is the bug?
Text Chunking processor in ingest pipeline while connecting to external embedding model like nvidia/nv-embedqa-mistral-7b-v2 is not sending the data to external model in correctly. It sends list of chunks in input instead of sending individual chunks to external mode. It sends chunks as list in input key as shown
original payload to remote model:
This is seen as single token stream in external model and it complains that token length exceeds limit.
How can one reproduce the bug?
Here is the connector definition
Pipeline definition is
Index definition is
What is the expected behavior?
Expected behaviour is send individual chunk to external model and receive the embedding and send the array of embeddings to the pos processor in ingest pipeline
What is your host/environment?
The text was updated successfully, but these errors were encountered: