Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add batch_size param for text_embedding processor #1298

Open
wants to merge 4 commits into
base: main
Choose a base branch
from

Conversation

YeonghyeonKO
Copy link

@YeonghyeonKO YeonghyeonKO commented Nov 17, 2024

Description

Unlike the document from text_embedding processor in ingest-pipelines describes batch_size parameter, interface in opensearch-java client doesn't include it. From OpenSearch 2.16, it's possible to add batch inference support in ingest processors which inherits from AbstractBatchingProcessor(opensearch-project/neural-search#820). From now on, OpenSearch java client doesn't support batch_size parameter(optional) when defining text_embedding processor.

Since @miguel-vila's contribution by adding TextEmbeddingProcessor has been merged, there was another big change in opensearch-project/neural-search(opensearch-project/neural-search#820). In line with this change, I attempted to modify the code of the text_embedding processor of the opensearch-java client.

Issues Resolved

This PR is related to #1297


By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
For more information on following Developer Certificate of Origin and signing off your commits, please check here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant