Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Paralism in generating embeddings #102

Open
Daniel-SicSo-Edinburgh opened this issue Aug 17, 2024 · 1 comment
Open

Paralism in generating embeddings #102

Daniel-SicSo-Edinburgh opened this issue Aug 17, 2024 · 1 comment

Comments

@Daniel-SicSo-Edinburgh
Copy link

Hi,

I was looking over the docs, and batch inference was mentioned. I looked at the code, and it is not batch inference. It is sequential inference.

I was really hoping for batch inference because I have a lot of samples I need to embed, and doing them in parallel would save a lot of time. Maybe this can be added as a feature in the future.

For now, I would advise that a disclaimer should be added to the docs and warn users that it is sequential and thus slow.

@aristotaloss
Copy link

Perhaps one of the library's API methods would help in your case? I think I've seen ways to get more low-level control over the parsing/processing before it's being fed into the text encoder. Worth a shot maybe?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants