Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement batch inference #30

Merged
merged 20 commits into from
Jul 12, 2023
Merged

Implement batch inference #30

merged 20 commits into from
Jul 12, 2023

Conversation

monatis
Copy link
Owner

@monatis monatis commented Jun 29, 2023

Closes #3

Currently batch inference is implemented only for image encoding. It'll be implemented for text encoding in a future PR.

@monatis monatis marked this pull request as draft June 29, 2023 20:39
@Green-Sky
Copy link
Collaborator

ggerganov/ggml#224 how much does this help?

@monatis
Copy link
Owner Author

monatis commented Jun 30, 2023

probably it's what we need --or at least a good starting point, but I'll try to have a look today.

@monatis monatis marked this pull request as ready for review July 3, 2023 21:23
@monatis monatis changed the title WIP: Implement batch inference Implement batch inference Jul 3, 2023
@monatis
Copy link
Owner Author

monatis commented Jul 3, 2023

CC @Green-Sky

@Green-Sky
Copy link
Collaborator

hm, wish we could use ggmls threading stubs

@monatis
Copy link
Owner Author

monatis commented Jul 7, 2023

I'm also thinking of it --this was a PoC, it's better to make use of GGML utils if it's possible. I'll take a look.

@monatis monatis mentioned this pull request Jul 7, 2023
5 tasks
@monatis monatis merged commit 018df28 into main Jul 12, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Experiment with batch inference
2 participants