-
Notifications
You must be signed in to change notification settings - Fork 842
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Accelerated preprocessing #1546
Labels
Comments
msaroufim
added
enhancement
New feature or request
help wanted
Extra attention is needed
labels
Apr 1, 2022
This was referenced May 4, 2022
For accelerated decoding, these are all the endpoints we can improve with faster decoding
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
This recent PR by @min-jean-cho #1545 has made me realize how much performance we're leaving on the table when it comes to acceleration of preprocessing operations. We could more easily support more torchvision backends, faster tokenizers and add more hardware specific configurations to make it for developers to pick and choose.
This will also help our #1457 efforts since there we'd only considered so far performance of torchserve and
inference()
The text was updated successfully, but these errors were encountered: