-
-
Notifications
You must be signed in to change notification settings - Fork 6.9k
[Frontend] Support embeddings in the run_batch API #7132
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, please make sure to run full CI as it is required to merge (or just use auto-merge). To run full CI, you can do one of these:
🚀 |
61b92fe
to
54acbc9
Compare
Looks like the way CI runs formatting is not totally consistent with simply running format.sh
bc923d2
to
10f4fcb
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great, do you mind adding a simple example to examples/offline_inference_openai.md
?
Thanks @wuisawesome |
Yep! |
Co-authored-by: Simon Mo <simon.mo@hey.com> Signed-off-by: Alvant <alvasian@yandex.ru>
Co-authored-by: Simon Mo <simon.mo@hey.com> Signed-off-by: LeiWang1999 <leiwang1999@outlook.com>
Currently run_batch which is the offline batching API only supports chat completion. This PR adds embedding supports to run_batch. The PR also adds support for empty lines (or lines with only whitespace) in the input file of run_batch.