-
-
Notifications
You must be signed in to change notification settings - Fork 8.5k
[Bugfix] Fix the issue where llm.generate cannot be called repeatedly after setting GuidedDecodingParams #16767
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix] Fix the issue where llm.generate cannot be called repeatedly after setting GuidedDecodingParams #16767
Conversation
Signed-off-by: chaunceyjiang <[email protected]>
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
/cc @DarkLight1337 @russellb PTAL. |
Thanks for the PR! I'm taking a closer look now. |
I meet the same problem, thanks a lot! |
Closes vllm-project#16738 Signed-off-by: Russell Bryant <[email protected]>
@chaunceyjiang can you take a look at the commit I added? It's an alternate solution that I prefer, but want to make sure you're OK with it. |
I think your solution is even better. @russellb |
Thanks for taking a look! Since I made changes, I asked for someone else to take a look for approval. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, however I would like to see a test case specifically for this
good point - I can do that pretty quick I think |
Signed-off-by: Russell Bryant <[email protected]>
added test coverage by slightly extending an existing test. |
/cc @russellb @DarkLight1337 I think this PR can be merged. |
… after setting GuidedDecodingParams (vllm-project#16767) Signed-off-by: chaunceyjiang <[email protected]> Signed-off-by: Russell Bryant <[email protected]> Co-authored-by: Russell Bryant <[email protected]> Signed-off-by: Frieda (Jingying) Huang <[email protected]>
When is this getting released? |
The next release will be around the end of this month. If you can't wait, you can install the nightly vLLM package. https://docs.vllm.ai/en/latest/getting_started/installation/gpu.html#pre-built-wheels |
Thanks! I was having the same problem and this solves it, I tested it with the nightly package. |
… after setting GuidedDecodingParams (vllm-project#16767) Signed-off-by: chaunceyjiang <[email protected]> Signed-off-by: Russell Bryant <[email protected]> Co-authored-by: Russell Bryant <[email protected]>
… after setting GuidedDecodingParams (vllm-project#16767) Signed-off-by: chaunceyjiang <[email protected]> Signed-off-by: Russell Bryant <[email protected]> Co-authored-by: Russell Bryant <[email protected]>
… after setting GuidedDecodingParams (vllm-project#16767) Signed-off-by: chaunceyjiang <[email protected]> Signed-off-by: Russell Bryant <[email protected]> Co-authored-by: Russell Bryant <[email protected]> Signed-off-by: Agata Dobrzyniewicz <[email protected]>
… after setting GuidedDecodingParams (vllm-project#16767) Signed-off-by: chaunceyjiang <[email protected]> Signed-off-by: Russell Bryant <[email protected]> Co-authored-by: Russell Bryant <[email protected]> Signed-off-by: Mu Huai <[email protected]>
… after setting GuidedDecodingParams (vllm-project#16767) Signed-off-by: chaunceyjiang <[email protected]> Signed-off-by: Russell Bryant <[email protected]> Co-authored-by: Russell Bryant <[email protected]> Signed-off-by: minpeter <[email protected]>
FIX #16738
Fix the issue where llm.generate cannot be called repeatedly after setting GuidedDecodingParams