-
-
Notifications
You must be signed in to change notification settings - Fork 5.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix] Fix decode tokens w. CUDA graph #6757
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, please make sure to run full CI as it is required to merge (or just use auto-merge). To run full CI, you can do one of these:
🚀 |
How do we know this fixes the issue? |
I tested it locally and will add a unit test to catch this later. Just put the PR here to unblock benchmarking first. |
Turns out that |
Signed-off-by: Alvant <[email protected]>
Fixes #6703.
This is not related to speculative decoding particularly but not sure why no unit test failed. Will try to add a unit test in this or a follow-up PR.
cc @cadedaniel @alexm-neuralmagic