-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
🐛 fix input text issue #97
Conversation
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## main #97 +/- ##
==========================================
+ Coverage 55.87% 55.98% +0.11%
==========================================
Files 24 24
Lines 1482 1486 +4
Branches 266 268 +2
==========================================
+ Hits 828 832 +4
Misses 579 579
Partials 75 75 ☔ View full report in Codecov by Sentry. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
I tried out this change out and noticed that the BOS token is included in the text output:
|
@tjohnson31415 thanks a lot for the suggestion to add the prompt directly instead of going down the complicated route of decoding the tokens! |
6e1c073
to
06347f7
Compare
Signed-off-by: Prashant Gupta <[email protected]>
The assumption that it remains a valid way to correlate the request and the response is because vllm does it that way Signed-off-by: Prashant Gupta <[email protected]>
70fcd0f
to
76f6594
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
Description
To fix the recent change in vllm - since output contains the input text they do not pass in the prompt with the result anymore.
How Has This Been Tested?
Tested by locally copying code over to the dev_pod and running a request
Single request
2 requests
Streaming
Merge criteria: