-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
b'S3 not supported' When Using RunAI Model Streamer with S3 in VLLM #34
Comments
Hey @purp1e-ace ! Though not entirely specified in the error what cause the engine process to fail, we have had issue with libz in version 0.12.0 (#32). We have had removed this version from PyPi, and now the latest version is 0.11.2. Please try to install it. |
I found that this happened because the linker failed to find libstreamers3.so because zlib is not installed, which is required by libstreamers3.so. So I solved this by install zlib. But after that, I got another error similar to #28 .
|
I got it. The vllm document here Loading models with Run:ai Model Streamer is incorrect. When you trying to run model from a S3 compatible object store, you should set env RUNAI_STREAMER_S3_ENDPOINT instead of AWS_ENDPOINT_URL as mentioned in vllm document. |
@purp1e-ace thanks for the update! Are you reading from AWS S3 bucket or from another object storage provider such as GCS (google cloud) or Minio? If reading from AWS S3, there is no need to specify any of the flags The flag You can both read from compatible S3 storage like GCS with the appropriate flags and use the memory limit
or by running
Setting the environment variable |
@noa-neria |
Thank you for noting this! In fact, currently both flags are needed We will fix this to use a single flag. Sorry for the inconvenience |
hello @noa-neria! sry for disturb on that closed issue. But I am still confused about this same issue. Here is some screenshot about my env and commands: ![]() ![]() Hoping for your advice! |
Can you please try with the latest vLLM (and runai) version? Also, in latest version |
Hi @noa-neria, thanks for quick reply! I am using runai with version For the second, I have noticed that in the lastest version |
Can you please run
If both exist, please run with enabled logs |
Thanks for advice! I tried with log, and found I did not satisfy |
I am encountering a persistent issue when attempting to serve a model from an S3 bucket using the vllm serve command with the --load-format runai_streamer option. Despite having proper access to the S3 bucket and all required files being present, the process fails with a b'S3 not supported'. Below are the details of the issue:
Command Used:
AWS_ACCESS_KEY_ID=my_ak AWS_SECRET_ACCESS_KEY=my_sk python -m vllm.entrypoints.openai.api_server --model s3://ai-peta-model-storage-bucket/model/qwen/Qwen2.5-1.5B-Instruct/main --load-format runai_streamer --dtype half
Error Message:
Environment Details:
VLLM version: 0.7.1
Python version: 3.10
RunAI Model Streamer version: 0.12.0
The text was updated successfully, but these errors were encountered: