Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CI/Build] bump minimum cmake version #6999

Merged
merged 1 commit into from
Aug 12, 2024

Conversation

dtrifiro
Copy link
Contributor

@dtrifiro dtrifiro commented Jul 31, 2024

  • find_python_from_executable uses find_package(PYTHON ... Development.SABIModule), which is only available since cmake>=3.26
  • cmake and ninja are build-time requirements, so they should not be in requirements-common.txt or requirements-openvino.txt

Copy link

👋 Hi! Thank you for contributing to the vLLM project.
Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which consists a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of default ones by unblocking the steps in your fast-check build on Buildkite UI.

Once the PR is approved and ready to go, please make sure to run full CI as it is required to merge (or just use auto-merge).

To run full CI, you can do one of these:

  • Comment /ready on the PR
  • Add ready label to the PR
  • Enable auto-merge.

🚀

@dtrifiro dtrifiro force-pushed the bump-minimum-cmake-version branch 2 times, most recently from f0462f1 to 9981b80 Compare July 31, 2024 17:51
@dtrifiro
Copy link
Contributor Author

Some extra context for this: having an older cmake version would result in a cryptic error when building

  Unable to find python matching:
  /home/dtrifiro/work/vllm/.venv/bin/python.
Call Stack (most recent call first):
  CMakeLists.txt:43 (find_python_from_executable)

Copy link
Contributor

@bnellnm bnellnm left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!

@dtrifiro
Copy link
Contributor Author

CI failures seem unrelated:

test_traces
...
ValueError: OpenTelemetry packages must be installed before configuring 'otlp_traces_endpoint'

@dtrifiro
Copy link
Contributor Author

dtrifiro commented Aug 1, 2024

/ready

@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Aug 1, 2024
@dtrifiro dtrifiro force-pushed the bump-minimum-cmake-version branch from 9981b80 to 4f4d5a0 Compare August 6, 2024 09:23
@dtrifiro
Copy link
Contributor Author

dtrifiro commented Aug 6, 2024

Anything holding this back?

@mgoin
Copy link
Member

mgoin commented Aug 6, 2024

@dtrifiro looking at the failing check logs, it seems the Intel CPU test cannot find its cmake executable with this change. It would be great if you could investigate this

@dtrifiro
Copy link
Contributor Author

dtrifiro commented Aug 8, 2024

@mgoin I opened #7298 to take care of the Dockerfile.cpu issues, this should go through once that's merged.

@dtrifiro dtrifiro force-pushed the bump-minimum-cmake-version branch from 4f4d5a0 to f4f3775 Compare August 9, 2024 08:16
@dtrifiro
Copy link
Contributor Author

dtrifiro commented Aug 9, 2024

Looking at the failure logs, it seems that they're unrelated:

entrypoints/openai/test_completion.py::test_single_completion[-zephyr-pa-8] INFO 08-09 08:48:57 logger.py:36] Received request cmpl-c29cba85cf3a4c7a930f4aa562d54614-0: prompt: 'Hello, my name is', params: SamplingParams(n=1, best_of=1, presence_penalty=0.0, frequency_penalty=0.0, repetition_penalty=1.0, temperature=0.0, top_p=1.0, top_k=-1, min_p=0.0, seed=None, use_beam_search=False, length_penalty=1.0, early_stopping=False, stop=[], stop_token_ids=[], include_stop_str_in_output=False, ignore_eos=False, max_tokens=5, min_tokens=0, logprobs=None, prompt_logprobs=None, skip_special_tokens=True, spaces_between_special_tokens=True, truncate_prompt_tokens=None), prompt_token_ids: [1, 22557, 28725, 586, 1141, 349], lora_request: None, prompt_adapter_request: PromptAdapterRequest(prompt_adapter_name='zephyr-pa', prompt_adapter_id=1, prompt_adapter_local_path='/root/.cache/huggingface/hub/models--swapnilbp--llama_tweet_ptune/snapshots/ad7faa5e642dd3f4311c927b0c423402444cd7a8', prompt_adapter_num_virtual_tokens=8).
[2024-08-09T08:48:57Z] INFO 08-09 08:48:57 async_llm_engine.py:199] Added request cmpl-c29cba85cf3a4c7a930f4aa562d54614-0.
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61] Engine background task failed
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61] Traceback (most recent call last):
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 51, in _log_task_completion
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     return_value = task.result()
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 772, in run_engine_loop
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     result = task.result()
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 715, in engine_step
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     request_outputs = await self.engine.step_async(virtual_engine)
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 282, in step_async
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     output = await self.model_executor.execute_model_async(
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/executor/gpu_executor.py", line 159, in execute_model_async
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     output = await make_async(self.driver_worker.execute_model
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/lib/python3.10/concurrent/futures/thread.py", line 58, in run
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     result = self.fn(*self.args, **self.kwargs)
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/worker_base.py", line 236, in execute_model
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     self.model_runner.prepare_model_input(
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 1435, in prepare_model_input
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     model_input = self._prepare_model_input_tensors(
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 1006, in _prepare_model_input_tensors
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     builder.add_seq_group(seq_group_metadata)
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 638, in add_seq_group
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     per_seq_group_fn(inter_data, seq_group_metadata)
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 590, in _compute_prompt_adapter_input
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61]     assert inter_data.n_seqs == 1
[2024-08-09T08:48:57Z] ERROR 08-09 08:48:57 async_llm_engine.py:61] AssertionError
[2024-08-09T08:48:57Z] Exception in callback _log_task_completion(error_callback=<bound method...7fec46856bc0>>)(<Task finishe...ertionError()>) at /usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py:41
[2024-08-09T08:48:57Z] handle: <Handle _log_task_completion(error_callback=<bound method...7fec46856bc0>>)(<Task finishe...ertionError()>) at /usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py:41>
[2024-08-09T08:48:57Z] Traceback (most recent call last):
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 51, in _log_task_completion
[2024-08-09T08:48:57Z]     return_value = task.result()
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 772, in run_engine_loop
[2024-08-09T08:48:57Z]     result = task.result()
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 715, in engine_step
[2024-08-09T08:48:57Z]     request_outputs = await self.engine.step_async(virtual_engine)
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 282, in step_async
[2024-08-09T08:48:57Z]     output = await self.model_executor.execute_model_async(
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/executor/gpu_executor.py", line 159, in execute_model_async
[2024-08-09T08:48:57Z]     output = await make_async(self.driver_worker.execute_model
[2024-08-09T08:48:57Z]   File "/usr/lib/python3.10/concurrent/futures/thread.py", line 58, in run
[2024-08-09T08:48:57Z]     result = self.fn(*self.args, **self.kwargs)
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/worker_base.py", line 236, in execute_model
[2024-08-09T08:48:57Z]     self.model_runner.prepare_model_input(
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 1435, in prepare_model_input
[2024-08-09T08:48:57Z]     model_input = self._prepare_model_input_tensors(
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 1006, in _prepare_model_input_tensors
[2024-08-09T08:48:57Z]     builder.add_seq_group(seq_group_metadata)
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 638, in add_seq_group
[2024-08-09T08:48:57Z]     per_seq_group_fn(inter_data, seq_group_metadata)
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/worker/model_runner.py", line 590, in _compute_prompt_adapter_input
[2024-08-09T08:48:57Z]     assert inter_data.n_seqs == 1
[2024-08-09T08:48:57Z] AssertionError
[2024-08-09T08:48:57Z]
[2024-08-09T08:48:57Z] The above exception was the direct cause of the following exception:
[2024-08-09T08:48:57Z]
[2024-08-09T08:48:57Z] Traceback (most recent call last):
[2024-08-09T08:48:57Z]   File "/usr/lib/python3.10/asyncio/events.py", line 80, in _run
[2024-08-09T08:48:57Z]     self._context.run(self._callback, *self._args)
[2024-08-09T08:48:57Z]   File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 63, in _log_task_completion
[2024-08-09T08:48:57Z]     raise AsyncEngineDeadError(
[2024-08-09T08:48:57Z] vllm.engine.async_llm_engine.AsyncEngineDeadError: Task finished unexpectedly. This should never happen! Please open an issue on Github. See stack trace above for the actual cause.
[2024-08-09T08:48:57Z] WARNING 08-09 08:48:57 server.py:103] Failed to abort request cmpl-c29cba85cf3a4c7a930f4aa562d54614-0

@dtrifiro dtrifiro force-pushed the bump-minimum-cmake-version branch from f4f3775 to 0a74f7b Compare August 12, 2024 08:48
- `find_python_from_executable` uses `find_package(PYTHON ... Development.SABIModule)`, which
is only available [since cmake>=3.26](https://cmake.org/cmake/help/latest/module/FindPython.html)
- `cmake` and `ninja` are build-time requirements, so they should not be
  in `requirements-common.txt` or `requirements-openvino.txt`
Copy link
Member

@njhill njhill left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @dtrifiro

@njhill njhill merged commit 774cd1d into vllm-project:main Aug 12, 2024
68 checks passed
kylesayrs pushed a commit to neuralmagic/vllm that referenced this pull request Aug 17, 2024
fialhocoelho pushed a commit to opendatahub-io/vllm that referenced this pull request Aug 22, 2024
Alvant pushed a commit to compressa-ai/vllm that referenced this pull request Oct 26, 2024
KuntaiDu pushed a commit to KuntaiDu/vllm that referenced this pull request Nov 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ready ONLY add when PR is ready to merge/full CI is needed
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants