Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Build] Disable sm_90a in cu11 #5141

Merged

Conversation

simon-mo
Copy link
Collaborator

@simon-mo simon-mo commented May 30, 2024

FAILED: CMakeFiles/_C.dir/csrc/quantization/cutlass_w8a8/scaled_mm_dq_c3x.cu.o 
ccache /usr/local/cuda-11.8/bin/nvcc -forward-unknown-to-host-compiler -DTORCH_EXTENSION_NAME=_C -DUSE_C10D_GLOO -DUSE_C10D_NCCL -DUSE_DISTRIBUTED -DUSE_RPC -DUSE_TENSORPIPE -D_C_EXPORTS -I/home/runner/work/vllm/vllm/csrc -I/home/runner/work/vllm/vllm/build/temp.linux-x86_64-cpython-310/_deps/cutlass-src/include -I/home/runner/work/vllm/vllm/build/temp.linux-x86_64-cpython-310/_deps/cutlass-src/tools/util/include -isystem /opt/hostedtoolcache/Python/3.10.14/x64/include/python3.10 -isystem /opt/hostedtoolcache/Python/3.10.14/x64/lib/python3.10/site-packages/torch/include -isystem /opt/hostedtoolcache/Python/3.10.14/x64/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda-11.8/include -DONNX_NAMESPACE=onnx_c2 -Xcudafe --diag_suppress=cc_clobber_ignored,--diag_suppress=field_without_dll_interface,--diag_suppress=base_class_has_different_dll_interface,--diag_suppress=dll_interface_conflict_none_assumed,--diag_suppress=dll_interface_conflict_dllexport_assumed,--diag_suppress=bad_friend_decl --expt-relaxed-constexpr --expt-extended-lambda -O3 -DNDEBUG -std=c++17 "--generate-code=arch=compute_70,code=[sm_70]" "--generate-code=arch=compute_75,code=[sm_75]" "--generate-code=arch=compute_80,code=[sm_80]" "--generate-code=arch=compute_86,code=[sm_86]" "--generate-code=arch=compute_89,code=[sm_89]" "--generate-code=arch=compute_90,code=[sm_90]" "--generate-code=arch=compute_90,code=[compute_90]" -Xcompiler=-fPIC -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr -DENABLE_FP8 --threads=1 -D_GLIBCXX_USE_CXX11_ABI=0 -gencode arch=compute_90a,code=sm_90a -MD -MT CMakeFiles/_C.dir/csrc/quantization/cutlass_w8a8/scaled_mm_dq_c3x.cu.o -MF CMakeFiles/_C.dir/csrc/quantization/cutlass_w8a8/scaled_mm_dq_c3x.cu.o.d -x cu -c /home/runner/work/vllm/vllm/csrc/quantization/cutlass_w8a8/scaled_mm_dq_c3x.cu -o CMakeFiles/_C.dir/csrc/quantization/cutlass_w8a8/scaled_mm_dq_c3x.cu.o
nvcc fatal   : Unsupported gpu architecture 'compute_90a'

https://github.com/vllm-project/vllm/actions/runs/9307272821/job/25618220413

@simon-mo simon-mo merged commit 45a1a69 into vllm-project:main May 30, 2024
20 of 24 checks passed
blinkbear pushed a commit to blinkbear/vllm that referenced this pull request May 31, 2024
dtrifiro pushed a commit to opendatahub-io/vllm that referenced this pull request May 31, 2024
blinkbear pushed a commit to blinkbear/vllm that referenced this pull request Jun 6, 2024
robertgshaw2-redhat pushed a commit to neuralmagic/nm-vllm that referenced this pull request Jun 8, 2024
joerunde pushed a commit to joerunde/vllm that referenced this pull request Jun 17, 2024
robertgshaw2-redhat pushed a commit to neuralmagic/nm-vllm that referenced this pull request Jul 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants