You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I originally had this as a discussion but as UV works it seems like a valid issue.
I need llama-cpp-python with cuda, according the installation docs I need to run CMAKE_ARGS="-DGGML_CUDA=on" pip install llama-cpp-python
I've tried CMAKE_ARGS="-DGGML_CUDA=on" rye add llama-cpp-python and this doesn't work, I've also tried setting export CMAKE_ARGS=-DGGML_CUDA=on and including this in a .env file and then running rye add llama-cpp-python. All of these approaches give me the CPU only version
FORCE_CMAKE="1" : Ensures that CMake is re-run even if existing build files are present, which is necessary when changing compiler settings. CMAKE_ARGS="-DGGML_CUDA=ON -DCMAKE_CXX_COMPILER=/usr/bin/g++-12" : -DGGML_CUDA=ON : Enables CUDA support in the build process. -DCMAKE_CXX_COMPILER=/usr/bin/g++-12 : Explicitly sets the C++ compiler to g++-12. This is required because your version of nvcc doesn't fully support GCC 13 yet, and you need to use the older g++-12 for compatibility.
Note: The NVCC_PREPEND_FLAGS approach is not the most reliable way to set the host compiler in newer CUDA versions. Directly setting CMAKE_CXX_COMPILER is the preferred method.
➜ nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2024 NVIDIA Corporation
Built on Wed_Aug_14_10:10:22_PDT_2024
Cuda compilation tools, release 12.6, V12.6.68
Build cuda_12.6.r12.6/compiler.34714021_0
gcc version :
➜ gcc --version
gcc (Ubuntu 13.2.0-23ubuntu4) 13.2.0
Copyright (C) 2023 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
I originally had this as a discussion but as UV works it seems like a valid issue.
I need llama-cpp-python with cuda, according the installation docs I need to run
CMAKE_ARGS="-DGGML_CUDA=on" pip install llama-cpp-python
I've tried
CMAKE_ARGS="-DGGML_CUDA=on" rye add llama-cpp-python
and this doesn't work, I've also tried settingexport CMAKE_ARGS=-DGGML_CUDA=on
and including this in a.env
file and then runningrye add llama-cpp-python
. All of these approaches give me the CPU only versionHowever, this works
Also this works
Originally posted by @this-josh in #1270
The text was updated successfully, but these errors were encountered: