-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Can't install GPU version for windows for many times. #1393
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I checked #1352 , and is there an issus related to windows 11? I just thought it this the problem of my installation steps or my machine. Is there an official explaination plz !!! |
Im having the same problem but with Linux, 20,04, using Kaggle Notebook, worked fine until yesterday. edit: |
Still not working, I have been trying 0.2.64, 0.2.60, 0.2.59 for many times and it seems to said:
|
The lama.dll is Win32 and we are 64 Bit on Windows 11, if I debug the C++ checker program in Win32 then the lama.dll loads successfully, but for 64 bit nope. int main() {
} |
|
cuda version i'm using I have tried this from Windows PowerShell and it works for me
|
I'm having the same issue. I have CUDA installed, nvcc works, and CUDA_PATH is set. Don't see any errors in the installation. AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 0 | Getting the same result with: |
It seems to be the problem of windows 11, i made it worked in windows 10 months ago (while the version of llama-cpp-python==0.1.72), but when i turn into the latest version with win11 it doesn't work :( |
Yeap, if you make it work plz let me know :) I will keep trying to find the solution as well. |
Although it didn't work initially, I was able to download the prebuilt wheel and it works, now I am getting GPU inference. It does seem like there is an issue with my environment in some way. |
i made a guide may be its helpful https://github.com/Granddyser/windows-llama-cpp-python-cuda-guide |
Issues
I am trying to install the lastest version of llama-cpp-python in my windows 11 with RTX-3090ti(24G). I have successfully installed llama-cpp-python=0.1.87 (can't exactly remember) months ago while using:
But when I want to access the latest version recently by using:
After loading the model, it is still using CPU with
BLAS=0
(or is another params = 1 instead ofBLAS
in new version?).I have been trying the pre-build wheel for CUDA 12.1 (
pip install llama-cpp-python --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu121
) and it still doesn't work. I add--verbose
to see the output:Environment
python=3.12
C++ compiler: viusal studio 2022 (with necessary C++ modules)
cmake --version
= 3.29.2nvcc -V
= CUDA 12.1 (whilenvidia-smi
cuda version is 12.3, i think it is not related to this issues)I have been download and install VS2022, CUDA toolkit, cmake and anaconda, I am wondering if some steps are missing. Considering my previous experience there is no need to git clone this code and cd into it to build (Though I did that on my mac to build a pth file to bin file months ago).
My system variables are listed below:
Questions
pip install llama-cpp-python
instead of loading model to checkBLAS=1.
pip install llama-cpp-py
?The text was updated successfully, but these errors were encountered: