-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
HipVMM bug prevents loading any model on desktop systems #105
Comments
Try adding like this: then save it, and in terminal run |
Sorry, still getting the same error on a clean install. I verified that the flag was present in all the steps of the compilation, however when loading the model it seems to still use HipVMM:
By the way, I tried with the AUR version (koboldcpp-hipblas), and it just works. The version in AUR is 1.82.4.yr0-1.
(Model loaded is DeepSeek-R1-Distill-Qwen-14B-Q6_K.gguf) |
I'm experiencing the same issue on Arch with my 6800 XT, even with the added HIPFLAG. Does not seem to matter which model I load.
|
Looks like the reason might be because the flag to disable VMM with HIP/ROCm is "GGML_HIP_NO_VMM" Could you please try it again, but change |
Sorry I took so long to respond. |
Awesome! I'm glad it fixed it for you. Thanks for informing me about the new flags |
Hi, I'm running koboldcpp-rocm in Arch Linux with my RX 6800 and so far I had no major issues unitl now. After a recent upgrade of system ROCm packages, I'm getting an out of memory error from ggml-cuda.cu,
ggml/src/ggml-cuda/ggml-cuda.cu:444: HipVMM Failure: out of memory
This is, as far as I can tell, a ROCm bug which has been mentioned also by developers of llama.cpp here. However, I'm not sure how to implement the proposed solution which is to avoid using Hip VMM by setting -DGGML_CUDA_NO_VMM=1, however this does not seem to work with koboldcpp-rocm (and specifically in arch, probably they are on Windows, I'm not sure). I have also tried setting
make GGML_USE_VMM=OFF LLAMA_HIPBLAS=1 GPU_TARGETS=gfx1030 -j10
but still get the same error, so I cannot load any model now.
What is the intended way to disable HipVMM?
The text was updated successfully, but these errors were encountered: