-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GGUF (Breaking Change to Model Files) #633
Conversation
Hello, excited about trying this out with the CodeLlama gguf model. Followed the MacOS (Sonoma beta) instructions. How do I get the 'llama' shared library? llama-cpp-python % python3 -m llama_cpp.server --model $MODEL --n_gpu_layers 1 Thanks! |
@sndani try reinstalling with the |
@abetlen thanks for the great work and thanks for responding. Turns out, cmake isn't getting to target libllama.so under vendor/llama.cpp (but 'make clean' try deleting it). This is a dev fix but will open an issue (or the next person who encounters this can) if this isn't only my environment for some reason. % make clean |
My model is now a lot slower... is there any solution to fix this? |
Hi, i always get this error.
|
Hi abdullrahmanesmael07, if you are using windows, install visual studio community with c++ options and reboot. It will works. |
GGUF support for llama.cpp Closes #628
Currently works to update your old ggml v3 llama models run
TODO
Move convert script into package to make it easier for people to migrateAdd docs link to conversion script in llama.cpp