Conversation
|
Hello, excited about trying this out with the CodeLlama gguf model. Followed the MacOS (Sonoma beta) instructions. How do I get the 'llama' shared library? llama-cpp-python % python3 -m llama_cpp.server --model $MODEL --n_gpu_layers 1 Thanks! |
|
@sndani try reinstalling with the |
|
@abetlen thanks for the great work and thanks for responding. Turns out, cmake isn't getting to target libllama.so under vendor/llama.cpp (but 'make clean' try deleting it). This is a dev fix but will open an issue (or the next person who encounters this can) if this isn't only my environment for some reason. % make clean |
|
My model is now a lot slower... is there any solution to fix this? |
|
Hi, i always get this error. |
|
Hi abdullrahmanesmael07, if you are using windows, install visual studio community with c++ options and reboot. It will works. |
GGUF support for llama.cpp Closes #628
Currently works to update your old ggml v3 llama models run
TODO
Move convert script into package to make it easier for people to migrateAdd docs link to conversion script in llama.cpp