Error: llama runner process has terminated: exit status 2 when running ollama
I run this command:
ollama run hf.co/openbmb/MiniCPM-o-2_6-gguf:Q4_K_M
and has this error:
Error: llama runner process has terminated: exit status 2
Specs:
- CUDA: v12
- GPU T4 Tesla
- Running in SageMaker instance (AWS)
- Python 3.11.11
Model has been downloaded but failed to run. I tried ollama list and confirmed that model has been downloaded:
hf.co/openbmb/MiniCPM-o-2_6-gguf:Q4_K_M 577cda2e3886 5.7 GB 21 seconds ago
I also tried running on llama-cpp-python but it also has a lot of errors.
Please help. I am desperate.
Thank you very much.
I get an error message as well running this on Ollama. I'm not desperate though, I hope this can be solved in the coming weeks. π
same problem
Hello, it is based on llama.cpp and can use gguf to inference, then my changes have not been merged into the official yet, and you may need to wait until the merge.