Local inference engine
Local inference engine (repository name changed to comply with dependency manager)
[!WARNING] Windows CLI is built without
libcurl. That means network feature such asllama-server -hf elyza/Llama-3-ELYZA-JP-8B-GGUF:Q4_K_Monly works on Mac.
x86_64GGML_CPU to FALSECMAKE_OSX_ARCHITECTURES to x86_64LLAMA_CURL to FALSEc.f. https://github.com/ggml-org/llama.cpp/issues/9937
BUILD_SHARED_LIBS to FALSE