
./llama-server -m ../../models/ToolACE-2-Llama-3.1-8B-Q4_K_M.ggml --n-gpu-layers 60 --host 0.0.0.0 --jinja