llama-rs
Rust bindings to llama.cpp, for macOS, with metal support, for testing and evaluating whether it would be worthwhile to run an Llama model locally in a Rust app.
Setup
- Clone llama.cpp into
vendor/llama.cpp
- Build llama.cpp:
LLAMA_METAL=1 make
- Download a llama2 model: https://huggingface.co/TheBloke/Llama-2-7B-GGML/tree/main
- Convert the model to llama.cpp's GGUF format using the script in the
llama.cpp
repo.