TinyLlama

Running a Local LLM on RISC-V: Building llama.cpp on a Banana Pi F3 (Part 1)

Running a Local LLM on RISC-V: Building llama.cpp on a Banana Pi F3 (Part 1)

TL;DR I built llama.cpp from source on a Banana Pi F3 (SpacemiT K1, riscv64), ran TinyLlama 1.1B, and got an OpenAI-compatible API server running at …