Ai

NanoClaw on RISC-V: Running an AI Agent Runtime on a Banana Pi F3

NanoClaw on RISC-V: Running an AI Agent Runtime on a Banana Pi F3

Photo by Tom Fisk on Pexels Can you run a full AI agent orchestration runtime on a $100 RISC-V board? My money was on "probably not," but I …

Benchmarking llama.cpp on SpacemiT K3: RISC-V AI Cores vs Standard RVV (Part 4)

Benchmarking llama.cpp on SpacemiT K3: RISC-V AI Cores vs Standard RVV (Part 4)

TL;DR SpacemiT's K3 has two core types: X100 (general-purpose, vlen 256) and A100 ("AI cores", vlen 1024). Standard llama.cpp runs 2.3x …

Your AI Agents Need Their Own Computers

Your AI Agents Need Their Own Computers

Part 1 of 3: Why the future of AI-assisted development runs on hypervisors Photo by Cookiecutter on Pexels You know that moment when you read …

Running a Local LLM on RISC-V: Building llama.cpp on a Banana Pi F3 (Part 1)

Running a Local LLM on RISC-V: Building llama.cpp on a Banana Pi F3 (Part 1)

TL;DR I built llama.cpp from source on a Banana Pi F3 (SpacemiT K1, riscv64), ran TinyLlama 1.1B, and got an OpenAI-compatible API server running at …

How Docker Gave My 'Unsupported' GPU a Second Life

How Docker Gave My 'Unsupported' GPU a Second Life

Photo by Vanessa Frosali on Unsplash I became a Docker Captain in January 2026. After more than a decade of using Docker, 37 images on Docker Hub, …

Bringing AI Agents to RISC-V64: The First cagent Distribution Package

Bringing AI Agents to RISC-V64: The First cagent Distribution Package

Bringing AI Agents to RISC-V64: The First cagent Distribution Package <span id="author" class="author">Bruno …