The easiest, smallest and fastest local LLM runtime and API server.

bash <(curl -sSfL 'https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/run-llm.sh')
Get Started

Powered by WasmEdge and Rust.

Lightweight

Total dependency of LlamaEdge is 30MB vs 5GB for Python

Very fast

Automagically use the device’s local hardware and software acceleration.