bash <(curl -sSfL 'https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/run-llm.sh')
Get StartedPowered by WasmEdge and Rust.
Lightweight
Total dependency of LlamaEdge is 30MB vs 5GB for Python
Very fast
Automagically use the device’s local hardware and software acceleration.
Write once run anywhere, for GPUs
Create an LLM web service on a MacBook, deploy it on a Nvidia device.
Native to the heterogeneous edge
Orchestrate and move an LLM app across CPUs, GPUs and NPUs.
Learn more about LlamaEdge