r2d4/react-llm

Easy-to-use headless React Hooks to run LLMs in the browser with WebGPU. Just useLLM().

31
/ 100
Emerging

Leverages Apache TVM and MLC Relax compiled to WebAssembly to execute Vicuna models entirely on-device, with model weights and tokenizer (SentencePiece) cached in browser storage for faster subsequent loads. Offloads inference to a WebWorker to avoid blocking the main thread, while conversation state persists to localStorage. The modular monorepo includes companion packages for pre-built retro UI components and a Chrome extension, allowing both headless integration and turnkey deployment.

702 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 9 / 25
Community 12 / 25

How are scores calculated?

Stars

702

Forks

32

Language

TypeScript

License

MIT

Last pushed

Jun 27, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/r2d4/react-llm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.