Local Chat
Run small language models directly in the browser with WebGPU/WASM fallback and OPFS caching.
Aether-SLM lets developers add local inference, local RAG, smart model defaults, and zero API-key AI to web apps. This Hub is both the public demo surface and the stable model delivery origin.
Run small language models directly in the browser with WebGPU/WASM fallback and OPFS caching.
Index private text in the browser and retrieve relevant context without sending documents to a server.
Use one production Hub origin for model provenance, range requests, CORS, and cache coordination.
Loading /models.json...checking