Run AI models directly in the browser — no server, no API keys, data never leaves your device. Each example loads a real model via Web Workers so the page stays responsive.
GitHub Repository →Instant hardware report with no model download. Shows whether your browser supports WebGPU GPU acceleration, WASM SIMD/Threads, and cross-tab model sharing via SharedWorker. inferis uses this report to pick the fastest execution path automatically.
import { detectCapabilities } from 'inferis-ml';
const caps = await detectCapabilities();
// caps.webgpu.supported → true/false
// caps.wasm.simd → true/false
// caps.hardwareConcurrency → 8
Detecting…