Local LLM inference engine supporting GGUF models with hardware acceleration on Metal, CUDA, ANE, WebGPU. Features Flash Attention, MicroLoRA, RoPE, quantization (Q4-Q8, π-Quantization), MoE routing, and streaming tokens for browser and edge deployment.