Browser-based LLM inference using WebGPU for RuVector ecosystem, enabling lightweight AI model execution in WASM environments.
Loading more......
ruvllm
Local LLM inference engine supporting GGUF models with hardware acceleration on Metal, CUDA, ANE, WebGPU. Features Flash Attention, MicroLoRA, RoPE, quantization (Q4-Q8, π-Quantization), MoE routing, and streaming tokens for browser and edge deployment.
micro-hnsw-wasm
WASM library for brain-inspired neuromorphic HNSW vector search in 11.8KB. Optimized for edge devices with spiking neurons for energy-efficient similarity search.
ClickHouse
ClickHouse is a columnar OLAP database with vector indexes (ANN via AMM, brute-force), supporting SQL queries over vectors + structured data at petabyte scale. Excels in aggregations with vectors. For analytics workloads with embeddings; faster ingestion than Postgres pgvector for big data.
RuVector
Self-optimizing on-device vector database with HNSW, graph RAG, and WASM deployment for low-latency edge AI ops across browsers/IoT/mobile. Supports real-time self-learning retrieval; lighter and offline vs cloud Qdrant.
ruvector-attention-unified-wasm
Unified WASM bindings for 18+ attention mechanisms including neural, DAG, and Mamba SSM, optimized for vector search and processing.
ruvector-economy-wasm
CRDT-based autonomous credit economy in WASM for decentralized vector resource allocation and AI agent economics.
WASM module for LLM inference in browsers with WebGPU acceleration.
Open-source, free.