Intelligence at the Edge,
In a Single API

The centralized cloud is too slow, too fragile, and too controlled. QVAC is the foundational SDK for local-first, decentralized AI. Embed intelligence that runs everywhere. Privately, instantly, and without permission.

SDK
npm install @qvac/sdk import { completion, LLAMA_3_2_1B_INST_Q4_0, loadModel, unloadModel } from "@qvac/sdk"; // Supports any Pear or HTTP URL const modelId = await loadModel({ modelSrc: LLAMA_3_2_1B_INST_Q4_0, modelType: "llm", }); const history = [ { role: "user", content: "QVAC, how may entropy be reversed?", }, ]; const result = completion({ modelId, history, stream: true, }); for await (const token of result.tokenStream) { console.log(token); } await unloadModel({ modelId });

Cross-platform AI for all your platforms

Run AI models natively across any operative system, any platform, and any device. Build on popular JavaScript platforms such as Node.js, Expo, Bare, or Bun. The SDK abstracts away platform complexity while providing consistent AI capabilities whether you're building on desktop, mobile, or a server.

Decentralization that doesn’t get in the way

We baked in the entire Pears.com stack to enable decentralized model sharing, delegated inference and allow for decentralized vector databases. P2P is native but optional. You can also run RAG using your favorite vector database (Chroma, LanceDB, SQLite-vector, and more). And we support fetching models from any HTTP provider, like HuggingFace.

SDK
import { loadModel, unloadModel, GTE_LARGE_FP16, ragSaveEmbeddings, ragSearch } from "@qvac/sdk"; const query = "machine learning algorithms"; const samples = [ "sample 1", "sample 2" ]; const modelId = await loadModel({ modelSrc: GTE_LARGE_FP16, modelType: "embeddings" }); const docs = await ragSaveEmbeddings({ modelId, documents: samples, chunk: false, }); const results = await ragSearch({ modelId, query, topK: 3, }); results.forEach((result, index) => { console.log(`${result.content}`); }); await unloadModel({ modelId });

Local AI that scales

Create distributed AI inference networks where devices can provide or consume AI services. Enable resource sharing across the network, allowing lightweight devices to access powerful AI models running on other peers.

SDK
import { startQVACProvider } from "@qvac/sdk"; const topic = "some-topic"; const response = await startQVACProvider({ topic, firewall: undefined, });

One SDK, All of AI

Seamlessly integrate multiple AI capabilities including completion, transcription, tool calling, embeddings, and retrieval, translation, vision or text-to-speech using a single entrypoint. It also supports streaming and multimodal inputs.

SDK
import { loadModel, textToSpeech, unloadModel, TTS_PIPER_NORMAN_EN_US_ONNX_MEDIUM, TTS_PIPER_NORMAN_EN_US_ONNX_MEDIUM_CONFIG } from "@qvac/sdk"; const eSpeakDataPath = "some path"; try { const modelId = await loadModel({ modelSrc: TTS_PIPER_NORMAN_EN_US_ONNX_MEDIUM, modelType: "tts", configSrc: TTS_PIPER_NORMAN_EN_US_ONNX_MEDIUM_CONFIG, eSpeakDataPath, modelConfig: { language: "en", } }); const result = textToSpeech({ modelId, text: `QVAC SDK is the canonical entry point to QVAC`, inputType: "text", stream: false, }); const audioBuffer = await result.buffer; // now you can convert to wav and play await unloadModel({ modelId }); }

FAQ

With cloud AI APIs, your data is sent to third-party servers for processing, you pay per request, and you need a constant internet connection. The QVAC SDK runs AI models directly on your own device. That means your data never leaves your hardware, there are no per-request costs, no rate limits, and no dependency on an internet connection once you have a model downloaded. You own the entire pipeline.