vai v1.31.0: voyage-4-nano local inference
vai v1.31.0 adds voyage-4-nano local inference, giving developers a zero-API-key path into Voyage AI embeddings through a lightweight Python bridge built directly into the CLI.
Release moment
The docs, CLI, and launch messaging now land as one story
This release is where the pixel robot and the product story line up cleanly: local-first onboarding, a friendly CLI, and documentation that carries the same visual cues for getting started, understanding the bridge, and shipping a working retrieval flow.
What shipped
This release makes local inference a first-class part of the vai story.
Highlights:
voyage-4-nanolocal embedding invaivai nano setupfor one-time environment and model setup--localsupport on core embedding and pipeline workflows- documentation updates across installation, models, commands, and guides
- a clearer "start local, scale later" path for the Voyage 4 family
Why nano matters
Before this release, the first-time vai experience usually started with an API key. With voyage-4-nano, you can now install the CLI, set up the local model, and generate embeddings immediately on your machine.
npm install -g voyageai-cli
vai nano setup
vai embed "What is vector search?" --local
That makes local inference the shortest path to a working embedding while still keeping the broader Voyage 4 workflow intact.
The Python bridge
vai is a Node.js CLI. voyage-4-nano runs in Python. This release connects those two pieces with a lightweight Python bridge that:
- creates the local runtime environment
- installs dependencies
- downloads and caches the model
- serves local embedding requests back to the CLI
This is intentionally user-visible. Local inference is easy to use, but it is not magic. Python is part of the local runtime story, and vai nano setup handles that setup for you.
Shared embedding space
The most important product detail is that voyage-4-nano shares embedding space with the rest of the Voyage 4 family:
voyage-4-largevoyage-4voyage-4-lite
That means local-first is not a dead end. You can prototype locally, then move to API-backed Voyage 4 models later as your needs change.
Try it
# One-time setup
vai nano setup
# Check readiness
vai nano status
# Smoke test local inference
vai nano test
# Embed locally
vai embed "What is vector search?" --local
# Run the ingestion pipeline with local embedding
vai pipeline ./docs/ --local --db myapp --collection knowledge --create-index
Updated docs
This release also updates the docs site to make local inference easier to discover and understand:
- Installation
- Local Inference Overview
- Local Inference Setup and Usage
- Voyage 4 Family
vai embedvai pipeline
What this changes for users
If you are new to vai, local inference is now the simplest way to get started.
If you already use the API-backed Voyage 4 models, nano gives you a local development path that fits into the same family and the same broader product story.
