This guide walks through setting up the AI-Q blueprint for local development. For containerized or production deployments, refer to Deployment.
| Requirement | Version | Notes |
|---|---|---|
| Python | 3.11 -- 3.13 | 3.13 recommended |
| uv | latest | Python package manager (installed automatically by the setup script if missing) |
| Git | 2.x+ | |
| Node.js | 22+ | Optional -- only needed for the web UI |
You also need at least one LLM API key. Refer to API key setup below.
Dependency Note: This release is pinned to NeMo Agent Toolkit (NAT) v1.4.0 (nvidia-nat==1.4.0). NAT v1.5 or later is not yet supported by AI-Q and upgrading may introduce breaking changes. The pin will be lifted in a future AI-Q release once compatibility has been validated.
When using NVIDIA API Catalog (the default), inference runs on NVIDIA-hosted infrastructure and there are no local GPU requirements. The hardware requirements below apply only when self-hosting models via NVIDIA NIM.
| Component | Default Model | Self-Hosted Hardware Reference |
|---|---|---|
| LLM (intent classifier, orchestrator, planner) | nvidia/nemotron-3-nano-30b-a3b |
Nemotron 3 Nano support matrix |
| LLM (deep research researcher) | nvidia/nemotron-3-nano-30b-a3b (default) or nvidia/nemotron-3-super-120b-a12b (optional) |
Nemotron 3 Nano support matrix, Nemotron 3 Super support matrix |
| LLM (deep research orchestrator/planner, optional) | openai/gpt-oss-120b |
GPT OSS support matrix |
| Document summary (optional) | nvidia/nemotron-mini-4b-instruct |
Nemotron Mini 4B |
| Text embedding | nvidia/llama-nemotron-embed-vl-1b-v2 |
NeMo Retriever embedding support matrix |
| VLM (image/chart extraction, optional) | nvidia/nemotron-nano-12b-v2-vl |
Vision language model support matrix |
| Knowledge layer (Foundational RAG, optional) | -- | RAG Blueprint support matrix |
The setup script handles everything -- virtual environment, Python dependencies, and UI dependencies:
git clone <repository-url>
cd aiq
./scripts/setup.shThe script performs the following steps:
- Installs
uvif not already present - Creates a Python 3.13 virtual environment at
.venv/ - Installs the core package with dev dependencies
- Installs all frontends (CLI, debug console, API server)
- Installs benchmark packages (freshqa, deepsearch_qa)
- Installs all data source plugins (Tavily, Google Scholar, knowledge layer)
- Sets up pre-commit hooks
- Copies
deploy/.env.exampletodeploy/.envif no.envfile exists - Installs UI npm dependencies (if Node.js is available)
After the script completes, activate the virtual environment:
source .venv/bin/activateIf you prefer to install components selectively, follow these steps.
git clone https://github.com/NVIDIA-AI-Blueprints/aiq.git
cd aiquv venv --python 3.13 .venv
source .venv/bin/activateInstall the core package and only the frontends, benchmarks, and data sources you need:
# Core with development dependencies
uv pip install -e ".[dev]"
# Frontends (pick what you need)
uv pip install -e ./frontends/cli # CLI interface
uv pip install -e ./frontends/debug # Debug console
uv pip install -e ./frontends/aiq_api # Unified API server (includes debug)
# Data sources (pick what you need)
uv pip install -e ./sources/tavily_web_search
uv pip install -e ./sources/google_scholar_paper_search
uv pip install -e "./sources/knowledge_layer[llamaindex,foundational_rag]"
# Benchmarks (optional)
uv pip install -e ./frontends/benchmarks/freshqa
uv pip install -e ./frontends/benchmarks/deepsearch_qapre-commit installAI-Q needs API keys to access LLMs and search providers. Create an environment file from the provided template:
cp deploy/.env.example deploy/.envThen edit deploy/.env and fill in your keys.
| Variable | Provider | How to obtain |
|---|---|---|
NVIDIA_API_KEY |
NVIDIA Build | Sign in, click any model, select Deploy > Get API Key > Generate Key |
| Variable | Provider | Purpose |
|---|---|---|
TAVILY_API_KEY |
Tavily | Web search |
SERPER_API_KEY |
Serper | Academic paper search (Google Scholar). To enable, uncomment paper_search_tool in your config file |
At minimum, you need NVIDIA_API_KEY for LLM inference and TAVILY_API_KEY for web search. Paper search (SERPER_API_KEY) is disabled by default in the shipped configs -- refer to the comments in your config file to enable it.
Confirm that the NeMo Agent Toolkit CLI is available and can find the project plugins:
# Must use the project venv, not the system nat
.venv/bin/nat --helpYou should observe the nat CLI help output with available commands (run, serve, eval, etc.).
To verify plugins are registered:
.venv/bin/nat run --helpThis should list available workflow configurations.
The project documentation is built with Sphinx and uses MyST-Parser for Markdown support. To build the HTML docs locally:
# Install docs dependencies and build in one step
uv run --extra docs sphinx-build -M html docs/source docs/buildThe generated site is written to docs/build/html/. Open docs/build/html/index.html in a browser to view it.
If you already have the virtual environment activated with docs extras installed, you can also run:
sphinx-build -M html docs/source docs/build- Quick Start -- Run your first research query in 5 minutes
- Developer Guide -- Recommended reading path through the documentation
- Deployment -- Docker Compose deployment