Skip to content

Latest commit

 

History

History
178 lines (123 loc) · 7.13 KB

File metadata and controls

178 lines (123 loc) · 7.13 KB

Installation

This guide walks through setting up the AI-Q blueprint for local development. For containerized or production deployments, refer to Deployment.

Prerequisites

Requirement Version Notes
Python 3.11 -- 3.13 3.13 recommended
uv latest Python package manager (installed automatically by the setup script if missing)
Git 2.x+
Node.js 22+ Optional -- only needed for the web UI

You also need at least one LLM API key. Refer to API key setup below.

Dependency Note: This release is pinned to NeMo Agent Toolkit (NAT) v1.4.0 (nvidia-nat==1.4.0). NAT v1.5 or later is not yet supported by AI-Q and upgrading may introduce breaking changes. The pin will be lifted in a future AI-Q release once compatibility has been validated.

Hardware Requirements

When using NVIDIA API Catalog (the default), inference runs on NVIDIA-hosted infrastructure and there are no local GPU requirements. The hardware requirements below apply only when self-hosting models via NVIDIA NIM.

Component Default Model Self-Hosted Hardware Reference
LLM (intent classifier, orchestrator, planner) nvidia/nemotron-3-nano-30b-a3b Nemotron 3 Nano support matrix
LLM (deep research researcher) nvidia/nemotron-3-nano-30b-a3b (default) or nvidia/nemotron-3-super-120b-a12b (optional) Nemotron 3 Nano support matrix, Nemotron 3 Super support matrix
LLM (deep research orchestrator/planner, optional) openai/gpt-oss-120b GPT OSS support matrix
Document summary (optional) nvidia/nemotron-mini-4b-instruct Nemotron Mini 4B
Text embedding nvidia/llama-nemotron-embed-vl-1b-v2 NeMo Retriever embedding support matrix
VLM (image/chart extraction, optional) nvidia/nemotron-nano-12b-v2-vl Vision language model support matrix
Knowledge layer (Foundational RAG, optional) -- RAG Blueprint support matrix

Automated Setup (Recommended)

The setup script handles everything -- virtual environment, Python dependencies, and UI dependencies:

git clone <repository-url>
cd aiq

./scripts/setup.sh

The script performs the following steps:

  1. Installs uv if not already present
  2. Creates a Python 3.13 virtual environment at .venv/
  3. Installs the core package with dev dependencies
  4. Installs all frontends (CLI, debug console, API server)
  5. Installs benchmark packages (freshqa, deepsearch_qa)
  6. Installs all data source plugins (Tavily, Google Scholar, knowledge layer)
  7. Sets up pre-commit hooks
  8. Copies deploy/.env.example to deploy/.env if no .env file exists
  9. Installs UI npm dependencies (if Node.js is available)

After the script completes, activate the virtual environment:

source .venv/bin/activate

Manual Setup

If you prefer to install components selectively, follow these steps.

1. Clone the Repository

git clone https://github.com/NVIDIA-AI-Blueprints/aiq.git
cd aiq

2. Create the Virtual Environment

uv venv --python 3.13 .venv
source .venv/bin/activate

3. Install Dependencies

Install the core package and only the frontends, benchmarks, and data sources you need:

# Core with development dependencies
uv pip install -e ".[dev]"

# Frontends (pick what you need)
uv pip install -e ./frontends/cli          # CLI interface
uv pip install -e ./frontends/debug        # Debug console
uv pip install -e ./frontends/aiq_api      # Unified API server (includes debug)

# Data sources (pick what you need)
uv pip install -e ./sources/tavily_web_search
uv pip install -e ./sources/google_scholar_paper_search
uv pip install -e "./sources/knowledge_layer[llamaindex,foundational_rag]"

# Benchmarks (optional)
uv pip install -e ./frontends/benchmarks/freshqa
uv pip install -e ./frontends/benchmarks/deepsearch_qa

4. Set Up Pre-Commit Hooks (Development)

pre-commit install

API Key Setup

AI-Q needs API keys to access LLMs and search providers. Create an environment file from the provided template:

cp deploy/.env.example deploy/.env

Then edit deploy/.env and fill in your keys.

Required Keys

Variable Provider How to obtain
NVIDIA_API_KEY NVIDIA Build Sign in, click any model, select Deploy > Get API Key > Generate Key

Optional Keys

Variable Provider Purpose
TAVILY_API_KEY Tavily Web search
SERPER_API_KEY Serper Academic paper search (Google Scholar). To enable, uncomment paper_search_tool in your config file

At minimum, you need NVIDIA_API_KEY for LLM inference and TAVILY_API_KEY for web search. Paper search (SERPER_API_KEY) is disabled by default in the shipped configs -- refer to the comments in your config file to enable it.

Verify Installation

Confirm that the NeMo Agent Toolkit CLI is available and can find the project plugins:

# Must use the project venv, not the system nat
.venv/bin/nat --help

You should observe the nat CLI help output with available commands (run, serve, eval, etc.).

To verify plugins are registered:

.venv/bin/nat run --help

This should list available workflow configurations.

Building the Documentation

The project documentation is built with Sphinx and uses MyST-Parser for Markdown support. To build the HTML docs locally:

# Install docs dependencies and build in one step
uv run --extra docs sphinx-build -M html docs/source docs/build

The generated site is written to docs/build/html/. Open docs/build/html/index.html in a browser to view it.

If you already have the virtual environment activated with docs extras installed, you can also run:

sphinx-build -M html docs/source docs/build

Next Steps