5 min readUpdated Mar 2, 2026

Usage & Tokens

All AI operations in Vantage consume tokens. This page explains how token usage works, how to monitor costs, and strategies for managing consumption.


What Are Tokens?

Tokens are the fundamental unit of AI processing. Every request to an AI provider — whether it's a question to the assistant, a tile summary, or a workflow enrichment — consumes tokens based on:

As a rough guideline, 1 token ≈ 4 characters of English text, or approximately ¾ of a word.


What Consumes Tokens

FeatureToken ImpactTypical Range
AI Assistant (single question)Low–Medium500–2,000 tokens
Tile SummaryLow–Medium300–1,500 tokens
Popup AI ChatMedium500–3,000 tokens per message
AI Enrichment (per row)Low100–500 tokens per row
AI Summary (batch)Medium–High1,000–5,000 tokens
AI Compliance Check (per row)Low100–300 tokens per row
AI Formatter (per row)Low50–200 tokens per row
AI Conditional (per row)Low100–300 tokens per row
AI TranscriberHighDepends on audio/video length

Note: Actual token usage depends on the size of the data, the length of context snippets, conversation history, and the model's response.


Cost Factors

Token costs vary based on several factors:

Provider & Model

More capable models cost more per token:

TierExamplesRelative Cost
BudgetGPT-4o-mini, Mistral Small, DeepSeek Chat$
StandardGPT-4o, Claude 3.5 Sonnet, Gemini Pro$$
PremiumClaude 3 Opus, Gemini Ultra, Mistral Large$$$

Data Volume

Context Length


Monitoring Usage

Usage Dashboard

Access the Usage Overview at Settings → Account → Usage & Tokens.

Scope Options

View usage at different levels:

ScopeShows
UserYour individual token consumption
ClientUsage across your client group
OrganizationTotal consumption for the entire organization

Time Period Options

What You'll See


Filtering & Analysis

The Usage Dashboard supports filtering to zero in on specific patterns:

FilterOptions
Operation typeAssistant, Summary, Workflow, Chat, etc.
CategoryDashboard AI, Workflow AI, System
Time granularityHourly, Daily

Exporting Usage Data

Download your usage data for external analysis or billing reconciliation from the Usage Dashboard.


Cost Management Strategies

1. Choose the Right Model

Use budget-tier models (GPT-4o-mini, Mistral Small) for routine tasks and reserve premium models for complex analysis.

2. Enable Data Sampling

Unless you need 100% data accuracy, disable Process Large Datasets in Settings → AI Features → Query Settings. Sampling processes a representative subset, significantly reducing token usage.

3. Optimize Context Snippets

Keep context snippets concise. A focused 2-sentence company overview is more cost-effective (and often more useful) than a 5-paragraph description.

4. Limit Conversation Length

Long multi-turn AI conversations accumulate token costs as the entire conversation history is resent with each message. Start new conversations for unrelated questions.

5. Review Workflow Volume

AI workflow nodes process data rows individually. A workflow processing 10,000 rows through an AI Enrichment node will make 10,000 AI calls. Consider:

6. Monitor Regularly

Check the Usage Dashboard weekly or monthly to catch unexpected consumption patterns early.


Usage Alerts

Set internal guidelines for token consumption and review the Usage Dashboard regularly. Consider establishing: