Skip to main content

Agent Page

Route: /agent — Source: ui/src/pages/AgentPage.tsx

Chat interface for the in-app assistant. Backed by Gemma 4 (4B) running natively on Apple Silicon via mlx-vlm, with Ollama as a fallback for non-Mac hosts.

Agent page

Opening the agent panel.

Capabilities

CapabilityWhat it does
InsightsSummarize a labeling job, highlight class imbalance, suggest prompts
ChatAnswer free-form questions about the current dataset
VisionAnalyze a frame or annotation directly (multimodal — pass an image as context)
SuggestRecommend the next action: more frames? new prompt? larger model?

The same model is wired into the AI Insights drawer on the Review page — the agent panel is just a freer interface to the same backend.

Configuration

VarDefaultPurpose
AGENT_MODEL_IDgoogle/gemma-4-e4b-itModel id for the MLX-VLM path
OLLAMA_URLhttp://localhost:11434Fallback Ollama endpoint
OLLAMA_MODELgemma2:4bFallback Ollama model

The page picks the MLX path automatically when mlx-vlm is importable; otherwise it falls back to Ollama. There is no third path — install one or the other.

Privacy

The agent runs locally. Nothing in your dataset, prompts, or frames leaves the host. That's the whole point of running it via MLX or Ollama instead of a hosted API.