Skip to main content

Roadmap

What we shipped, what we are building next, and what we plan to build.

Last Shipped

PDF Support in the Playground
12/17/2025
PlaygroundEvaluationObservability
Attach PDF documents to chat messages in the playground. Upload files, provide URLs, or use file IDs from provider APIs. Works with OpenAI, Gemini, and Claude models. PDFs are supported in evaluations and observability traces.
Provider Built-in Tools in the Playground
12/11/2025
Playground
Use provider built-in tools like web search, code execution, and file search directly in the Playground. Supported providers include OpenAI, Anthropic, and Gemini. Tools are saved with prompts and automatically used via the LLM gateway.
Projects within Organizations
12/4/2025
Misc
Create projects within organizations to divide work between different AI products. Each project scopes its prompts, traces, and evaluations independently.
Jinja2 Template Support in the Playground
11/17/2025
Playground
Use Jinja2 templating in prompts to add conditional logic, filters, and template blocks. The template format is stored in the configuration schema, and the SDK handles rendering automatically.
Programmatic Evaluation through the SDK
11/11/2025
Evaluation
Run evaluations programmatically from code with full control over test data and evaluation logic. Evaluate agents built with any framework and view results in the Agenta dashboard.
Online Evaluation
11/11/2025
Evaluation
Automatically evaluate every request to your LLM application in production. Catch hallucinations and off-brand responses as they happen instead of discovering them through user complaints.
Customize LLM-as-a-Judge Output Schemas
11/10/2025
Evaluation
Configure LLM-as-a-Judge evaluators with custom output schemas. Use binary, multiclass, or custom JSON formats. Enable reasoning for better evaluation quality.

In progress

Planned

Feature Requests

Upvote or comment on the features you care about or request a new feature.