Managed Datasets¶
Experimental Feature
Managed datasets are an experimental feature currently gated behind a feature flag. Reach out to us on Slack or contact us to learn how to enable it for your project.
Managed datasets let you build and maintain collections of test cases for evaluating your AI systems. You can create datasets through the Logfire web UI or programmatically via the SDK, then export them as pydantic-evals Dataset objects to run evaluations.
Relationship with Pydantic Evals
Managed datasets are the server-side complement to pydantic-evals file-based datasets. While pydantic-evals stores datasets as local YAML files, managed datasets are stored in Logfire and can be created from production traces, edited collaboratively in the UI, and exported for evaluation. The SDK is designed so you can move seamlessly between the two.
Why Managed Datasets?¶
When evaluating AI systems, you need test cases that reflect real-world usage. Managed datasets solve several problems:
- Production-driven test cases: Create test cases directly from production traces using the "Add to Dataset" action in Live View, so your evaluations reflect real user inputs.
- Collaborative editing: Team members can add, review, and refine test cases through the web UI without touching code.
- Typed schemas: Define input, output, and metadata schemas so that every case is validated, preventing malformed test data.
- SDK integration: Export datasets as typed
pydantic_evals.Datasetobjects and run evaluations with a single method call.
The Evaluation Workflow¶
Managed datasets fit into a continuous evaluation loop:
- Observe production behavior in Live View.
- Curate test cases by adding interesting traces to a managed dataset.
- Evaluate your system against the dataset using pydantic-evals.
- Analyze the results in the Logfire Evals tab.
- Improve your system and repeat.
Next Steps¶
- Web UI Guide --- Create and manage datasets through the Logfire interface.
- SDK Guide --- Manage datasets programmatically with Python.
- Running Evaluations --- Export datasets and run evaluations with pydantic-evals.
- SDK Reference --- Complete method and exception reference.