Overview
End-to-end evaluation workflow.
This page shows the core eval workflow with the Evidently library and links to guides.
Define and run the eval
To log the evaluation results to the Evidently Platform, first connect to Evidently Cloud or your local workspace and create a Project. It’s optional: you can also run evals locally.
Prepare the input data
Get your data in a table like a pandas.DataFrame
. More on data requirements. You can also load data from Evidently Platform, like tracing or synthetic datasets.
Create a Dataset object
Create a Dataset object with DataDefinition()
that specifies column role and types. You can also use default type detection. How to set Data Definition.
(Optional) Add descriptors
For text evals, choose and compute row-level descriptors
. How to use Descriptors.
Configure Report
For dataset-level evals (classification, data drift) or to summarize descriptors, create a Report
with chosen metrics
or presets
. How to configure Reports.
(Optional) Add Test conditions
Add Pass/Fail conditions, like to check if text length in < 100 symbols. How to configure Tests.
(Optional) Add Tags and Timestamps
Add tags
or metadata
to identify specific evaluation runs or datasets, or override the default timestamp
. How to add metadata.
Run the Report
To execute the eval, run
the Report on the Dataset
(or two).
Explore the results
Quickstarts
Check for end-to-end examples: