Recommendations
Overview of the Recommender Systems Preset
This Preset is coming soon to the new Evidently API! Check the old docs for now.
Report. To run a Preset on your data for a single current dataset for top-k recommendations:
Test Suite. To add pass/fail ranking quality Tests, auto-generated from the ref
dataset:
Overview
RecsysPreset
evaluates the quality of the recommender system by generating multiple metrics to assess the quality of ranking and diversity of recommendations. You must provide the k
parameter to evaluate the Top-K recommendations.
It includes 10+ metrics like NDCG at K, MAP at K, HitRate, diversity, serendipity, etc. Metric selection depends on the provided data since some of the Metrics require additional dataset (training data) or item / user features.
Metric explainers. Check the Ranking and RecSys Metrics to see how each Metric works.
Test Suite. If you enable Tests, this will automatically run checks to assess if the model performance metrics are within bounds.
Tests are auto-generated based on reference dataset. If the reference dataset is provided, conditions like expected ranking accuracy will be derived from it.
How Tests work. Read about Tests and check defaults for each Test in the reference table.
Use case
These Presets are useful in various scenarios:
-
Experimental evaluations as you iterate on building your recommender system.
-
Side-by-side comparison for two different models or periods.
-
Production monitoring checks after you acquire ground truth labels.
-
Debugging. If you notice a drop in performance, use the visual Report to understand changes.
Data requirements
-
Prediction. Recommended items with rank or score.
-
Target. True relevance score or interaction result.
-
(Optional) Input/user features. For some diversity metrics.
-
(Optional) Training data. For some diversity metrics.
-
(Optional) Reference dataset. To get a side-by-side comparison or auto-generate test conditions.
Data schema mapping. Use the data definition to map your input data.
Report Customization
You can customize the Report in several ways:
-
Change Test conditions. To modify the auto-generated conditions, you can set yours: either a different condition relative to the reference or any custom conditions.
-
Modify Report composition. You can add additional metrics, such as compute Data Drift for user or item feature distributions, or to evaluate prediction drift.