Monitor Metrics
Monitor Performance, Drift, Data Quality, and Custom Metrics
Overview
Monitors automatically detect drift, data quality issues, or anomalous performance degradations with highly configurable dimensions based on both common KPIs and custom metrics.
Performance
AUC, LogLoss, Mean Error, MAE, MAPE, SMAPE, WAPE, RMSE, MSE, RSquared, Accuracy, Precision, Recall, f_1, Sensitivity, Specificity, False Negative Rate, False Positive Rate
Drift
PSI, KL Divergence, JS Distance, KS Statistic
Data Quality
Percent Empty, Cardinality, New Values, Missing Values, Quantiles (P99.9, P95, P50, P99
Learn how to set up your monitors here!
Model performance metrics measure how well your model performs in production. Monitor model performance with daily or hourly checks using an evaluation metric. Your model type determines your performance metric.
Performance Metrics
Metrics are batched into Metric Groups that align with model types and their variants.
Metric Group | Metrics |
---|---|
Classification | Accuracy, Recall, Precision, FPR, FNR, F1, Sensitivity, Specificity |
Regression | MAPE, MAE, RMSE, MSE, R-Squared, Mean Error |
Ranking | NDCG@k, AUC@k |
Ranking Labels | MAP@k, MRR |
AUC / LogLoss | AUC, PR-AUC, Log Loss |
Computer Vision / Object Detection | Accuracy (MAP & IoU coming soon) |
Valid Model Type & Metric Group Combinations
Model Type | Metric Group Combination |
---|---|
| Regression |
| Classification and/or Regression and/or AUC/LogLoss |
| Ranking and/or Ranking Labels |
| Ranking and/or AUC/LogLoss |
Map performance metrics relevant to your model type within each model type page.
Metric | Metric Family |
---|---|
| |
| |
Mean Error |
|
| |
| |
| |
| |
| |
| |
rSquared |
|
| |
| |
| |
| |
| |
| |
| |
| |
|
Drift Monitors
Drift monitors measure distribution drift, which is the difference between two statistical distributions.
Arize offers various distributional drift metrics to choose from when setting up a monitor. Each metric is tailored to a specific use case; refer to this guide to help choose the appropriate metric for various ML use cases.
Drift Metrics
Metric | Data Type | Description |
---|---|---|
|
| |
Embedding Vectors | Euclidean distance check determines if the group of production data’s average centroid has moved away from the baseline group For unstructured data types, learn more here | |
|
| |
|
| |
|
|
Model health depends on high-quality data that powers model features. Data quality monitors help identify key data quality issues such as cardinality shifts, data type mismatch, missing data, and more.
Data Quality Metrics
Metric | Data Type | Description |
---|---|---|
Percent Empty |
| The percent of nulls in your model features |
Cardinality (Count Distinct) |
| The cardinality of your categorical features |
| Count of new unique values that appear in production but not in baseline Note: this monitor requires a baseline to compare against | |
| Count of new unique values that appear in baseline but not in production Note: this monitor requires a baseline to compare against | |
| p99.9, p99, p95, p50 | |
Sum |
| Sum of your numeric data over the evaluation window |
Count |
| Traffic count of predictions, features, etc. Can be used with filters |
Average |
| Average of your numeric data over the evaluation window |
Couldn't find your metric above? Arize supports the ability to monitor custom metrics using SQL. Here is an example of a custom metric for the percent of a loan that is outstanding:
Learn how to create custom metrics here.
pageCustom Metrics Query LanguageLast updated