Alerts#

Alerts are a mechanism for informing you about possible problem situations.

In this section

System configuration#

These variables control the basic alert behavior:

  • alerts.mode — Enables/disables the feature. Enabled by default.

  • alerts.max_allowed — Maximum number of alerts allowed to be configured, by default 10000. Any new alerts above this limit return an error.

  • alerts.max_criteria_count — Maximum number of events. By default, 100.

These values can be modified by the support team.

SDK#

The SDK supports these alert operations:

Predefined events (EventKind)#

The predefined event types are:

  • data-drift-detected — A detected change in model input data that potentially leads to model performance degradation.

  • data-drift-suspected — A suspected change in model input data that potentially leads to model performance degradation.

  • concept-drift-detected — A detected change, over time, of statistical properties of the target variable (what the model is predicting).

  • concept-drift-suspected — A suspected change, over time, of statistical properties of the target variable (what the model is predicting).

  • model-performance-detected — A detected change of the overall model performance and/or feature-level performance.

  • model-performance-suspected — A suspected change of the overall model performance and/or feature-level performance.

  • model-serving-performance-detected — A detected change in how much time the prediction takes (i.e. the latency, measured in time units).

  • model-serving-performance-suspected — A suspected change in how much time the prediction takes (i.e. the latency, measured in time units).

  • mm-app-anomaly-detected — An alert based on user-defined metrics/results.

  • mm-app-anomaly-suspected — An alert based on user-defined metrics/results.

  • failed — The job failed.

See Model monitoring for more details on drift and performance.

Creating an alert#

When creating an alert you can select an event type for a specific model, for example data_drift_suspected or any of the predefined events above. You can optionally specify the frequency of the alert through the criteria field in the configuration (how many times in what time window, etc.). If not specified, it uses the default. See all of the alert configuration parameters. You can configure Git, Slack, and webhook notifications for the alert. For alerts on model endpoints, see Creating a model monitoring alert.

This example illustrates creating an alert with a Slack notification for a job failure with defined criteria. This alert gets triggered if the job fails 3 times in a 10 minute period.

notification = mlrun.model.Notification(
    kind="slack",
    name="slack_notification",
    secret_params={
        "webhook": "https://hooks.slack.com/",
    },
).to_dict()

notifications = [alert_objects.AlertNotification(notification=notification)]
alert_name = "failure-alert"
alert_summary = "Running a job has failed"
entity_kind = alert_objects.EventEntityKind.JOB
event_name = alert_objects.EventKind.FAILED

# The job's run id that will be tracked
run_id = "run-id"

alert_data = mlrun.alerts.alert.AlertConfig(
    project=project_name,
    name=alert_name,
    summary=alert_summary,
    severity=alert_objects.AlertSeverity.HIGH,
    entities=alert_objects.EventEntities(
        kind=entity_kind, project=project_name, ids=[run_id]
    ),
    trigger=alert_objects.AlertTrigger(events=[event_name]),
    criteria=alert_objects.AlertCriteria(period="10m", count=3),
    notifications=notifications,
)

# Save (and activate) the alert config:
project.store_alert_config(alert_data)

Creating a model monitoring alert#

Model monitoring alerts notify you when measured input data and/or statistic/result produce unexpected results, the same as other alerts. The difference is that the configuration of a model monitoring alert is based on specific model endpoints and optionally result names, including wildcards. See the full parameter details in create_model_monitoring_alert_configs(). (You could also use mlrun.alerts.alert.AlertConfig to configure ModelEndpoint alerts, but create_model_monitoring_alert_configs is much easier to configure).

Important

Create model monitoring alerts after your serving function is deployed. When using a wildcard or when not specifying exact name of app+result (for example when not specifying results at all), the apps in question need to already be running and generating some metrics, so that the get_model_endpoint_monitoring_metrics API call is able to extract the details for the specific ModelEndpoint.

This example illustrates creating a model monitoring alert to detect data drift, with a webhook notification for the alert.

alert_configs = myproject.create_model_monitoring_alert_configs(
    # Name of the AlertConfig template
    name="alert-name",
    summary="user_template_summary_EventKind.DATA_DRIFT_DETECTED",
    # Retrieve metrics from these endpoints to configure the alert
    endpoints=myproject.list_model_endpoints(),
    # AlertTrigger event type
    events=[EventKind.DATA_DRIFT_DETECTED],
    notifications=[notifications],
    result_names=[],  # Can use wildcards
    severity=alert_constants.AlertSeverity.LOW,
    criteria=None,
    reset_policy=mlrun.common.schemas.alert.ResetPolicy.MANUAL,
)
for alert_config in alert_configs:
    myproject.store_alert_config(alert_config)

Modifying an alert#

When you run store_alert_config on an existing alert:

  • The alert is reset if you modify a field that affects the conditions that trigger the alert. These fields are:

    • Entity

    • Trigger

    • Criteria

  • The alert is not reset if you modify a field that affects the notifications that are sent or the result of the alert being activated. These fields are:

    • Description

    • Summary

    • Severity

    • Notifications

You can use the force_reset option when running store_alert_config to force a reset for fields that, by default, do not reset the alert. By default, force_reset is set to false.

Alert reset policy#

The mlrun.common.schemas.alert.ResetPolicy specifies when to clear the alert and change the alert's status from active to inactive. When an alert becomes inactive, its notifications cease. When it is re-activated, notifications are renewed. The ResetPolicy options are:

  • manual — for manual reset of the alert

  • auto — if the criteria contains a time period such that the alert is reset once there are no more invocations in the relevant time window.

Note: If an alert is in an active state and its reset-policy is changed from manual to auto, the alert is immediately reset. This ensures that the behavior aligns with the auto-reset behavior.

Alert templates#

Alert templates simplify the creation of alerts by providing a predefined set of configurations. The system comes with several predefined templates that can be used with MLRun applications. If you use non-MLRun applications (for example, with model monitoring), you must configure an application-specific alert. The templates are cross-project objects. When generating an alert, you must assign the project to it. See the alert template parameters.

Creating an alert with a template#

The system has a few pre-defined templates: JobFailed, DataDriftDetected, DataDriftSuspected. When using a pre-defined template, you only need to supply:

  • name: str

  • project: str

  • entity: EventEntity

  • NotificationKind: a list of at least one notification

summary, severity, trigger, and reset policy, are pre-configured in the template.
You can customize one or more of these fields when creating an alert from a template.

See the AlertTemplate parameters.

This example illustrates a Slack notification for a job failure alert, using the predefined system template JobFailed:

job_fail_template = project.get_alert_template("JobFailed")
alert_from_template = mlrun.alerts.alert.AlertConfig(
    project=project_name,
    name="failure",
    template=job_fail_template,
)
entities = alert_objects.EventEntities(
    kind=alert_objects.EventEntityKind.JOB,
    project=project_name,
    ids=[run_id],
)
alert_from_template.with_entities(entities=entities)

notification = mlrun.model.Notification(
    kind="slack",
    name="slack_notification",
    secret_params={
        "webhook": "https://hooks.slack.com/",
    },
).to_dict()

notifications = [alert_objects.AlertNotification(notification=notification)]

alert_from_template.with_notifications(notifications=notifications)

project.store_alert_config(alert_from_template)

Alert activations#

When an existing alert is activated by its configured trigger, MLRun saves the activation record. There is a way to get all activations of alerts - see alert activations tutorial

Alert activation records are stored in a partitioned table. The table is partitioned weekly and supports retention. The default retention period is 14 weeks, but it can be adjusted by setting object_retentions.alert_activations in the MLRun configuration. The value should be specified in days.