From 68d49af6f80aed9f1d0d06dcb4880d5740cb6c5c Mon Sep 17 00:00:00 2001 From: NoyaOffer Date: Thu, 9 Oct 2025 16:34:48 +0300 Subject: [PATCH 1/3] Migrate from syntax to Mintlify import syntax - Convert all usage to ES6 import statements - Update 124 files across the documentation - Use PascalCase component names derived from filenames - Maintain absolute paths starting with /snippets/ - Preserve all existing content and functionality This migration aligns with Mintlify's recommended snippet syntax for better developer experience and standard React/JSX conventions. --- docs/_snippets/add-connection-profile.mdx | 4 +- docs/_snippets/cli/bigquery-cli.mdx | 4 +- docs/_snippets/cloud/integrations/athena.mdx | 4 +- .../_snippets/cloud/integrations/bigquery.mdx | 10 +++-- .../cards-groups/cloud-integrations-cards.mdx | 19 ++++++--- .../cloud/integrations/clickhouse.mdx | 7 +++- .../create-user-operation-snowflake.mdx | 4 +- .../integrations/create-user-operation.mdx | 4 +- .../cloud/integrations/databricks.mdx | 10 +++-- .../_snippets/cloud/integrations/postgres.mdx | 7 +++- .../_snippets/cloud/integrations/redshift.mdx | 7 +++- .../cloud/integrations/snowflake.mdx | 4 +- docs/_snippets/cloud/introduction.mdx | 4 +- .../dwh/bigquery/cli_service_account.mdx | 10 +++-- .../dwh/bigquery/cloud_service_account.mdx | 13 +++++-- docs/_snippets/faq/question-schema.mdx | 4 +- .../guides/alerts-code-configuration.mdx | 13 +++++-- .../quickstart/quickstart-elementary-prod.mdx | 4 +- docs/cloud/features.mdx | 6 ++- .../alerts-and-incidents/alert-rules.mdx | 6 ++- .../alerts-and-incidents-overview.mdx | 9 ++++- .../incident-management.mdx | 6 ++- .../alerts-and-incidents/incidents.mdx | 6 ++- .../owners-and-subscribers.mdx | 9 ++++- .../anomaly-detection/automated-freshness.mdx | 9 ++++- .../anomaly-detection/automated-monitors.mdx | 12 ++++-- .../anomaly-detection/automated-volume.mdx | 9 ++++- .../monitors-configuration.mdx | 12 ++++-- .../anomaly-detection/monitors-overview.mdx | 12 ++++-- docs/cloud/features/ci.mdx | 6 ++- .../catalog.mdx | 6 ++- .../data-health.mdx | 12 ++++-- .../data-quality-dimensions.mdx | 9 ++++- .../data-lineage/column-level-lineage.mdx | 6 ++- .../data-lineage/exposures-lineage.mdx | 9 ++++- docs/cloud/features/data-lineage/lineage.mdx | 6 ++- .../data-tests/data-tests-overview.mdx | 6 ++- docs/cloud/features/data-tests/dbt-tests.mdx | 12 ++++-- docs/cloud/features/integrations.mdx | 6 ++- docs/cloud/features/multi-env.mdx | 6 ++- docs/cloud/general/security-and-privacy.mdx | 6 ++- docs/cloud/guides/alerts-configuration.mdx | 6 ++- docs/cloud/guides/collect-job-data.mdx | 6 ++- .../cloud/guides/collect-source-freshness.mdx | 6 ++- docs/cloud/guides/enable-slack-alerts.mdx | 9 ++++- .../cloud/integrations/bi/connect-bi-tool.mdx | 6 ++- .../integrations/code-repo/bitbucket.mdx | 6 ++- .../code-repo/connect-code-repo.mdx | 6 ++- docs/cloud/integrations/code-repo/github.mdx | 6 ++- docs/cloud/integrations/code-repo/gitlab.mdx | 6 ++- docs/cloud/integrations/dwh/athena.mdx | 9 ++++- docs/cloud/integrations/dwh/bigquery.mdx | 9 ++++- docs/cloud/integrations/dwh/clickhouse.mdx | 9 ++++- docs/cloud/integrations/dwh/databricks.mdx | 9 ++++- docs/cloud/integrations/dwh/dremio.mdx | 6 ++- docs/cloud/integrations/dwh/postgres.mdx | 12 ++++-- docs/cloud/integrations/dwh/redshift.mdx | 9 ++++- docs/cloud/integrations/dwh/snowflake.mdx | 9 ++++- .../integrations/elementary-integrations.mdx | 6 ++- docs/cloud/main_introduction.mdx | 9 ++++- .../onboarding/connect-data-warehouse.mdx | 6 ++- .../onboarding/quickstart-dbt-package.mdx | 6 ++- docs/cloud/quickstart.mdx | 6 ++- .../anomaly-direction.mdx | 6 ++- .../anomaly-exclude-metrics.mdx | 6 ++- .../anomaly-params.mdx | 6 ++- .../anomaly-sensitivity.mdx | 6 ++- .../column-anomalies.mdx | 9 ++++- .../detection-delay.mdx | 6 ++- .../detection-period.mdx | 6 ++- .../dimensions.mdx | 6 ++- .../event_timestamp_column.mdx | 6 ++- .../exclude-final-results.mdx | 6 ++- .../exclude_prefix.mdx | 6 ++- .../exclude_regexp.mdx | 6 ++- .../fail_on_zero.mdx | 6 ++- .../ignore_small_changes.mdx | 6 ++- .../seasonality.mdx | 6 ++- .../time-bucket.mdx | 6 ++- .../timestamp-column.mdx | 6 ++- .../training-period.mdx | 6 ++- .../update_timestamp_column.mdx | 6 ++- .../where-expression.mdx | 6 ++- .../Anomaly-troubleshooting-guide.mdx | 6 ++- .../all-columns-anomalies.mdx | 9 ++++- .../column-anomalies.mdx | 9 ++++- .../dimension-anomalies.mdx | 6 ++- .../event-freshness-anomalies.mdx | 6 ++- .../freshness-anomalies.mdx | 6 ++- .../volume-anomalies.mdx | 6 ++- docs/data-tests/dbt/dbt-package.mdx | 6 ++- docs/data-tests/dbt/quickstart-package.mdx | 6 ++- docs/data-tests/introduction.mdx | 6 ++- docs/oss/cli-install.mdx | 15 +++++-- .../elementary-in-production.mdx | 6 ++- .../deployment-and-configuration/slack.mdx | 6 ++- .../deployment-and-configuration/teams.mdx | 6 ++- docs/oss/general/community-and-support.mdx | 6 ++- docs/oss/general/faq.mdx | 39 +++++++++++++------ docs/oss/general/troubleshooting.mdx | 6 ++- .../guides/alerts/alerts-configuration.mdx | 6 ++- docs/oss/guides/alerts/elementary-alerts.mdx | 6 ++- docs/oss/guides/alerts/send-slack-alerts.mdx | 6 ++- docs/oss/guides/alerts/send-teams-alerts.mdx | 6 ++- .../guides/collect-dbt-source-freshness.mdx | 6 ++- docs/oss/guides/collect-job-data.mdx | 6 ++- docs/oss/guides/generate-report-ui.mdx | 15 +++++-- docs/oss/integrations/bigquery.mdx | 9 ++++- docs/oss/integrations/clickhouse.mdx | 6 ++- docs/oss/integrations/databricks.mdx | 9 ++++- docs/oss/integrations/dbt.mdx | 9 ++++- docs/oss/integrations/dremio.mdx | 9 ++++- docs/oss/integrations/postgres.mdx | 9 ++++- docs/oss/integrations/redshift.mdx | 9 ++++- docs/oss/integrations/snowflake.mdx | 9 ++++- docs/oss/oss-introduction.mdx | 6 ++- docs/oss/quickstart/quickstart-alerts.mdx | 6 ++- .../oss/quickstart/quickstart-cli-package.mdx | 6 ++- docs/oss/quickstart/quickstart-cli.mdx | 18 ++++++--- docs/oss/quickstart/quickstart-prod.mdx | 6 ++- docs/oss/quickstart/quickstart-report.mdx | 6 ++- docs/overview/elementary-cloud.mdx | 9 ++++- docs/tutorial/installing-elementary.mdx | 6 ++- docs/tutorial/running-elementary.mdx | 12 ++++-- 124 files changed, 744 insertions(+), 204 deletions(-) diff --git a/docs/_snippets/add-connection-profile.mdx b/docs/_snippets/add-connection-profile.mdx index 9f7a24289..964050ace 100644 --- a/docs/_snippets/add-connection-profile.mdx +++ b/docs/_snippets/add-connection-profile.mdx @@ -1,3 +1,5 @@ +import AllProfiles from '/snippets/profiles/all-profiles.mdx'; + ## Configuring the Elementary Profile In order to connect, Elementary needs a [connection profile](https://docs.getdbt.com/dbt-cli/configure-your-profile) in a file named `profiles.yml`. @@ -40,4 +42,4 @@ Here is a demonstration: - Profile name: `elementary` - Schema name: The schema of elementary models, default is `_elementary` - + diff --git a/docs/_snippets/cli/bigquery-cli.mdx b/docs/_snippets/cli/bigquery-cli.mdx index da1f0152b..9771ed96f 100644 --- a/docs/_snippets/cli/bigquery-cli.mdx +++ b/docs/_snippets/cli/bigquery-cli.mdx @@ -1,3 +1,5 @@ +import CliServiceAccount from '/snippets/dwh/bigquery/cli_service_account.mdx'; + ### BigQuery connection profile After installing Elementary's dbt package upon running `dbt deps`, @@ -38,7 +40,7 @@ elementary: We support the same format and connection methods as dbt. Please refer to dbt's documentation of [BigQuery](https://docs.getdbt.com/reference/warehouse-setups/bigquery-setup) for further details. - + Add the full path of this JSON file to your connection profile under 'keyfile'. diff --git a/docs/_snippets/cloud/integrations/athena.mdx b/docs/_snippets/cloud/integrations/athena.mdx index aaa7e3460..8bbf98986 100644 --- a/docs/_snippets/cloud/integrations/athena.mdx +++ b/docs/_snippets/cloud/integrations/athena.mdx @@ -1,6 +1,8 @@ +import CreateUserOperation from '/snippets/cloud/integrations/create-user-operation.mdx'; + You will connect Elementary Cloud to Athena for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + ## AWS Setup diff --git a/docs/_snippets/cloud/integrations/bigquery.mdx b/docs/_snippets/cloud/integrations/bigquery.mdx index 8e35e235e..6e56069c5 100644 --- a/docs/_snippets/cloud/integrations/bigquery.mdx +++ b/docs/_snippets/cloud/integrations/bigquery.mdx @@ -1,8 +1,12 @@ +import CloudServiceAccount from '/snippets/dwh/bigquery/cloud_service_account.mdx'; +import PermissionsAndSecurity from '/snippets/cloud/integrations/permissions-and-security.mdx'; +import IpAllowlist from '/snippets/cloud/integrations/ip-allowlist.mdx'; + You will connect Elementary Cloud to Bigquery for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + - + ### Fill the connection form @@ -13,4 +17,4 @@ Provide the following fields: - **Elementary dataset**: The name of your Elementary dataset. Usually `[dataset name]_elementary`. - **Location**: Use this field to configure the location of BigQuery datasets as per [the BigQuery documentation](https://cloud.google.com/bigquery/docs/locations). - + diff --git a/docs/_snippets/cloud/integrations/cards-groups/cloud-integrations-cards.mdx b/docs/_snippets/cloud/integrations/cards-groups/cloud-integrations-cards.mdx index 121020434..f28a58ebc 100644 --- a/docs/_snippets/cloud/integrations/cards-groups/cloud-integrations-cards.mdx +++ b/docs/_snippets/cloud/integrations/cards-groups/cloud-integrations-cards.mdx @@ -1,23 +1,30 @@ +import ConnectDwhCards from '/snippets/cloud/integrations/cards-groups/connect-dwh-cards.mdx'; +import TransformationAndOrchestrationCards from '/snippets/cloud/integrations/cards-groups/transformation-and-orchestration-cards.mdx'; +import BiCards from '/snippets/cloud/integrations/cards-groups/bi-cards.mdx'; +import ReverseEtlCards from '/snippets/cloud/integrations/cards-groups/reverse-etl-cards.mdx'; +import CodeRepoCards from '/snippets/cloud/integrations/cards-groups/code-repo-cards.mdx'; +import AlertsDestinationCards from '/snippets/cloud/integrations/cards-groups/alerts-destination-cards.mdx'; + ### Data warehouses - + ### Transformation and orchestration - + ### Data visualization - + ### Reverse ETL - + ### Code repositories - + ### Alerts & incidents - \ No newline at end of file + \ No newline at end of file diff --git a/docs/_snippets/cloud/integrations/clickhouse.mdx b/docs/_snippets/cloud/integrations/clickhouse.mdx index c9dfdf50a..21013479c 100644 --- a/docs/_snippets/cloud/integrations/clickhouse.mdx +++ b/docs/_snippets/cloud/integrations/clickhouse.mdx @@ -1,6 +1,9 @@ +import CreateUserOperation from '/snippets/cloud/integrations/create-user-operation.mdx'; +import IpAllowlist from '/snippets/cloud/integrations/ip-allowlist.mdx'; + You will connect Elementary Cloud to Clickhouse for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + ### Fill the connection form @@ -12,7 +15,7 @@ Provide the following fields: - **User**: The name of the for Elementary user. - **Password**: The password associated with the provided user. - + ### Connect via SSH tunnel diff --git a/docs/_snippets/cloud/integrations/create-user-operation-snowflake.mdx b/docs/_snippets/cloud/integrations/create-user-operation-snowflake.mdx index 01aac8f02..2bf22da30 100644 --- a/docs/_snippets/cloud/integrations/create-user-operation-snowflake.mdx +++ b/docs/_snippets/cloud/integrations/create-user-operation-snowflake.mdx @@ -1,3 +1,5 @@ +import PermissionsAndSecurity from '/snippets/cloud/integrations/permissions-and-security.mdx'; + ### Create a user for Elementary cloud * Please create a Snowflake key-pair (private and public key) using [this](https://docs.snowflake.com/en/user-guide/key-pair-auth#configuring-key-pair-authentication) guide. @@ -14,4 +16,4 @@ dbt run-operation create_elementary_user --args "{'public_key': '$SNOWFLAKE_PUBL This command will generate a query to create a user with the necessary permissions. Run this query on your data warehouse with **admin permissions** to create the user. - + diff --git a/docs/_snippets/cloud/integrations/create-user-operation.mdx b/docs/_snippets/cloud/integrations/create-user-operation.mdx index bbc685a73..1d60ba049 100644 --- a/docs/_snippets/cloud/integrations/create-user-operation.mdx +++ b/docs/_snippets/cloud/integrations/create-user-operation.mdx @@ -1,3 +1,5 @@ +import PermissionsAndSecurity from '/snippets/cloud/integrations/permissions-and-security.mdx'; + ### Create a user for Elementary cloud On your dbt project, run: @@ -9,4 +11,4 @@ dbt run-operation create_elementary_user This command will generate a query to create a user with the necessary permissions. Run this query on your data warehouse with **admin permissions** to create the user. - + diff --git a/docs/_snippets/cloud/integrations/databricks.mdx b/docs/_snippets/cloud/integrations/databricks.mdx index 04dd98f97..10f447a3c 100644 --- a/docs/_snippets/cloud/integrations/databricks.mdx +++ b/docs/_snippets/cloud/integrations/databricks.mdx @@ -1,8 +1,12 @@ +import CreateServicePrincipal from '/snippets/dwh/databricks/create_service_principal.mdx'; +import PermissionsAndSecurity from '/snippets/cloud/integrations/permissions-and-security.mdx'; +import IpAllowlist from '/snippets/cloud/integrations/ip-allowlist.mdx'; + You will connect Elementary Cloud to Databricks for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + - + ### Fill the connection form @@ -14,4 +18,4 @@ Provide the following fields: - **Catalog (optional)**: The name of the Databricks Catalog. - **Elementary schema**: The name of your Elementary schema. Usually `[schema name]_elementary`. - + diff --git a/docs/_snippets/cloud/integrations/postgres.mdx b/docs/_snippets/cloud/integrations/postgres.mdx index 9e7b7fd5f..741ad0b67 100644 --- a/docs/_snippets/cloud/integrations/postgres.mdx +++ b/docs/_snippets/cloud/integrations/postgres.mdx @@ -1,6 +1,9 @@ +import CreateUserOperation from '/snippets/cloud/integrations/create-user-operation.mdx'; +import IpAllowlist from '/snippets/cloud/integrations/ip-allowlist.mdx'; + You will connect Elementary Cloud to Postgres for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + ### Fill the connection form @@ -13,7 +16,7 @@ Provide the following fields: - **User**: The name of the for Elementary user. - **Password**: The password associated with the provided user. - + ### Connect via SSH tunnel diff --git a/docs/_snippets/cloud/integrations/redshift.mdx b/docs/_snippets/cloud/integrations/redshift.mdx index 3604ce7a0..e220cbc7a 100644 --- a/docs/_snippets/cloud/integrations/redshift.mdx +++ b/docs/_snippets/cloud/integrations/redshift.mdx @@ -1,6 +1,9 @@ +import CreateUserOperation from '/snippets/cloud/integrations/create-user-operation.mdx'; +import IpAllowlist from '/snippets/cloud/integrations/ip-allowlist.mdx'; + You will connect Elementary Cloud to Redshift for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + ### Fill the connection form @@ -13,7 +16,7 @@ Provide the following fields: - **User**: The name of the for Elementary user. - **Password**: The password associated with the provided user. - + ### Connect via SSH tunnel diff --git a/docs/_snippets/cloud/integrations/snowflake.mdx b/docs/_snippets/cloud/integrations/snowflake.mdx index 6c2dea745..9248a2ef3 100644 --- a/docs/_snippets/cloud/integrations/snowflake.mdx +++ b/docs/_snippets/cloud/integrations/snowflake.mdx @@ -1,6 +1,8 @@ +import CreateUserOperationSnowflake from '/snippets/cloud/integrations/create-user-operation-snowflake.mdx'; + You will connect Elementary Cloud to Snowflake for syncing the Elementary schema (created by the [Elementary dbt package](/cloud/onboarding/quickstart-dbt-package)). - + ### Fill the connection form diff --git a/docs/_snippets/cloud/introduction.mdx b/docs/_snippets/cloud/introduction.mdx index dc6c95d3d..ebc05faaf 100644 --- a/docs/_snippets/cloud/introduction.mdx +++ b/docs/_snippets/cloud/introduction.mdx @@ -1,3 +1,5 @@ +import CloudIntegrationsCards from '/snippets/cloud/integrations/cards-groups/cloud-integrations-cards.mdx'; + ### Why choose Elementary Cloud? @@ -40,7 +42,7 @@ ### Elementary Cloud integrations - + ### How does it work? diff --git a/docs/_snippets/dwh/bigquery/cli_service_account.mdx b/docs/_snippets/dwh/bigquery/cli_service_account.mdx index 2d66aba18..8b634e0d4 100644 --- a/docs/_snippets/dwh/bigquery/cli_service_account.mdx +++ b/docs/_snippets/dwh/bigquery/cli_service_account.mdx @@ -1,3 +1,7 @@ - - - +import CreateServiceAccount from '/snippets/dwh/bigquery/create_service_account.mdx'; +import CliPermissions from '/snippets/dwh/bigquery/cli_permissions.mdx'; +import CreateKey from '/snippets/dwh/bigquery/create_key.mdx'; + + + + diff --git a/docs/_snippets/dwh/bigquery/cloud_service_account.mdx b/docs/_snippets/dwh/bigquery/cloud_service_account.mdx index 703186db4..23ac51d6d 100644 --- a/docs/_snippets/dwh/bigquery/cloud_service_account.mdx +++ b/docs/_snippets/dwh/bigquery/cloud_service_account.mdx @@ -1,4 +1,9 @@ - - - - +import CreateServiceAccount from '/snippets/dwh/bigquery/create_service_account.mdx'; +import CloudPermissions from '/snippets/dwh/bigquery/cloud_permissions.mdx'; +import CreateKey from '/snippets/dwh/bigquery/create_key.mdx'; +import GrantUserAccessOnDatasetLevel from '/snippets/dwh/bigquery/grant_user_access_on_dataset_level.mdx'; + + + + + diff --git a/docs/_snippets/faq/question-schema.mdx b/docs/_snippets/faq/question-schema.mdx index 5a80fc069..8e4533c8c 100644 --- a/docs/_snippets/faq/question-schema.mdx +++ b/docs/_snippets/faq/question-schema.mdx @@ -1,5 +1,7 @@ +import QuestionSchemaNoAccordion from '/snippets/faq/question-schema-no-accordion.mdx'; + - + diff --git a/docs/_snippets/guides/alerts-code-configuration.mdx b/docs/_snippets/guides/alerts-code-configuration.mdx index aab2625ab..339e73634 100644 --- a/docs/_snippets/guides/alerts-code-configuration.mdx +++ b/docs/_snippets/guides/alerts-code-configuration.mdx @@ -1,3 +1,8 @@ +import Owner from '/snippets/alerts/owner.mdx'; +import Subscribers from '/snippets/alerts/subscribers.mdx'; +import Description from '/snippets/alerts/description.mdx'; +import Tags from '/snippets/alerts/tags.mdx'; + You can enrich your alerts by adding properties to tests, models and sources in your `.yml` files. The supported attributes are: [owner](#Owner), [subscribers](#Subscribers), @@ -45,19 +50,19 @@ Elementary prioritizes configuration in the following order: #### Owner - + #### Subscribers - + #### Test description - + #### Tags - + ### Alerts distribution diff --git a/docs/_snippets/quickstart/quickstart-elementary-prod.mdx b/docs/_snippets/quickstart/quickstart-elementary-prod.mdx index 82ddb206c..682357f85 100644 --- a/docs/_snippets/quickstart/quickstart-elementary-prod.mdx +++ b/docs/_snippets/quickstart/quickstart-elementary-prod.mdx @@ -1,3 +1,5 @@ +import QuestionSchema from '/snippets/faq/question-schema.mdx'; + @@ -48,7 +50,7 @@ On your dbt project, make sure that Elementary dbt package can: - Write to the elementary schema - Create a schema (alternatively, you can create the elementary schema in advance) - + ## When to run Elementary? diff --git a/docs/cloud/features.mdx b/docs/cloud/features.mdx index 4a27276bb..ae375317f 100644 --- a/docs/cloud/features.mdx +++ b/docs/cloud/features.mdx @@ -3,4 +3,8 @@ title: "Platform features" icon: "browsers" --- - \ No newline at end of file +import Features from '/snippets/cloud/features.mdx'; + + + + \ No newline at end of file diff --git a/docs/cloud/features/alerts-and-incidents/alert-rules.mdx b/docs/cloud/features/alerts-and-incidents/alert-rules.mdx index 6ecd26666..7e606423e 100644 --- a/docs/cloud/features/alerts-and-incidents/alert-rules.mdx +++ b/docs/cloud/features/alerts-and-incidents/alert-rules.mdx @@ -2,7 +2,11 @@ title: "Alert rules" --- - +import CloudFeatureTag from '/_snippets/cloud/cloud-feature-tag.mdx'; + + + + **Alert Rules** help you control where alerts are sent and when they are triggered, so you can stay focused on the incidents that matter most. In Elementary Cloud, each rule combines **filters** (what to alert on) and **destinations** (where to send the alert). diff --git a/docs/cloud/features/alerts-and-incidents/alerts-and-incidents-overview.mdx b/docs/cloud/features/alerts-and-incidents/alerts-and-incidents-overview.mdx index 41e3eec25..5bfbc9b60 100644 --- a/docs/cloud/features/alerts-and-incidents/alerts-and-incidents-overview.mdx +++ b/docs/cloud/features/alerts-and-incidents/alerts-and-incidents-overview.mdx @@ -3,6 +3,11 @@ title: Alerts and Incidents Overview sidebarTitle: Alerts & incidents overview --- +import AlertTypes from '/snippets/cloud/features/alerts-and-incidents/alert-types.mdx'; +import AlertsDestinationCards from '/snippets/cloud/integrations/cards-groups/alerts-destination-cards.mdx'; + + + Alerts and incidents in Elementary are designed to shorten your time to response and time to resolution when data issues occur. - **Alert -** Notification about an event that indicates a data issue. @@ -51,8 +56,8 @@ you will be able to track all open and historical incidents, and get metrics on ## Alert types - + ## Supported alert integrations - + diff --git a/docs/cloud/features/alerts-and-incidents/incident-management.mdx b/docs/cloud/features/alerts-and-incidents/incident-management.mdx index e85a80c2d..4b1b69d0d 100644 --- a/docs/cloud/features/alerts-and-incidents/incident-management.mdx +++ b/docs/cloud/features/alerts-and-incidents/incident-management.mdx @@ -3,7 +3,11 @@ title: Incident Management sidebarTitle: Incident management --- - +import CloudFeatureTag from '/snippets/cloud/cloud-feature-tag.mdx'; + + + + The `Incidents` page is designed to enable your team to stay on top of open incidents and collaborate on resolving them. The page gives a comprehensive overview of all current and previous incidents, where users can view the status, prioritize, assign and resolve incidents. diff --git a/docs/cloud/features/alerts-and-incidents/incidents.mdx b/docs/cloud/features/alerts-and-incidents/incidents.mdx index 10809c345..d11752de6 100644 --- a/docs/cloud/features/alerts-and-incidents/incidents.mdx +++ b/docs/cloud/features/alerts-and-incidents/incidents.mdx @@ -3,7 +3,11 @@ title: Incidents in Elementary sidebarTitle: Incidents --- - +import CloudFeatureTag from '/snippets/cloud/cloud-feature-tag.mdx'; + + + + One of the challenges data teams face is tracking and understand and collaborate on the status of data issues. Tests fail daily, pipelines are executed frequently, alerts are sent to different channels. diff --git a/docs/cloud/features/alerts-and-incidents/owners-and-subscribers.mdx b/docs/cloud/features/alerts-and-incidents/owners-and-subscribers.mdx index 362615eea..4486f302c 100644 --- a/docs/cloud/features/alerts-and-incidents/owners-and-subscribers.mdx +++ b/docs/cloud/features/alerts-and-incidents/owners-and-subscribers.mdx @@ -2,6 +2,11 @@ title: "Owners and subscribers" --- +import Owner from '/snippets/alerts/owner.mdx'; +import Subscribers from '/snippets/alerts/subscribers.mdx'; + + + We highly recommend configuring owners and subscribers for your models and/or tests. An owner is the person responsible for the model, and subscribers are the people who are interested in getting the alerts on the model or test. Owners and subscribers will be mentioned (tagged) in the Slack alerts. Also their name will appear in the alerts and the UI. @@ -12,8 +17,8 @@ configuring owners and subscribers: Owners can be easily added or edited in the Catalog screen. [Learn more about managing your assets' metadata.](/cloud/features/data-governance/manage-metadata) #### Configuting owners in code: - + ### Subscribers - + diff --git a/docs/cloud/features/anomaly-detection/automated-freshness.mdx b/docs/cloud/features/anomaly-detection/automated-freshness.mdx index 8b0357e0e..c86e0f77a 100644 --- a/docs/cloud/features/anomaly-detection/automated-freshness.mdx +++ b/docs/cloud/features/anomaly-detection/automated-freshness.mdx @@ -3,6 +3,11 @@ title: Automated Freshness Monitor sidebarTitle: "Automated freshness" --- +import FreshnessConfiguration from '/snippets/cloud/features/anomaly-detection/freshness-configuration.mdx'; +import AllAnomaliesConfiguration from '/snippets/cloud/features/anomaly-detection/all-anomalies-configuration.mdx'; + + + The purpose of the Freshness monitor is to alert when a data asset hasn't been updated in a period of time that exceeds the update SLA of that table. Freshness monitors are by default created for all sources in your dbt project. They can be created for additional tables upon request. @@ -59,5 +64,5 @@ Use the `Anomaly settings` and `result feedback` buttons to impact the monitor. ### Anomaly settings - - + + diff --git a/docs/cloud/features/anomaly-detection/automated-monitors.mdx b/docs/cloud/features/anomaly-detection/automated-monitors.mdx index 428bce61d..588cbcd95 100644 --- a/docs/cloud/features/anomaly-detection/automated-monitors.mdx +++ b/docs/cloud/features/anomaly-detection/automated-monitors.mdx @@ -3,9 +3,15 @@ title: Automated Freshness & Volume Monitors sidebarTitle: "Introduction" --- - +import CloudFeatureTag from '/snippets/cloud/cloud-feature-tag.mdx'; +import AutomatedMonitorsIntro from '/snippets/cloud/features/anomaly-detection/automated-monitors-intro.mdx'; +import AutomatedMonitorsCards from '/snippets/cloud/features/anomaly-detection/automated-monitors-cards.mdx'; - + + + + + Once your environment is set up, we automatically collect metadata from your warehouse, which our ML models run on. The models are operational when the initial backfill is completed, there is no "loading / training period" - Elementary will collect enough historical data after setup to train the models. @@ -36,7 +42,7 @@ You can fine tune the [configuration](/cloud/features/anomaly-detection/monitors ## Automated Monitors - + ## Alerts on Failures diff --git a/docs/cloud/features/anomaly-detection/automated-volume.mdx b/docs/cloud/features/anomaly-detection/automated-volume.mdx index 15b66da75..6e45c285a 100644 --- a/docs/cloud/features/anomaly-detection/automated-volume.mdx +++ b/docs/cloud/features/anomaly-detection/automated-volume.mdx @@ -3,6 +3,11 @@ title: Automated Volume Monitor sidebarTitle: "Automated volume" --- +import VolumeConfiguration from '/snippets/cloud/features/anomaly-detection/volume-configuration.mdx'; +import AllAnomaliesConfiguration from '/snippets/cloud/features/anomaly-detection/all-anomalies-configuration.mdx'; + + + The volume monitor tracks the **total row count** of a table over time, rather than individual table updates. This means that Elementary will not consider a single update as anomalous, but rather a continuous anomalous trend occurring over a period of time. @@ -31,8 +36,8 @@ Use the `Anomaly settings` and `result feedback` buttons to impact the monitor. ### Anonmaly settings - - + +