You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/concepts/models/sql_models.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -281,7 +281,7 @@ JOIN countries
281
281
282
282
SQLMesh will detect that the model depends on both `employees` and `countries`. When executing this model, it will ensure that `employees` and `countries` are executed first.
283
283
284
-
External dependencies not defined in SQLMesh are also supported. SQLMesh can either depend on them implicitly through the order in which they are executed, or through signals if you are using [Airflow](../../integrations/airflow.md).
284
+
External dependencies not defined in SQLMesh are also supported. SQLMesh can either depend on them implicitly through the order in which they are executed, or through [signals](../../guides/signals.md).
285
285
286
286
Although automatic dependency detection works most of the time, there may be specific cases for which you want to define dependencies manually. You can do so in the `MODEL` DDL with the [dependencies property](./overview.md#properties).
Copy file name to clipboardExpand all lines: docs/concepts/overview.md
+1-3Lines changed: 1 addition & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -68,6 +68,4 @@ SQLMesh automatically runs audits when you apply a `plan` to an environment, or
68
68
## Infrastructure and orchestration
69
69
Every company's data infrastructure is different. SQLMesh is flexible with regard to which engines and orchestration frameworks you use — its only requirement is access to the target SQL/analytics engine.
70
70
71
-
SQLMesh keeps track of model versions and processed data intervals using your existing infrastructure. If SQLMesh is configured without an external orchestrator (such as Airflow), it automatically creates a `sqlmesh` schema in your data warehouse for its internal metadata.
72
-
73
-
If SQLMesh is configured with Airflow, then it will store all its metadata in the Airflow database. Read more about how [SQLMesh integrates with Airflow](../integrations/airflow.md).
71
+
SQLMesh keeps track of model versions and processed data intervals using your existing infrastructure. SQLMesh it automatically creates a `sqlmesh` schema in your data warehouse for its internal metadata.
Copy file name to clipboardExpand all lines: docs/faq/faq.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -167,14 +167,17 @@
167
167
## Scheduling
168
168
169
169
??? question "How do I run SQLMesh models on a schedule?"
170
-
You can run SQLMesh models using the [built-in scheduler](../guides/scheduling.md#built-in-scheduler) or with the native [Airflow integration](../integrations/airflow.md).
170
+
You can run SQLMesh models using the [built-in scheduler](../guides/scheduling.md#built-in-scheduler) or using [Tobiko Cloud](../cloud/features/scheduler/scheduler.md)
171
171
172
172
Both approaches use each model's `cron` parameter to determine when the model should run - see the [question about `cron` above](#cron-question) for more information.
173
173
174
174
The built-in scheduler works by executing the command `sqlmesh run`. A sensible approach to running on your project on a schedule is to use Linux’s `cron` tool to execute `sqlmesh run` on a cadence at least as frequent as your briefest SQLMesh model `cron` parameter. For example, if your most frequent model’s `cron` is hour, the `cron` tool should execute `sqlmesh run` at least every hour.
175
175
176
176
??? question "How do I use SQLMesh with Airflow?"
177
-
SQLMesh has first-class support for Airflow - learn more [here](../integrations/airflow.md).
177
+
Tobiko Cloud offers first-class support for Airflow - learn more [here](../cloud/features/scheduler/airflow.md)
178
+
179
+
??? question "How do I use SQLMesh with Dagster?"
180
+
Tobiko Cloud offers first-class support for Dagster - learn more [here](../cloud/features/scheduler/dagster.md)
Copy file name to clipboardExpand all lines: docs/guides/configuration.md
+2-85Lines changed: 2 additions & 85 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -501,7 +501,7 @@ These pages describe the connection configuration options for each execution eng
501
501
502
502
Configuration for the state backend connection if different from the data warehouse connection.
503
503
504
-
The data warehouse connection is used to store SQLMesh state if the `state_connection` key is not specified, unless the configuration uses an Airflow or Google Cloud Composer scheduler. If using one of those schedulers, the scheduler's database is used (not the data warehouse) unless an [Airflow Connection has been configured](../integrations/airflow.md#state-connection).
504
+
The data warehouse connection is used to store SQLMesh state if the `state_connection` key is not specified.
505
505
506
506
Unlike data transformations, storing state information requires database transactions. Data warehouses aren’t optimized for executing transactions, and storing state information in them can slow down your project or produce corrupted data due to simultaneous writes to the same table. Therefore, production SQLMesh deployments should use a dedicated state connection.
507
507
@@ -675,7 +675,7 @@ Configuration for a connection used to run unit tests. An in-memory DuckDB datab
675
675
676
676
### Scheduler
677
677
678
-
Identifies which scheduler backend to use. The scheduler backend is used both for storing metadata and for executing [plans](../concepts/plans.md). By default, the scheduler type is set to `builtin`, which uses the existing SQL engine to store metadata. Use the `airflow` type integrate with Airflow.
678
+
Identifies which scheduler backend to use. The scheduler backend is used both for storing metadata and for executing [plans](../concepts/plans.md). By default, the scheduler type is set to `builtin`, which uses the existing SQL engine to store metadata.
679
679
680
680
These options are in the [scheduler](../reference/configuration.md#scheduler) section of the configuration reference page.
681
681
@@ -716,89 +716,6 @@ Example configuration:
716
716
717
717
No additional configuration options are supported by this scheduler type.
718
718
719
-
#### Airflow
720
-
721
-
Example configuration:
722
-
723
-
=== "YAML"
724
-
725
-
```yaml linenums="1"
726
-
gateways:
727
-
my_gateway:
728
-
scheduler:
729
-
type: airflow
730
-
airflow_url: <airflow_url>
731
-
username: <username>
732
-
password: <password>
733
-
```
734
-
735
-
=== "Python"
736
-
737
-
An Airflow scheduler is specified with an `AirflowSchedulerConfig` object.
See [Airflow Integration Guide](../integrations/airflow.md) for information about how to integrate Airflow with SQLMesh. See the [configuration reference page](../reference/configuration.md#airflow) for a list of all parameters.
762
-
763
-
#### Cloud Composer
764
-
765
-
The Google Cloud Composer scheduler type shares the same configuration options as the `airflow` type, except for `username` and `password`. Cloud Composer relies on `gcloud` authentication, so the `username` and `password` options are not required.
766
-
767
-
Example configuration:
768
-
769
-
=== "YAML"
770
-
771
-
```yaml linenums="1"
772
-
gateways:
773
-
my_gateway:
774
-
scheduler:
775
-
type: cloud_composer
776
-
airflow_url: <airflow_url>
777
-
```
778
-
779
-
=== "Python"
780
-
781
-
An Google Cloud Composer scheduler is specified with an `CloudComposerSchedulerConfig` object.
Copy file name to clipboardExpand all lines: docs/guides/connections.md
+1-3Lines changed: 1 addition & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,8 +2,6 @@
2
2
3
3
## Overview
4
4
5
-
**Note:** The following guide only applies when using the built-in scheduler. Connections are configured differently when using an external scheduler such as Airflow. See the [Scheduling guide](scheduling.md) for more details.
6
-
7
5
In order to deploy models and to apply changes to them, you must configure a connection to your Data Warehouse and, optionally, connection to the database where the SQLMesh state is stored. This can be done in either the `config.yaml` file in your project folder, or the one in `~/.sqlmesh`.
8
6
9
7
Each connection is configured as part of a gateway which has a unique name associated with it. The gateway name can be used to select a specific combination of connection settings when using the CLI. For example:
@@ -23,7 +21,7 @@ sqlmesh --gateway local_db plan
23
21
24
22
## State connection
25
23
26
-
By default, the data warehouse connection is also used to store the SQLMesh state, unless the configuration uses an Airflow or Google Cloud Composer scheduler. If using one of those schedulers, the state connection defaults to the scheduler's database.
24
+
By default, the data warehouse connection is also used to store the SQLMesh state.
27
25
28
26
The state connection can be changed by providing different connection settings in the `state_connection` key of the gateway configuration:
0 commit comments