This repository contains the code for the Data Engineering Pipelines with Snowpark Python Snowflake Quickstart.
Data Engineering Pipelines with Snowpark Python on quickstarts.snowflake.com.
Run the commands in the below script by clicking the execute all run button
Path : steps/01_setup_snowflake.sql
Loading the raw Tasty Bytes POS and Customer loyalty data from raw Parquet files in S3 bucket s3://sfquickstarts/data-engineering-with-snowpark-python/ to our RAW_POS and RAW_CUSTOMER schemas in Snowflake
python 02_load_raw.py`Getting the free weather dats from snowflake marketplace : "Weather Source LLC: frostbyte"
Creating a view to simplify the raw POS schema by joining together 6 different tables and picking only the columns we need
python 04_create_pos_view.pyTo run: python fahrenheit_to_celsius_udf/function.py
Deploy:
snow snowpark build
snow snowpark deployTo run: python orders_update_sp/procedure.py
Deploy:
snow snowpark build
snow snowpark deployTo run: python daily_city_metrics_update_sp/procedure.py
Deploy:
snow snowpark build
snow snowpark deployCreate two tasks, one for each stored procedure, and chain them together. We will then run the tasks.
Run the commands in the below script by clicking the execute all run button
Path : steps/08_orchestrate_jobs.sql
Adding new data to our POS order tables and then running our entire end-to-end pipeline to process the new data
Run the commands in the below script by clicking the execute all run button
Path : steps/09_process_incrementally.sql
Update the Fahrenheit to Celsius UDF
path: steps/05_fahrenheit_to_celsius_udf/fahrenheit_to_celsius_udf/function.py
Add scipy to our requirements.txt Run:
pip install -r requirements.txtStore Snowflake credentials in GitHub
Deploy through Action's tab - workflow














