diff --git a/delta-sharing/change-data-feed/Oracle Delta Sharing CDF.ipynb b/delta-sharing/change-data-feed/Oracle Delta Sharing CDF.ipynb index 72c46cd..7a93af0 100644 --- a/delta-sharing/change-data-feed/Oracle Delta Sharing CDF.ipynb +++ b/delta-sharing/change-data-feed/Oracle Delta Sharing CDF.ipynb @@ -18,14 +18,6 @@ "This notebook reads **Change Data Feed (CDF)** from an Oracle Autonomous Database\n", "that publishes a Delta Sharing endpoint, and displays the raw change rows.\n", "\n", - "## Why a custom REST approach?\n", - "\n", - "| Problem | Root cause | Workaround |\n", - "|---|---|---|\n", - "| `spark.read.format(\"deltaSharing\")` throws `InvocationTargetException` | Spark's Java Delta Sharing connector is incompatible with Oracle endpoints on serverless compute | Use the Python `delta_sharing` REST client instead |\n", - "| `load_table_changes_as_pandas()` throws `KeyError: '_commit_timestamp'` | Oracle's file-level CDF omits the `_commit_timestamp` column that the library expects | Call the REST API directly and parse with `DeltaSharingReader._to_pandas()` |\n", - "| `spark.read.parquet(*urls)` throws `UNSUPPORTED_FILE_SYSTEM` | Spark can't read HTTPS pre-signed URLs from Oracle object storage | Download via HTTP with `_to_pandas()`, then convert to Spark DataFrame |\n", - "\n", "## Oracle's file-level CDF\n", "\n", "Oracle implements CDF at the **file level**, not the row level. When *any* row in a\n", @@ -247,4 +239,4 @@ }, "nbformat": 4, "nbformat_minor": 0 -} \ No newline at end of file +}