This repository was archived by the owner on Apr 1, 2026. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 68
Expand file tree
/
Copy pathio.py
More file actions
94 lines (85 loc) · 3.76 KB
/
io.py
File metadata and controls
94 lines (85 loc) · 3.76 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
# Copyright 2026 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import annotations
from typing import Mapping, Optional, Union
import pandas as pd
from bigframes.bigquery._operations.table import _get_table_metadata
import bigframes.core.logging.log_adapter as log_adapter
import bigframes.core.sql.io
import bigframes.session
@log_adapter.method_logger(custom_base_name="bigquery_io")
def load_data(
table_name: str,
*,
write_disposition: str = "INTO",
columns: Optional[Mapping[str, str]] = None,
partition_by: Optional[list[str]] = None,
cluster_by: Optional[list[str]] = None,
table_options: Optional[Mapping[str, Union[str, int, float, bool, list]]] = None,
from_files_options: Mapping[str, Union[str, int, float, bool, list]],
with_partition_columns: Optional[Mapping[str, str]] = None,
connection_name: Optional[str] = None,
session: Optional[bigframes.session.Session] = None,
) -> pd.Series:
"""
Loads data into a BigQuery table.
See the `BigQuery LOAD DATA DDL syntax
<https://docs.cloud.google.com/bigquery/docs/reference/standard-sql/load-statements>`_
for additional reference.
Args:
table_name (str):
The name of the table in BigQuery.
write_disposition (str, default "INTO"):
Whether to replace the table if it already exists ("OVERWRITE") or append to it ("INTO").
columns (Mapping[str, str], optional):
The table's schema.
partition_by (list[str], optional):
A list of partition expressions to partition the table by. See https://docs.cloud.google.com/bigquery/docs/reference/standard-sql/load-statements#partition_expression.
cluster_by (list[str], optional):
A list of columns to cluster the table by.
table_options (Mapping[str, Union[str, int, float, bool, list]], optional):
The table options.
from_files_options (Mapping[str, Union[str, int, float, bool, list]]):
The options for loading data from files.
with_partition_columns (Mapping[str, str], optional):
The table's partition columns.
connection_name (str, optional):
The connection to use for the table.
session (bigframes.session.Session, optional):
The session to use. If not provided, the default session is used.
Returns:
pandas.Series:
A Series with object dtype containing the table metadata. Reference
the `BigQuery Table REST API reference
<https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#Table>`_
for available fields.
"""
import bigframes.pandas as bpd
sql = bigframes.core.sql.io.load_data_ddl(
table_name=table_name,
write_disposition=write_disposition,
columns=columns,
partition_by=partition_by,
cluster_by=cluster_by,
table_options=table_options,
from_files_options=from_files_options,
with_partition_columns=with_partition_columns,
connection_name=connection_name,
)
if session is None:
bpd.read_gbq_query(sql)
session = bpd.get_global_session()
else:
session.read_gbq_query(sql)
return _get_table_metadata(bqclient=session.bqclient, table_name=table_name)