Skip to content

Commit 82e748a

Browse files
1 parent 4db652d commit 82e748a

File tree

5 files changed

+219
-6
lines changed

5 files changed

+219
-6
lines changed

clients/google-api-services-dataproc/v1/2.0.0/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ Add the following lines to your `pom.xml` file:
2222
<dependency>
2323
<groupId>com.google.apis</groupId>
2424
<artifactId>google-api-services-dataproc</artifactId>
25-
<version>v1-rev20260327-2.0.0</version>
25+
<version>v1-rev20260407-2.0.0</version>
2626
</dependency>
2727
</dependencies>
2828
</project>
@@ -35,7 +35,7 @@ repositories {
3535
mavenCentral()
3636
}
3737
dependencies {
38-
implementation 'com.google.apis:google-api-services-dataproc:v1-rev20260327-2.0.0'
38+
implementation 'com.google.apis:google-api-services-dataproc:v1-rev20260407-2.0.0'
3939
}
4040
```
4141

clients/google-api-services-dataproc/v1/2.0.0/com/google/api/services/dataproc/model/Batch.java

Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -81,6 +81,13 @@ public final class Batch extends com.google.api.client.json.GenericJson {
8181
@com.google.api.client.util.Key
8282
private PySparkBatch pysparkBatch;
8383

84+
/**
85+
* Optional. PySpark notebook batch config.
86+
* The value may be {@code null}.
87+
*/
88+
@com.google.api.client.util.Key
89+
private PySparkNotebookBatch pysparkNotebookBatch;
90+
8491
/**
8592
* Optional. Runtime configuration for the batch execution.
8693
* The value may be {@code null}.
@@ -277,6 +284,23 @@ public Batch setPysparkBatch(PySparkBatch pysparkBatch) {
277284
return this;
278285
}
279286

287+
/**
288+
* Optional. PySpark notebook batch config.
289+
* @return value or {@code null} for none
290+
*/
291+
public PySparkNotebookBatch getPysparkNotebookBatch() {
292+
return pysparkNotebookBatch;
293+
}
294+
295+
/**
296+
* Optional. PySpark notebook batch config.
297+
* @param pysparkNotebookBatch pysparkNotebookBatch or {@code null} for none
298+
*/
299+
public Batch setPysparkNotebookBatch(PySparkNotebookBatch pysparkNotebookBatch) {
300+
this.pysparkNotebookBatch = pysparkNotebookBatch;
301+
return this;
302+
}
303+
280304
/**
281305
* Optional. Runtime configuration for the batch execution.
282306
* @return value or {@code null} for none
Lines changed: 189 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,189 @@
1+
/*
2+
* Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except
3+
* in compliance with the License. You may obtain a copy of the License at
4+
*
5+
* http://www.apache.org/licenses/LICENSE-2.0
6+
*
7+
* Unless required by applicable law or agreed to in writing, software distributed under the License
8+
* is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express
9+
* or implied. See the License for the specific language governing permissions and limitations under
10+
* the License.
11+
*/
12+
/*
13+
* This code was generated by https://github.com/googleapis/google-api-java-client-services/
14+
* Modify at your own risk.
15+
*/
16+
17+
package com.google.api.services.dataproc.model;
18+
19+
/**
20+
* A configuration for running a PySpark Notebook batch workload.
21+
*
22+
* <p> This is the Java data model class that specifies how to parse/serialize into the JSON that is
23+
* transmitted over HTTP when working with the Cloud Dataproc API. For a detailed explanation see:
24+
* <a href="https://developers.google.com/api-client-library/java/google-http-java-client/json">https://developers.google.com/api-client-library/java/google-http-java-client/json</a>
25+
* </p>
26+
*
27+
* @author Google, Inc.
28+
*/
29+
@SuppressWarnings("javadoc")
30+
public final class PySparkNotebookBatch extends com.google.api.client.json.GenericJson {
31+
32+
/**
33+
* Optional. HCFS URIs of archives to be extracted into the working directory of each executor.
34+
* Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
35+
* The value may be {@code null}.
36+
*/
37+
@com.google.api.client.util.Key
38+
private java.util.List<java.lang.String> archiveUris;
39+
40+
/**
41+
* Optional. HCFS URIs of files to be placed in the working directory of each executor
42+
* The value may be {@code null}.
43+
*/
44+
@com.google.api.client.util.Key
45+
private java.util.List<java.lang.String> fileUris;
46+
47+
/**
48+
* Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
49+
* The value may be {@code null}.
50+
*/
51+
@com.google.api.client.util.Key
52+
private java.util.List<java.lang.String> jarFileUris;
53+
54+
/**
55+
* Required. The HCFS URI of the notebook file to execute.
56+
* The value may be {@code null}.
57+
*/
58+
@com.google.api.client.util.Key
59+
private java.lang.String notebookFileUri;
60+
61+
/**
62+
* Optional. The parameters to pass to the notebook.
63+
* The value may be {@code null}.
64+
*/
65+
@com.google.api.client.util.Key
66+
private java.util.Map<String, java.lang.String> params;
67+
68+
/**
69+
* Optional. HCFS URIs of Python files to pass to the PySpark framework.
70+
* The value may be {@code null}.
71+
*/
72+
@com.google.api.client.util.Key
73+
private java.util.List<java.lang.String> pythonFileUris;
74+
75+
/**
76+
* Optional. HCFS URIs of archives to be extracted into the working directory of each executor.
77+
* Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
78+
* @return value or {@code null} for none
79+
*/
80+
public java.util.List<java.lang.String> getArchiveUris() {
81+
return archiveUris;
82+
}
83+
84+
/**
85+
* Optional. HCFS URIs of archives to be extracted into the working directory of each executor.
86+
* Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
87+
* @param archiveUris archiveUris or {@code null} for none
88+
*/
89+
public PySparkNotebookBatch setArchiveUris(java.util.List<java.lang.String> archiveUris) {
90+
this.archiveUris = archiveUris;
91+
return this;
92+
}
93+
94+
/**
95+
* Optional. HCFS URIs of files to be placed in the working directory of each executor
96+
* @return value or {@code null} for none
97+
*/
98+
public java.util.List<java.lang.String> getFileUris() {
99+
return fileUris;
100+
}
101+
102+
/**
103+
* Optional. HCFS URIs of files to be placed in the working directory of each executor
104+
* @param fileUris fileUris or {@code null} for none
105+
*/
106+
public PySparkNotebookBatch setFileUris(java.util.List<java.lang.String> fileUris) {
107+
this.fileUris = fileUris;
108+
return this;
109+
}
110+
111+
/**
112+
* Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
113+
* @return value or {@code null} for none
114+
*/
115+
public java.util.List<java.lang.String> getJarFileUris() {
116+
return jarFileUris;
117+
}
118+
119+
/**
120+
* Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
121+
* @param jarFileUris jarFileUris or {@code null} for none
122+
*/
123+
public PySparkNotebookBatch setJarFileUris(java.util.List<java.lang.String> jarFileUris) {
124+
this.jarFileUris = jarFileUris;
125+
return this;
126+
}
127+
128+
/**
129+
* Required. The HCFS URI of the notebook file to execute.
130+
* @return value or {@code null} for none
131+
*/
132+
public java.lang.String getNotebookFileUri() {
133+
return notebookFileUri;
134+
}
135+
136+
/**
137+
* Required. The HCFS URI of the notebook file to execute.
138+
* @param notebookFileUri notebookFileUri or {@code null} for none
139+
*/
140+
public PySparkNotebookBatch setNotebookFileUri(java.lang.String notebookFileUri) {
141+
this.notebookFileUri = notebookFileUri;
142+
return this;
143+
}
144+
145+
/**
146+
* Optional. The parameters to pass to the notebook.
147+
* @return value or {@code null} for none
148+
*/
149+
public java.util.Map<String, java.lang.String> getParams() {
150+
return params;
151+
}
152+
153+
/**
154+
* Optional. The parameters to pass to the notebook.
155+
* @param params params or {@code null} for none
156+
*/
157+
public PySparkNotebookBatch setParams(java.util.Map<String, java.lang.String> params) {
158+
this.params = params;
159+
return this;
160+
}
161+
162+
/**
163+
* Optional. HCFS URIs of Python files to pass to the PySpark framework.
164+
* @return value or {@code null} for none
165+
*/
166+
public java.util.List<java.lang.String> getPythonFileUris() {
167+
return pythonFileUris;
168+
}
169+
170+
/**
171+
* Optional. HCFS URIs of Python files to pass to the PySpark framework.
172+
* @param pythonFileUris pythonFileUris or {@code null} for none
173+
*/
174+
public PySparkNotebookBatch setPythonFileUris(java.util.List<java.lang.String> pythonFileUris) {
175+
this.pythonFileUris = pythonFileUris;
176+
return this;
177+
}
178+
179+
@Override
180+
public PySparkNotebookBatch set(String fieldName, Object value) {
181+
return (PySparkNotebookBatch) super.set(fieldName, value);
182+
}
183+
184+
@Override
185+
public PySparkNotebookBatch clone() {
186+
return (PySparkNotebookBatch) super.clone();
187+
}
188+
189+
}

clients/google-api-services-dataproc/v1/2.0.0/pom.xml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,8 @@
88

99
<groupId>com.google.apis</groupId>
1010
<artifactId>google-api-services-dataproc</artifactId>
11-
<version>v1-rev20260327-2.0.0</version>
12-
<name>Cloud Dataproc API v1-rev20260327-2.0.0</name>
11+
<version>v1-rev20260407-2.0.0</version>
12+
<name>Cloud Dataproc API v1-rev20260407-2.0.0</name>
1313
<packaging>jar</packaging>
1414

1515
<inceptionYear>2011</inceptionYear>

clients/google-api-services-dataproc/v1/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ Add the following lines to your `pom.xml` file:
2222
<dependency>
2323
<groupId>com.google.apis</groupId>
2424
<artifactId>google-api-services-dataproc</artifactId>
25-
<version>v1-rev20260327-2.0.0</version>
25+
<version>v1-rev20260407-2.0.0</version>
2626
</dependency>
2727
</dependencies>
2828
</project>
@@ -35,7 +35,7 @@ repositories {
3535
mavenCentral()
3636
}
3737
dependencies {
38-
implementation 'com.google.apis:google-api-services-dataproc:v1-rev20260327-2.0.0'
38+
implementation 'com.google.apis:google-api-services-dataproc:v1-rev20260407-2.0.0'
3939
}
4040
```
4141

0 commit comments

Comments
 (0)