Send feedback
Class PySparkBatch (5.21.0)
Stay organized with collections
Save and categorize content based on your preferences.
Version latestkeyboard_arrow_down
PySparkBatch ( mapping = None , * , ignore_unknown_fields = False , ** kwargs )
A configuration for running an Apache
PySpark <https://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html>
__
batch workload.
Attributes
Name
Description
main_python_file_uri
str
Required. The HCFS URI of the main Python
file to use as the Spark driver. Must be a .py
file.
args
MutableSequence[str]
Optional. The arguments to pass to the driver. Do not
include arguments that can be set as batch properties, such
as --conf
, since a collision can occur that causes an
incorrect batch submission.
python_file_uris
MutableSequence[str]
Optional. HCFS file URIs of Python files to pass to the
PySpark framework. Supported file types: .py
, .egg
,
and .zip
.
jar_file_uris
MutableSequence[str]
Optional. HCFS URIs of jar files to add to
the classpath of the Spark driver and tasks.
file_uris
MutableSequence[str]
Optional. HCFS URIs of files to be placed in
the working directory of each executor.
archive_uris
MutableSequence[str]
Optional. HCFS URIs of archives to be extracted into the
working directory of each executor. Supported file types:
.jar
, .tar
, .tar.gz
, .tgz
, and .zip
.
Send feedback
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License , and code samples are licensed under the Apache 2.0 License . For details, see the Google Developers Site Policies . Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2025-08-07 UTC.
Need to tell us more?
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-07 UTC."],[],[],null,["# Class PySparkBatch (5.21.0)\n\nVersion latestkeyboard_arrow_down\n\n- [5.21.0 (latest)](/python/docs/reference/dataproc/latest/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.20.0](/python/docs/reference/dataproc/5.20.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.18.1](/python/docs/reference/dataproc/5.18.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.17.1](/python/docs/reference/dataproc/5.17.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.16.0](/python/docs/reference/dataproc/5.16.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.15.1](/python/docs/reference/dataproc/5.15.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.14.0](/python/docs/reference/dataproc/5.14.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.13.0](/python/docs/reference/dataproc/5.13.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.12.0](/python/docs/reference/dataproc/5.12.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.10.2](/python/docs/reference/dataproc/5.10.2/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.9.3](/python/docs/reference/dataproc/5.9.3/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.8.0](/python/docs/reference/dataproc/5.8.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.7.0](/python/docs/reference/dataproc/5.7.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.6.0](/python/docs/reference/dataproc/5.6.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.5.1](/python/docs/reference/dataproc/5.5.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.4.3](/python/docs/reference/dataproc/5.4.3/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.3.0](/python/docs/reference/dataproc/5.3.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.2.0](/python/docs/reference/dataproc/5.2.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.1.0](/python/docs/reference/dataproc/5.1.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [5.0.3](/python/docs/reference/dataproc/5.0.3/google.cloud.dataproc_v1.types.PySparkBatch)\n- [4.0.3](/python/docs/reference/dataproc/4.0.3/google.cloud.dataproc_v1.types.PySparkBatch)\n- [3.3.2](/python/docs/reference/dataproc/3.3.2/google.cloud.dataproc_v1.types.PySparkBatch)\n- [3.2.0](/python/docs/reference/dataproc/3.2.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [3.1.1](/python/docs/reference/dataproc/3.1.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [3.0.0](/python/docs/reference/dataproc/3.0.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [2.6.2](/python/docs/reference/dataproc/2.6.2/google.cloud.dataproc_v1.types.PySparkBatch)\n- [2.5.0](/python/docs/reference/dataproc/2.5.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [2.4.0](/python/docs/reference/dataproc/2.4.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [2.3.1](/python/docs/reference/dataproc/2.3.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [2.2.0](/python/docs/reference/dataproc/2.2.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [2.0.2](/python/docs/reference/dataproc/2.0.2/google.cloud.dataproc_v1.types.PySparkBatch)\n- [1.1.3](/python/docs/reference/dataproc/1.1.3/google.cloud.dataproc_v1.types.PySparkBatch)\n- [1.0.1](/python/docs/reference/dataproc/1.0.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [0.8.2](/python/docs/reference/dataproc/0.8.2/google.cloud.dataproc_v1.types.PySparkBatch)\n- [0.7.0](/python/docs/reference/dataproc/0.7.0/google.cloud.dataproc_v1.types.PySparkBatch)\n- [0.6.1](/python/docs/reference/dataproc/0.6.1/google.cloud.dataproc_v1.types.PySparkBatch)\n- [0.5.0](/python/docs/reference/dataproc/0.5.0/google.cloud.dataproc_v1.types.PySparkBatch) \n\n PySparkBatch(mapping=None, *, ignore_unknown_fields=False, **kwargs)\n\nA configuration for running an `Apache\nPySpark \u003chttps://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html\u003e`__\nbatch workload."]]