Optional. HCFS URIs of archives to be extracted into the working directory
of each executor. Supported file types:
.jar, .tar, .tar.gz, .tgz, and .zip.
Optional. The arguments to pass to the driver. Do not include arguments
that can be set as batch properties, such as --conf, since a collision
can occur that causes an incorrect batch submission.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-07 UTC."],[[["\u003cp\u003eThe latest version available for the Google Cloud Dataproc v1 API class PySparkBatch is 5.17.0, and multiple previous versions, ranging from 3.1.0 to 5.16.0, are also documented.\u003c/p\u003e\n"],["\u003cp\u003eThe \u003ccode\u003ePySparkBatch\u003c/code\u003e class is used for configuring Apache PySpark batch workloads within the Google Cloud Dataproc environment, as shown through the use of this class in the .NET library.\u003c/p\u003e\n"],["\u003cp\u003eThe \u003ccode\u003ePySparkBatch\u003c/code\u003e class implements interfaces such as \u003ccode\u003eIMessage\u003c/code\u003e, \u003ccode\u003eIEquatable\u003c/code\u003e, \u003ccode\u003eIDeepCloneable\u003c/code\u003e, and \u003ccode\u003eIBufferMessage\u003c/code\u003e and it inherits from the \u003ccode\u003eobject\u003c/code\u003e class.\u003c/p\u003e\n"],["\u003cp\u003eKey properties within the \u003ccode\u003ePySparkBatch\u003c/code\u003e class include \u003ccode\u003eMainPythonFileUri\u003c/code\u003e, \u003ccode\u003eArchiveUris\u003c/code\u003e, \u003ccode\u003eArgs\u003c/code\u003e, \u003ccode\u003eFileUris\u003c/code\u003e, \u003ccode\u003eJarFileUris\u003c/code\u003e, and \u003ccode\u003ePythonFileUris\u003c/code\u003e, which allow users to define various parameters for their PySpark batch jobs, such as file locations, arguments, and archive files.\u003c/p\u003e\n"],["\u003cp\u003eThe main file type for this class is \u003ccode\u003e.py\u003c/code\u003e, and the supported archive file types are \u003ccode\u003e.jar\u003c/code\u003e, \u003ccode\u003e.tar\u003c/code\u003e, \u003ccode\u003e.tar.gz\u003c/code\u003e, \u003ccode\u003e.tgz\u003c/code\u003e, and \u003ccode\u003e.zip\u003c/code\u003e, while \u003ccode\u003e.egg\u003c/code\u003e, \u003ccode\u003e.zip\u003c/code\u003e are supported as python file types.\u003c/p\u003e\n"]]],[],null,["# Google Cloud Dataproc v1 API - Class PySparkBatch (5.20.0)\n\nVersion latestkeyboard_arrow_down\n\n- [5.20.0 (latest)](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/latest/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.19.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.19.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.18.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.18.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.17.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.17.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.16.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.16.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.15.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.15.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.14.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.14.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.13.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.13.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.12.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.12.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.11.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.11.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.10.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.10.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.9.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.9.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.8.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.8.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.7.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.7.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.6.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.6.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.5.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.5.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.4.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.4.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.3.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.3.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.2.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.2.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.1.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.1.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [5.0.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/5.0.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [4.0.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/4.0.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [3.4.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/3.4.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [3.3.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/3.3.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [3.2.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/3.2.0/Google.Cloud.Dataproc.V1.PySparkBatch)\n- [3.1.0](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/3.1.0/Google.Cloud.Dataproc.V1.PySparkBatch) \n\n public sealed class PySparkBatch : IMessage\u003cPySparkBatch\u003e, IEquatable\u003cPySparkBatch\u003e, IDeepCloneable\u003cPySparkBatch\u003e, IBufferMessage, IMessage\n\nReference documentation and code samples for the Google Cloud Dataproc v1 API class PySparkBatch.\n\nA configuration for running an\n[Apache\nPySpark](https://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html)\nbatch workload. \n\nInheritance\n-----------\n\n[object](https://learn.microsoft.com/dotnet/api/system.object) \\\u003e PySparkBatch \n\nImplements\n----------\n\n[IMessage](https://cloud.google.com/dotnet/docs/reference/Google.Protobuf/latest/Google.Protobuf.IMessage-1.html)[PySparkBatch](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/latest/Google.Cloud.Dataproc.V1.PySparkBatch), [IEquatable](https://learn.microsoft.com/dotnet/api/system.iequatable-1)[PySparkBatch](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/latest/Google.Cloud.Dataproc.V1.PySparkBatch), [IDeepCloneable](https://cloud.google.com/dotnet/docs/reference/Google.Protobuf/latest/Google.Protobuf.IDeepCloneable-1.html)[PySparkBatch](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/latest/Google.Cloud.Dataproc.V1.PySparkBatch), [IBufferMessage](https://cloud.google.com/dotnet/docs/reference/Google.Protobuf/latest/Google.Protobuf.IBufferMessage.html), [IMessage](https://cloud.google.com/dotnet/docs/reference/Google.Protobuf/latest/Google.Protobuf.IMessage.html) \n\nInherited Members\n-----------------\n\n[object.GetHashCode()](https://learn.microsoft.com/dotnet/api/system.object.gethashcode) \n[object.GetType()](https://learn.microsoft.com/dotnet/api/system.object.gettype) \n[object.ToString()](https://learn.microsoft.com/dotnet/api/system.object.tostring)\n\nNamespace\n---------\n\n[Google.Cloud.Dataproc.V1](/dotnet/docs/reference/Google.Cloud.Dataproc.V1/latest/Google.Cloud.Dataproc.V1)\n\nAssembly\n--------\n\nGoogle.Cloud.Dataproc.V1.dll\n\nConstructors\n------------\n\n### PySparkBatch()\n\n public PySparkBatch()\n\n### PySparkBatch(PySparkBatch)\n\n public PySparkBatch(PySparkBatch other)\n\nProperties\n----------\n\n### ArchiveUris\n\n public RepeatedField\u003cstring\u003e ArchiveUris { get; }\n\nOptional. HCFS URIs of archives to be extracted into the working directory\nof each executor. Supported file types:\n`.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`.\n\n### Args\n\n public RepeatedField\u003cstring\u003e Args { get; }\n\nOptional. The arguments to pass to the driver. Do not include arguments\nthat can be set as batch properties, such as `--conf`, since a collision\ncan occur that causes an incorrect batch submission.\n\n### FileUris\n\n public RepeatedField\u003cstring\u003e FileUris { get; }\n\nOptional. HCFS URIs of files to be placed in the working directory of\neach executor.\n\n### JarFileUris\n\n public RepeatedField\u003cstring\u003e JarFileUris { get; }\n\nOptional. HCFS URIs of jar files to add to the classpath of the\nSpark driver and tasks.\n\n### MainPythonFileUri\n\n public string MainPythonFileUri { get; set; }\n\nRequired. The HCFS URI of the main Python file to use as the Spark driver.\nMust be a .py file.\n\n### PythonFileUris\n\n public RepeatedField\u003cstring\u003e PythonFileUris { get; }\n\nOptional. HCFS file URIs of Python files to pass to the PySpark\nframework. Supported file types: `.py`, `.egg`, and `.zip`."]]