Cloud Dataproc V1 API - Class Google::Cloud::Dataproc::V1::PySparkNotebookBatch (v1.12.0)

Reference documentation and code samples for the Cloud Dataproc V1 API class Google::Cloud::Dataproc::V1::PySparkNotebookBatch.

A configuration for running a PySpark Notebook batch workload.

Inherits

  • Object

Extended By

  • Google::Protobuf::MessageExts::ClassMethods

Includes

  • Google::Protobuf::MessageExts

Methods

#archive_uris

def archive_uris() -> ::Array<::String>
Returns
  • (::Array<::String>) — Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

#archive_uris=

def archive_uris=(value) -> ::Array<::String>
Parameter
  • value (::Array<::String>) — Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
Returns
  • (::Array<::String>) — Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

#file_uris

def file_uris() -> ::Array<::String>
Returns
  • (::Array<::String>) — Optional. HCFS URIs of files to be placed in the working directory of each executor

#file_uris=

def file_uris=(value) -> ::Array<::String>
Parameter
  • value (::Array<::String>) — Optional. HCFS URIs of files to be placed in the working directory of each executor
Returns
  • (::Array<::String>) — Optional. HCFS URIs of files to be placed in the working directory of each executor

#jar_file_uris

def jar_file_uris() -> ::Array<::String>
Returns
  • (::Array<::String>) — Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

#jar_file_uris=

def jar_file_uris=(value) -> ::Array<::String>
Parameter
  • value (::Array<::String>) — Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
Returns
  • (::Array<::String>) — Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

#notebook_file_uri

def notebook_file_uri() -> ::String
Returns
  • (::String) — Required. The HCFS URI of the notebook file to execute.

#notebook_file_uri=

def notebook_file_uri=(value) -> ::String
Parameter
  • value (::String) — Required. The HCFS URI of the notebook file to execute.
Returns
  • (::String) — Required. The HCFS URI of the notebook file to execute.

#params

def params() -> ::Google::Protobuf::Map{::String => ::String}
Returns
  • (::Google::Protobuf::Map{::String => ::String}) — Optional. The parameters to pass to the notebook.

#params=

def params=(value) -> ::Google::Protobuf::Map{::String => ::String}
Parameter
  • value (::Google::Protobuf::Map{::String => ::String}) — Optional. The parameters to pass to the notebook.
Returns
  • (::Google::Protobuf::Map{::String => ::String}) — Optional. The parameters to pass to the notebook.

#python_file_uris

def python_file_uris() -> ::Array<::String>
Returns
  • (::Array<::String>) — Optional. HCFS URIs of Python files to pass to the PySpark framework.

#python_file_uris=

def python_file_uris=(value) -> ::Array<::String>
Parameter
  • value (::Array<::String>) — Optional. HCFS URIs of Python files to pass to the PySpark framework.
Returns
  • (::Array<::String>) — Optional. HCFS URIs of Python files to pass to the PySpark framework.