[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-07 UTC."],[[["\u003cp\u003eThis webpage provides documentation for the \u003ccode\u003eJob.TypeJobOneofCase\u003c/code\u003e enum within the Google Cloud Dataproc v1 API, detailing its various fields.\u003c/p\u003e\n"],["\u003cp\u003eThe enum represents different types of jobs that can be run on Dataproc, including Hadoop, Hive, Pig, Presto, Pyspark, Spark, SparkR, SparkSql, and Trino jobs, along with a None option.\u003c/p\u003e\n"],["\u003cp\u003eThe page offers access to the reference documentation across numerous versions of the API, spanning from version 3.1.0 up to the latest release, 5.17.0.\u003c/p\u003e\n"],["\u003cp\u003eThe documentation for \u003ccode\u003eJob.TypeJobOneofCase\u003c/code\u003e is located in the \u003ccode\u003eGoogle.Cloud.Dataproc.V1\u003c/code\u003e namespace, and it is included in the \u003ccode\u003eGoogle.Cloud.Dataproc.V1.dll\u003c/code\u003e assembly.\u003c/p\u003e\n"]]],[],null,[]]