[[["易于理解","easyToUnderstand","thumb-up"],["解决了我的问题","solvedMyProblem","thumb-up"],["其他","otherUp","thumb-up"]],[["很难理解","hardToUnderstand","thumb-down"],["信息或示例代码不正确","incorrectInformationOrSampleCode","thumb-down"],["没有我需要的信息/示例","missingTheInformationSamplesINeed","thumb-down"],["翻译问题","translationIssue","thumb-down"],["其他","otherDown","thumb-down"]],["最后更新时间 (UTC):2025-04-03。"],[[["This guide demonstrates how to submit a Spark job to an existing Dataproc cluster using a Google APIs Explorer template."],["Before submitting a job, a Dataproc cluster must be created using methods like the APIs Explorer, Google Cloud console, gcloud CLI, or Cloud Client Libraries."],["The Spark job example provided calculates a rough value for pi, and requires parameters such as projectId, region, clusterName, and specific job details like task count, jar file path, and main class."],["After submitting the job through the API, you can view the job output in the Dataproc Jobs page in the Google Cloud console."],["To avoid incurring charges, delete the Dataproc cluster using one of the provided methods if it is no longer needed."]]],[]]