# Functions
New creates a new Service.
NewService creates a new Service.
# Constants
See, edit, configure, and delete your Google Cloud data and see the email address for your Google Account.
# Structs
AcceleratorConfig: Specifies the type and number of accelerator cards attached to the instances of an instance.
AccessSessionSparkApplicationEnvironmentInfoResponse: Environment details of a Saprk Application.
AccessSessionSparkApplicationJobResponse: Details of a particular job associated with Spark Application.
AccessSessionSparkApplicationResponse: A summary of Spark Application.
AccessSessionSparkApplicationSqlQueryResponse: Details of a query for a Spark Application.
AccessSessionSparkApplicationSqlSparkPlanGraphResponse: SparkPlanGraph for a Spark Application execution limited to maximum 10000 clusters.
AccessSessionSparkApplicationStageAttemptResponse: Stage Attempt for a Stage of a Spark Application.
AccessSessionSparkApplicationStageRddOperationGraphResponse: RDD operation graph for a Spark Application Stage limited to maximum 10000 clusters.
AccessSparkApplicationEnvironmentInfoResponse: Environment details of a Saprk Application.
AccessSparkApplicationJobResponse: Details of a particular job associated with Spark Application.
AccessSparkApplicationResponse: A summary of Spark Application.
AccessSparkApplicationSqlQueryResponse: Details of a query for a Spark Application.
AccessSparkApplicationSqlSparkPlanGraphResponse: SparkPlanGraph for a Spark Application execution limited to maximum 10000 clusters.
AccessSparkApplicationStageAttemptResponse: Stage Attempt for a Stage of a Spark Application.
AccessSparkApplicationStageRddOperationGraphResponse: RDD operation graph for a Spark Application Stage limited to maximum 10000 clusters.
AnalyzeBatchRequest: A request to analyze a batch workload.
AnalyzeOperationMetadata: Metadata describing the Analyze operation.
ApplicationAttemptInfo: Specific attempt of an application.
ApplicationEnvironmentInfo: Details about the Environment that the application is running in.
ApplicationInfo: High level information corresponding to an application.
AutoscalingConfig: Autoscaling Policy config associated with the cluster.
AutoscalingPolicy: Describes an autoscaling policy for Dataproc cluster autoscaler.
AutotuningConfig: Autotuning configuration of the workload.
AuxiliaryNodeGroup: Node group identification and configuration information.
AuxiliaryServicesConfig: Auxiliary services configuration for a Cluster.
BasicAutoscalingAlgorithm: Basic algorithm for autoscaling.
BasicYarnAutoscalingConfig: Basic autoscaling configurations for YARN.
Batch: A representation of a batch workload in the service.
BatchOperationMetadata: Metadata describing the Batch operation.
Binding: Associates members, or principals, with a role.
BuildInfo: Native Build Info.
CancelJobRequest: A request to cancel a job.
Cluster: Describes the identifying information, config, and status of a Dataproc cluster.
ClusterConfig: The cluster config.
ClusterMetrics: Contains cluster daemon metrics, such as HDFS and YARN stats.Beta Feature: This report is available for testing purposes only.
ClusterOperation: The cluster operation triggered by a workflow.
ClusterOperationMetadata: Metadata describing the operation.
ClusterOperationStatus: The status of the operation.
ClusterSelector: A selector that chooses target cluster for jobs based on metadata.
ClusterStatus: The status of a cluster and its instances.
ClusterToRepair: Cluster to be repaired.
ConfidentialInstanceConfig: Confidential Instance Config for clusters using Confidential VMs (https://cloud.google.com/compute/confidential-vm/docs).
ConsolidatedExecutorSummary: Consolidated summary about executors used by the application.
DataprocMetricConfig: Dataproc metric config.
DiagnoseClusterRequest: A request to collect cluster diagnostic information.
DiagnoseClusterResults: The location of diagnostic output.
DiskConfig: Specifies the config of disk options for a group of VM instances.
DriverSchedulingConfig: Driver scheduling configuration.
Empty: A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs.
EncryptionConfig: Encryption settings for the cluster.
EndpointConfig: Endpoint config for this cluster.
EnvironmentConfig: Environment configuration for a workload.
ExecutionConfig: Execution configuration for a workload.
ExecutorResourceRequest: Resources used per executor used by the application.
ExecutorStageSummary: Executor resources consumed by a stage.
ExecutorSummary: Details about executors used by the application.
Expr: Represents a textual expression in the Common Expression Language (CEL) syntax.
FallbackReason: Native SQL Execution Data.
FlinkJob: A Dataproc job for running Apache Flink applications on YARN.
GceClusterConfig: Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster.
GetIamPolicyRequest: Request message for GetIamPolicy method.
GetPolicyOptions: Encapsulates settings provided to GetIamPolicy.
GkeClusterConfig: The cluster's GKE config.
GkeNodeConfig: Parameters that describe cluster nodes.
GkeNodePoolAcceleratorConfig: A GkeNodeConfigAcceleratorConfig represents a Hardware Accelerator request for a node pool.
GkeNodePoolAutoscalingConfig: GkeNodePoolAutoscaling contains information the cluster autoscaler needs to adjust the size of the node pool to the current cluster usage.
GkeNodePoolConfig: The configuration of a GKE node pool used by a Dataproc-on-GKE cluster (https://cloud.google.com/dataproc/docs/concepts/jobs/dataproc-gke#create-a-dataproc-on-gke-cluster).
GkeNodePoolTarget: GKE node pools that Dataproc workloads run on.
GoogleCloudDataprocV1WorkflowTemplateEncryptionConfig: Encryption settings for encrypting workflow template job arguments.
HadoopJob: A Dataproc job for running Apache Hadoop MapReduce (https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html).
HiveJob: A Dataproc job for running Apache Hive (https://hive.apache.org/) queries on YARN.
IdentityConfig: Identity related configuration, including service account based secure multi-tenancy user mappings.
InjectCredentialsRequest: A request to inject credentials into a cluster.
InputMetrics: Metrics about the input data read by the task.
InstanceFlexibilityPolicy: Instance flexibility Policy allowing a mixture of VM shapes and provisioning models.
InstanceGroupAutoscalingPolicyConfig: Configuration for the size bounds of an instance group, including its proportional size to other groups.
InstanceGroupConfig: The config settings for Compute Engine resources in an instance group, such as a master or worker group.
InstanceReference: A reference to a Compute Engine instance.
InstanceSelection: Defines machines types and a rank to which the machines types belong.
InstanceSelectionResult: Defines a mapping from machine types to the number of VMs that are created with each machine type.
InstantiateWorkflowTemplateRequest: A request to instantiate a workflow template.
Interval: Represents a time interval, encoded as a Timestamp start (inclusive) and a Timestamp end (exclusive).The start must be less than or equal to the end.
Job: A Dataproc job resource.
JobData: Data corresponding to a spark job.
JobMetadata: Job Operation metadata.
JobPlacement: Dataproc job config.
JobReference: Encapsulates the full scoping used to reference a job.
JobScheduling: Job scheduling options.
JobsSummary: Data related to Jobs page summary.
JobStatus: Dataproc job status.
JupyterConfig: Jupyter configuration for an interactive session.
KerberosConfig: Specifies Kerberos related configuration.
KubernetesClusterConfig: The configuration for running the Dataproc cluster on Kubernetes.
KubernetesSoftwareConfig: The software configuration for this Dataproc cluster running on Kubernetes.
LifecycleConfig: Specifies the cluster auto-delete schedule configuration.
ListAutoscalingPoliciesResponse: A response to a request to list autoscaling policies in a project.
ListBatchesResponse: A list of batch workloads.
ListClustersResponse: The list of all clusters in a project.
ListJobsResponse: A list of jobs in a project.
ListOperationsResponse: The response message for Operations.ListOperations.
ListSessionsResponse: A list of interactive sessions.
ListSessionTemplatesResponse: A list of session templates.
ListWorkflowTemplatesResponse: A response to a request to list workflow templates in a project.
LoggingConfig: The runtime logging config of the job.
ManagedCluster: Cluster that is managed by the workflow.
ManagedGroupConfig: Specifies the resources used to actively manage an instance group.
MetastoreConfig: Specifies a Metastore configuration.
Metric: A Dataproc custom metric.
NamespacedGkeDeploymentTarget: Deprecated.
NativeSqlExecutionUiData: Native SQL Execution Data.
NodeGroup: Dataproc Node Group.
NodeGroupAffinity: Node Group Affinity for clusters using sole-tenant node groups.
NodeGroupOperationMetadata: Metadata describing the node group operation.
NodeInitializationAction: Specifies an executable to run on a fully configured node and a timeout period for executable completion.
NodePool: indicating a list of workers of same type.
Operation: This resource represents a long-running operation that is the result of a network API call.
OrderedJob: A job executed by the workflow.
OutputMetrics: Metrics about the data written by the task.
ParameterValidation: Configuration for parameter validation.
PeripheralsConfig: Auxiliary services configuration for a workload.
PigJob: A Dataproc job for running Apache Pig (https://pig.apache.org/) queries on YARN.
Policy: An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources.A Policy is a collection of bindings.
PoolData: Pool Data.
PrestoJob: A Dataproc job for running Presto (https://prestosql.io/) queries.
ProcessSummary: Process Summary.
ProvisioningModelMix: Defines how Dataproc should create VMs with a mixture of provisioning models.
PyPiRepositoryConfig: Configuration for PyPi repository.
PySparkBatch: A configuration for running an Apache PySpark (https://spark.apache.org/docs/latest/api/python/getting_started/quickstart.html) batch workload.
PySparkJob: A Dataproc job for running Apache PySpark (https://spark.apache.org/docs/latest/api/python/index.html#pyspark-overview) applications on YARN.
Quantiles: Quantile metrics data related to Tasks.
QueryList: A list of queries to run on a cluster.
RddDataDistribution: Details about RDD usage.
RddOperationCluster: A grouping of nodes representing higher level constructs (stage, job etc.).
RddOperationEdge: A directed edge representing dependency between two RDDs.
RddOperationGraph: Graph representing RDD dependencies.
RddOperationNode: A node in the RDD operation graph.
RddPartitionInfo: Information about RDD partitions.
RddStorageInfo: Overall data about RDD storage.
RegexValidation: Validation based on regular expressions.
RepairClusterRequest: A request to repair a cluster.
RepositoryConfig: Configuration for dependency repositories.
ReservationAffinity: Reservation Affinity for consuming Zonal reservation.
ResizeNodeGroupRequest: A request to resize a node group.
ResourceProfileInfo: Resource profile that contains information about all the resources required by executors and tasks.
RuntimeConfig: Runtime configuration for a workload.
RuntimeInfo: Runtime information about workload execution.
SearchSessionSparkApplicationExecutorsResponse: List of Executors associated with a Spark Application.
SearchSessionSparkApplicationExecutorStageSummaryResponse: List of Executors associated with a Spark Application Stage.
SearchSessionSparkApplicationJobsResponse: A list of Jobs associated with a Spark Application.
SearchSessionSparkApplicationSqlQueriesResponse: List of all queries for a Spark Application.
SearchSessionSparkApplicationsResponse: A list of summary of Spark Applications.
SearchSessionSparkApplicationStageAttemptsResponse: A list of Stage Attempts for a Stage of a Spark Application.
SearchSessionSparkApplicationStageAttemptTasksResponse: List of tasks for a stage of a Spark Application.
SearchSessionSparkApplicationStagesResponse: A list of stages associated with a Spark Application.
SearchSparkApplicationExecutorsResponse: List of Executors associated with a Spark Application.
SearchSparkApplicationExecutorStageSummaryResponse: List of Executors associated with a Spark Application Stage.
SearchSparkApplicationJobsResponse: A list of Jobs associated with a Spark Application.
SearchSparkApplicationSqlQueriesResponse: List of all queries for a Spark Application.
SearchSparkApplicationsResponse: A list of summary of Spark Applications.
SearchSparkApplicationStageAttemptsResponse: A list of Stage Attempts for a Stage of a Spark Application.
SearchSparkApplicationStageAttemptTasksResponse: List of tasks for a stage of a Spark Application.
SearchSparkApplicationStagesResponse: A list of stages associated with a Spark Application.
SecurityConfig: Security related configuration, including encryption, Kerberos, etc.
Session: A representation of a session.
SessionOperationMetadata: Metadata describing the Session operation.
SessionStateHistory: Historical state information.
SessionTemplate: A representation of a session template.
SetIamPolicyRequest: Request message for SetIamPolicy method.
ShieldedInstanceConfig: Shielded Instance Config for clusters using Compute Engine Shielded VMs (https://cloud.google.com/security/shielded-cloud/shielded-vm).
ShuffleReadMetrics: Shuffle data read by the task.
ShuffleWriteMetrics: Shuffle data written by task.
SoftwareConfig: Specifies the selection and config of software inside the cluster.
SparkApplication: A summary of Spark Application.
SparkBatch: A configuration for running an Apache Spark (https://spark.apache.org/) batch workload.
SparkConnectConfig: Spark connect configuration for an interactive session.
SparkHistoryServerConfig: Spark History Server configuration for the workload.
SparkJob: A Dataproc job for running Apache Spark (https://spark.apache.org/) applications on YARN.
SparkPlanGraph: A graph used for storing information of an executionPlan of DataFrame.
SparkPlanGraphCluster: Represents a tree of spark plan.
SparkPlanGraphEdge: Represents a directed edge in the spark plan tree from child to parent.
SparkPlanGraphNode: Represents a node in the spark plan tree.
SparkPlanGraphNodeWrapper: Wrapper user to represent either a node or a cluster.
SparkRBatch: A configuration for running an Apache SparkR (https://spark.apache.org/docs/latest/sparkr.html) batch workload.
SparkRJob: A Dataproc job for running Apache SparkR (https://spark.apache.org/docs/latest/sparkr.html) applications on YARN.
SparkSqlBatch: A configuration for running Apache Spark SQL (https://spark.apache.org/sql/) queries as a batch workload.
SparkSqlJob: A Dataproc job for running Apache Spark SQL (https://spark.apache.org/sql/) queries.
SparkStandaloneAutoscalingConfig: Basic autoscaling configurations for Spark Standalone.
SparkWrapperObject: Outer message that contains the data obtained from spark listener, packaged with information that is required to process it.
SpeculationStageSummary: Details of the speculation task when speculative execution is enabled.
SqlExecutionUiData: SQL Execution Data.
SqlPlanMetric: Metrics related to SQL execution.
StageAttemptTasksSummary: Data related to tasks summary for a Spark Stage Attempt.
StageData: Data corresponding to a stage.
StageInputMetrics: Metrics about the input read by the stage.
StageMetrics: Stage Level Aggregated Metrics.
StageOutputMetrics: Metrics about the output written by the stage.
StageShuffleReadMetrics: Shuffle data read for the stage.
StageShuffleWriteMetrics: Shuffle data written for the stage.
StagesSummary: Data related to Stages page summary.
StartClusterRequest: A request to start a cluster.
StartupConfig: Configuration to handle the startup of instances during cluster create and update process.
StateHistory: Historical state information.
Status: The Status type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs.
StopClusterRequest: A request to stop a cluster.
StreamBlockData: Stream Block Data.
StreamingQueryData: Streaming.
SubmitJobRequest: A request to submit a job.
SummarizeSessionSparkApplicationExecutorsResponse: Consolidated summary of executors for a Spark Application.
SummarizeSessionSparkApplicationJobsResponse: Summary of a Spark Application jobs.
SummarizeSessionSparkApplicationStageAttemptTasksResponse: Summary of tasks for a Spark Application stage attempt.
SummarizeSessionSparkApplicationStagesResponse: Summary of a Spark Application stages.
SummarizeSparkApplicationExecutorsResponse: Consolidated summary of executors for a Spark Application.
SummarizeSparkApplicationJobsResponse: Summary of a Spark Application jobs.
SummarizeSparkApplicationStageAttemptTasksResponse: Summary of tasks for a Spark Application stage attempt.
SummarizeSparkApplicationStagesResponse: Summary of a Spark Application stages.
TaskData: Data corresponding to tasks created by spark.
TaskMetrics: Executor Task Metrics.
TaskResourceRequest: Resources used per task created by the application.
TemplateParameter: A configurable parameter that replaces one or more fields in the template.
TerminateSessionRequest: A request to terminate an interactive session.
TestIamPermissionsRequest: Request message for TestIamPermissions method.
TestIamPermissionsResponse: Response message for TestIamPermissions method.
TrinoJob: A Dataproc job for running Trino (https://trino.io/) queries.
UsageMetrics: Usage metrics represent approximate total resources consumed by a workload.
UsageSnapshot: The usage snapshot represents the resources consumed by a workload at a specified time.
ValueValidation: Validation based on a list of allowed values.
VirtualClusterConfig: The Dataproc cluster config for a cluster that does not directly control the underlying compute resources, such as a Dataproc-on-GKE cluster (https://cloud.google.com/dataproc/docs/guides/dpgke/dataproc-gke-overview).
WorkflowGraph: The workflow graph.
WorkflowMetadata: A Dataproc workflow template resource.
WorkflowNode: The workflow node.
WorkflowTemplate: A Dataproc workflow template resource.
WorkflowTemplatePlacement: Specifies workflow execution target.Either managed_cluster or cluster_selector is required.
WriteSessionSparkApplicationContextRequest: Write Spark Application data to internal storage systems.
WriteSessionSparkApplicationContextResponse: Response returned as an acknowledgement of receipt of data.
WriteSparkApplicationContextRequest: Write Spark Application data to internal storage systems.
WriteSparkApplicationContextResponse: Response returned as an acknowledgement of receipt of data.
YarnApplication: A YARN application created by a job.