dataproc

package
v0.0.0-...-063164c Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Feb 24, 2021 License: Apache-2.0 Imports: 15 Imported by: 29

Documentation

Index

Constants

This section is empty.

Variables

View Source
var (
	InstanceGroupConfig_Preemptibility_name = map[int32]string{
		0: "PREEMPTIBILITY_UNSPECIFIED",
		1: "NON_PREEMPTIBLE",
		2: "PREEMPTIBLE",
	}
	InstanceGroupConfig_Preemptibility_value = map[string]int32{
		"PREEMPTIBILITY_UNSPECIFIED": 0,
		"NON_PREEMPTIBLE":            1,
		"PREEMPTIBLE":                2,
	}
)

Enum value maps for InstanceGroupConfig_Preemptibility.

View Source
var (
	ClusterStatus_State_name = map[int32]string{
		0: "UNKNOWN",
		1: "CREATING",
		2: "RUNNING",
		3: "ERROR",
		4: "DELETING",
		5: "UPDATING",
	}
	ClusterStatus_State_value = map[string]int32{
		"UNKNOWN":  0,
		"CREATING": 1,
		"RUNNING":  2,
		"ERROR":    3,
		"DELETING": 4,
		"UPDATING": 5,
	}
)

Enum value maps for ClusterStatus_State.

View Source
var (
	ClusterStatus_Substate_name = map[int32]string{
		0: "UNSPECIFIED",
		1: "UNHEALTHY",
		2: "STALE_STATUS",
	}
	ClusterStatus_Substate_value = map[string]int32{
		"UNSPECIFIED":  0,
		"UNHEALTHY":    1,
		"STALE_STATUS": 2,
	}
)

Enum value maps for ClusterStatus_Substate.

View Source
var (
	ReservationAffinity_Type_name = map[int32]string{
		0: "TYPE_UNSPECIFIED",
		1: "NO_RESERVATION",
		2: "ANY_RESERVATION",
		3: "SPECIFIC_RESERVATION",
	}
	ReservationAffinity_Type_value = map[string]int32{
		"TYPE_UNSPECIFIED":     0,
		"NO_RESERVATION":       1,
		"ANY_RESERVATION":      2,
		"SPECIFIC_RESERVATION": 3,
	}
)

Enum value maps for ReservationAffinity_Type.

View Source
var (
	LoggingConfig_Level_name = map[int32]string{
		0: "LEVEL_UNSPECIFIED",
		1: "ALL",
		2: "TRACE",
		3: "DEBUG",
		4: "INFO",
		5: "WARN",
		6: "ERROR",
		7: "FATAL",
		8: "OFF",
	}
	LoggingConfig_Level_value = map[string]int32{
		"LEVEL_UNSPECIFIED": 0,
		"ALL":               1,
		"TRACE":             2,
		"DEBUG":             3,
		"INFO":              4,
		"WARN":              5,
		"ERROR":             6,
		"FATAL":             7,
		"OFF":               8,
	}
)

Enum value maps for LoggingConfig_Level.

View Source
var (
	JobStatus_State_name = map[int32]string{
		0: "STATE_UNSPECIFIED",
		1: "PENDING",
		8: "SETUP_DONE",
		2: "RUNNING",
		3: "CANCEL_PENDING",
		7: "CANCEL_STARTED",
		4: "CANCELLED",
		5: "DONE",
		6: "ERROR",
		9: "ATTEMPT_FAILURE",
	}
	JobStatus_State_value = map[string]int32{
		"STATE_UNSPECIFIED": 0,
		"PENDING":           1,
		"SETUP_DONE":        8,
		"RUNNING":           2,
		"CANCEL_PENDING":    3,
		"CANCEL_STARTED":    7,
		"CANCELLED":         4,
		"DONE":              5,
		"ERROR":             6,
		"ATTEMPT_FAILURE":   9,
	}
)

Enum value maps for JobStatus_State.

View Source
var (
	JobStatus_Substate_name = map[int32]string{
		0: "UNSPECIFIED",
		1: "SUBMITTED",
		2: "QUEUED",
		3: "STALE_STATUS",
	}
	JobStatus_Substate_value = map[string]int32{
		"UNSPECIFIED":  0,
		"SUBMITTED":    1,
		"QUEUED":       2,
		"STALE_STATUS": 3,
	}
)

Enum value maps for JobStatus_Substate.

View Source
var (
	YarnApplication_State_name = map[int32]string{
		0: "STATE_UNSPECIFIED",
		1: "NEW",
		2: "NEW_SAVING",
		3: "SUBMITTED",
		4: "ACCEPTED",
		5: "RUNNING",
		6: "FINISHED",
		7: "FAILED",
		8: "KILLED",
	}
	YarnApplication_State_value = map[string]int32{
		"STATE_UNSPECIFIED": 0,
		"NEW":               1,
		"NEW_SAVING":        2,
		"SUBMITTED":         3,
		"ACCEPTED":          4,
		"RUNNING":           5,
		"FINISHED":          6,
		"FAILED":            7,
		"KILLED":            8,
	}
)

Enum value maps for YarnApplication_State.

View Source
var (
	ListJobsRequest_JobStateMatcher_name = map[int32]string{
		0: "ALL",
		1: "ACTIVE",
		2: "NON_ACTIVE",
	}
	ListJobsRequest_JobStateMatcher_value = map[string]int32{
		"ALL":        0,
		"ACTIVE":     1,
		"NON_ACTIVE": 2,
	}
)

Enum value maps for ListJobsRequest_JobStateMatcher.

View Source
var (
	ClusterOperationStatus_State_name = map[int32]string{
		0: "UNKNOWN",
		1: "PENDING",
		2: "RUNNING",
		3: "DONE",
	}
	ClusterOperationStatus_State_value = map[string]int32{
		"UNKNOWN": 0,
		"PENDING": 1,
		"RUNNING": 2,
		"DONE":    3,
	}
)

Enum value maps for ClusterOperationStatus_State.

View Source
var (
	Component_name = map[int32]string{
		0: "COMPONENT_UNSPECIFIED",
		5: "ANACONDA",
		3: "HIVE_WEBHCAT",
		1: "JUPYTER",
		6: "PRESTO",
		4: "ZEPPELIN",
		8: "ZOOKEEPER",
	}
	Component_value = map[string]int32{
		"COMPONENT_UNSPECIFIED": 0,
		"ANACONDA":              5,
		"HIVE_WEBHCAT":          3,
		"JUPYTER":               1,
		"PRESTO":                6,
		"ZEPPELIN":              4,
		"ZOOKEEPER":             8,
	}
)

Enum value maps for Component.

View Source
var (
	WorkflowMetadata_State_name = map[int32]string{
		0: "UNKNOWN",
		1: "PENDING",
		2: "RUNNING",
		3: "DONE",
	}
	WorkflowMetadata_State_value = map[string]int32{
		"UNKNOWN": 0,
		"PENDING": 1,
		"RUNNING": 2,
		"DONE":    3,
	}
)

Enum value maps for WorkflowMetadata_State.

View Source
var (
	WorkflowNode_NodeState_name = map[int32]string{
		0: "NODE_STATE_UNSPECIFIED",
		1: "BLOCKED",
		2: "RUNNABLE",
		3: "RUNNING",
		4: "COMPLETED",
		5: "FAILED",
	}
	WorkflowNode_NodeState_value = map[string]int32{
		"NODE_STATE_UNSPECIFIED": 0,
		"BLOCKED":                1,
		"RUNNABLE":               2,
		"RUNNING":                3,
		"COMPLETED":              4,
		"FAILED":                 5,
	}
)

Enum value maps for WorkflowNode_NodeState.

View Source
var File_google_cloud_dataproc_v1_autoscaling_policies_proto protoreflect.FileDescriptor
View Source
var File_google_cloud_dataproc_v1_clusters_proto protoreflect.FileDescriptor
View Source
var File_google_cloud_dataproc_v1_jobs_proto protoreflect.FileDescriptor
View Source
var File_google_cloud_dataproc_v1_operations_proto protoreflect.FileDescriptor
View Source
var File_google_cloud_dataproc_v1_shared_proto protoreflect.FileDescriptor
View Source
var File_google_cloud_dataproc_v1_workflow_templates_proto protoreflect.FileDescriptor

Functions

func RegisterAutoscalingPolicyServiceServer

func RegisterAutoscalingPolicyServiceServer(s *grpc.Server, srv AutoscalingPolicyServiceServer)

func RegisterClusterControllerServer

func RegisterClusterControllerServer(s *grpc.Server, srv ClusterControllerServer)

func RegisterJobControllerServer

func RegisterJobControllerServer(s *grpc.Server, srv JobControllerServer)

func RegisterWorkflowTemplateServiceServer

func RegisterWorkflowTemplateServiceServer(s *grpc.Server, srv WorkflowTemplateServiceServer)

Types

type AcceleratorConfig

type AcceleratorConfig struct {

	// Full URL, partial URI, or short name of the accelerator type resource to
	// expose to this instance. See
	// [Compute Engine
	// AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes).
	//
	// Examples:
	//
	// * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80`
	// * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80`
	// * `nvidia-tesla-k80`
	//
	// **Auto Zone Exception**: If you are using the Dataproc
	// [Auto Zone
	// Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
	// feature, you must use the short name of the accelerator type
	// resource, for example, `nvidia-tesla-k80`.
	AcceleratorTypeUri string `protobuf:"bytes,1,opt,name=accelerator_type_uri,json=acceleratorTypeUri,proto3" json:"accelerator_type_uri,omitempty"`
	// The number of the accelerator cards of this type exposed to this instance.
	AcceleratorCount int32 `protobuf:"varint,2,opt,name=accelerator_count,json=acceleratorCount,proto3" json:"accelerator_count,omitempty"`
	// contains filtered or unexported fields
}

Specifies the type and number of accelerator cards attached to the instances of an instance. See [GPUs on Compute Engine](https://cloud.google.com/compute/docs/gpus/).

func (*AcceleratorConfig) Descriptor deprecated

func (*AcceleratorConfig) Descriptor() ([]byte, []int)

Deprecated: Use AcceleratorConfig.ProtoReflect.Descriptor instead.

func (*AcceleratorConfig) GetAcceleratorCount

func (x *AcceleratorConfig) GetAcceleratorCount() int32

func (*AcceleratorConfig) GetAcceleratorTypeUri

func (x *AcceleratorConfig) GetAcceleratorTypeUri() string

func (*AcceleratorConfig) ProtoMessage

func (*AcceleratorConfig) ProtoMessage()

func (*AcceleratorConfig) ProtoReflect

func (x *AcceleratorConfig) ProtoReflect() protoreflect.Message

func (*AcceleratorConfig) Reset

func (x *AcceleratorConfig) Reset()

func (*AcceleratorConfig) String

func (x *AcceleratorConfig) String() string

type AutoscalingConfig

type AutoscalingConfig struct {

	// Optional. The autoscaling policy used by the cluster.
	//
	// Only resource names including projectid and location (region) are valid.
	// Examples:
	//
	// * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
	// * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
	//
	// Note that the policy must be in the same project and Dataproc region.
	PolicyUri string `protobuf:"bytes,1,opt,name=policy_uri,json=policyUri,proto3" json:"policy_uri,omitempty"`
	// contains filtered or unexported fields
}

Autoscaling Policy config associated with the cluster.

func (*AutoscalingConfig) Descriptor deprecated

func (*AutoscalingConfig) Descriptor() ([]byte, []int)

Deprecated: Use AutoscalingConfig.ProtoReflect.Descriptor instead.

func (*AutoscalingConfig) GetPolicyUri

func (x *AutoscalingConfig) GetPolicyUri() string

func (*AutoscalingConfig) ProtoMessage

func (*AutoscalingConfig) ProtoMessage()

func (*AutoscalingConfig) ProtoReflect

func (x *AutoscalingConfig) ProtoReflect() protoreflect.Message

func (*AutoscalingConfig) Reset

func (x *AutoscalingConfig) Reset()

func (*AutoscalingConfig) String

func (x *AutoscalingConfig) String() string

type AutoscalingPolicy

type AutoscalingPolicy struct {

	// Required. The policy id.
	//
	// The id must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). Cannot begin or end with underscore
	// or hyphen. Must consist of between 3 and 50 characters.
	//
	Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
	// Output only. The "resource name" of the autoscaling policy, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.autoscalingPolicies`, the resource name of the
	//   policy has the following format:
	//   `projects/{project_id}/regions/{region}/autoscalingPolicies/{policy_id}`
	//
	// * For `projects.locations.autoscalingPolicies`, the resource name of the
	//   policy has the following format:
	//   `projects/{project_id}/locations/{location}/autoscalingPolicies/{policy_id}`
	Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"`
	// Autoscaling algorithm for policy.
	//
	// Types that are assignable to Algorithm:
	//	*AutoscalingPolicy_BasicAlgorithm
	Algorithm isAutoscalingPolicy_Algorithm `protobuf_oneof:"algorithm"`
	// Required. Describes how the autoscaler will operate for primary workers.
	WorkerConfig *InstanceGroupAutoscalingPolicyConfig `protobuf:"bytes,4,opt,name=worker_config,json=workerConfig,proto3" json:"worker_config,omitempty"`
	// Optional. Describes how the autoscaler will operate for secondary workers.
	SecondaryWorkerConfig *InstanceGroupAutoscalingPolicyConfig `` /* 126-byte string literal not displayed */
	// contains filtered or unexported fields
}

Describes an autoscaling policy for Dataproc cluster autoscaler.

func (*AutoscalingPolicy) Descriptor deprecated

func (*AutoscalingPolicy) Descriptor() ([]byte, []int)

Deprecated: Use AutoscalingPolicy.ProtoReflect.Descriptor instead.

func (*AutoscalingPolicy) GetAlgorithm

func (m *AutoscalingPolicy) GetAlgorithm() isAutoscalingPolicy_Algorithm

func (*AutoscalingPolicy) GetBasicAlgorithm

func (x *AutoscalingPolicy) GetBasicAlgorithm() *BasicAutoscalingAlgorithm

func (*AutoscalingPolicy) GetId

func (x *AutoscalingPolicy) GetId() string

func (*AutoscalingPolicy) GetName

func (x *AutoscalingPolicy) GetName() string

func (*AutoscalingPolicy) GetSecondaryWorkerConfig

func (x *AutoscalingPolicy) GetSecondaryWorkerConfig() *InstanceGroupAutoscalingPolicyConfig

func (*AutoscalingPolicy) GetWorkerConfig

func (*AutoscalingPolicy) ProtoMessage

func (*AutoscalingPolicy) ProtoMessage()

func (*AutoscalingPolicy) ProtoReflect

func (x *AutoscalingPolicy) ProtoReflect() protoreflect.Message

func (*AutoscalingPolicy) Reset

func (x *AutoscalingPolicy) Reset()

func (*AutoscalingPolicy) String

func (x *AutoscalingPolicy) String() string

type AutoscalingPolicyServiceClient

type AutoscalingPolicyServiceClient interface {
	// Creates new autoscaling policy.
	CreateAutoscalingPolicy(ctx context.Context, in *CreateAutoscalingPolicyRequest, opts ...grpc.CallOption) (*AutoscalingPolicy, error)
	// Updates (replaces) autoscaling policy.
	//
	// Disabled check for update_mask, because all updates will be full
	// replacements.
	UpdateAutoscalingPolicy(ctx context.Context, in *UpdateAutoscalingPolicyRequest, opts ...grpc.CallOption) (*AutoscalingPolicy, error)
	// Retrieves autoscaling policy.
	GetAutoscalingPolicy(ctx context.Context, in *GetAutoscalingPolicyRequest, opts ...grpc.CallOption) (*AutoscalingPolicy, error)
	// Lists autoscaling policies in the project.
	ListAutoscalingPolicies(ctx context.Context, in *ListAutoscalingPoliciesRequest, opts ...grpc.CallOption) (*ListAutoscalingPoliciesResponse, error)
	// Deletes an autoscaling policy. It is an error to delete an autoscaling
	// policy that is in use by one or more clusters.
	DeleteAutoscalingPolicy(ctx context.Context, in *DeleteAutoscalingPolicyRequest, opts ...grpc.CallOption) (*emptypb.Empty, error)
}

AutoscalingPolicyServiceClient is the client API for AutoscalingPolicyService service.

For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.

type AutoscalingPolicyServiceServer

type AutoscalingPolicyServiceServer interface {
	// Creates new autoscaling policy.
	CreateAutoscalingPolicy(context.Context, *CreateAutoscalingPolicyRequest) (*AutoscalingPolicy, error)
	// Updates (replaces) autoscaling policy.
	//
	// Disabled check for update_mask, because all updates will be full
	// replacements.
	UpdateAutoscalingPolicy(context.Context, *UpdateAutoscalingPolicyRequest) (*AutoscalingPolicy, error)
	// Retrieves autoscaling policy.
	GetAutoscalingPolicy(context.Context, *GetAutoscalingPolicyRequest) (*AutoscalingPolicy, error)
	// Lists autoscaling policies in the project.
	ListAutoscalingPolicies(context.Context, *ListAutoscalingPoliciesRequest) (*ListAutoscalingPoliciesResponse, error)
	// Deletes an autoscaling policy. It is an error to delete an autoscaling
	// policy that is in use by one or more clusters.
	DeleteAutoscalingPolicy(context.Context, *DeleteAutoscalingPolicyRequest) (*emptypb.Empty, error)
}

AutoscalingPolicyServiceServer is the server API for AutoscalingPolicyService service.

type AutoscalingPolicy_BasicAlgorithm

type AutoscalingPolicy_BasicAlgorithm struct {
	BasicAlgorithm *BasicAutoscalingAlgorithm `protobuf:"bytes,3,opt,name=basic_algorithm,json=basicAlgorithm,proto3,oneof"`
}

type BasicAutoscalingAlgorithm

type BasicAutoscalingAlgorithm struct {

	// Required. YARN autoscaling configuration.
	YarnConfig *BasicYarnAutoscalingConfig `protobuf:"bytes,1,opt,name=yarn_config,json=yarnConfig,proto3" json:"yarn_config,omitempty"`
	// Optional. Duration between scaling events. A scaling period starts after
	// the update operation from the previous event has completed.
	//
	// Bounds: [2m, 1d]. Default: 2m.
	CooldownPeriod *durationpb.Duration `protobuf:"bytes,2,opt,name=cooldown_period,json=cooldownPeriod,proto3" json:"cooldown_period,omitempty"`
	// contains filtered or unexported fields
}

Basic algorithm for autoscaling.

func (*BasicAutoscalingAlgorithm) Descriptor deprecated

func (*BasicAutoscalingAlgorithm) Descriptor() ([]byte, []int)

Deprecated: Use BasicAutoscalingAlgorithm.ProtoReflect.Descriptor instead.

func (*BasicAutoscalingAlgorithm) GetCooldownPeriod

func (x *BasicAutoscalingAlgorithm) GetCooldownPeriod() *durationpb.Duration

func (*BasicAutoscalingAlgorithm) GetYarnConfig

func (*BasicAutoscalingAlgorithm) ProtoMessage

func (*BasicAutoscalingAlgorithm) ProtoMessage()

func (*BasicAutoscalingAlgorithm) ProtoReflect

func (*BasicAutoscalingAlgorithm) Reset

func (x *BasicAutoscalingAlgorithm) Reset()

func (*BasicAutoscalingAlgorithm) String

func (x *BasicAutoscalingAlgorithm) String() string

type BasicYarnAutoscalingConfig

type BasicYarnAutoscalingConfig struct {

	// Required. Timeout for YARN graceful decommissioning of Node Managers.
	// Specifies the duration to wait for jobs to complete before forcefully
	// removing workers (and potentially interrupting jobs). Only applicable to
	// downscaling operations.
	//
	// Bounds: [0s, 1d].
	GracefulDecommissionTimeout *durationpb.Duration `` /* 144-byte string literal not displayed */
	// Required. Fraction of average YARN pending memory in the last cooldown period
	// for which to add workers. A scale-up factor of 1.0 will result in scaling
	// up so that there is no pending memory remaining after the update (more
	// aggressive scaling). A scale-up factor closer to 0 will result in a smaller
	// magnitude of scaling up (less aggressive scaling).
	// See [How autoscaling
	// works](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/autoscaling#how_autoscaling_works)
	// for more information.
	//
	// Bounds: [0.0, 1.0].
	ScaleUpFactor float64 `protobuf:"fixed64,1,opt,name=scale_up_factor,json=scaleUpFactor,proto3" json:"scale_up_factor,omitempty"`
	// Required. Fraction of average YARN pending memory in the last cooldown period
	// for which to remove workers. A scale-down factor of 1 will result in
	// scaling down so that there is no available memory remaining after the
	// update (more aggressive scaling). A scale-down factor of 0 disables
	// removing workers, which can be beneficial for autoscaling a single job.
	// See [How autoscaling
	// works](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/autoscaling#how_autoscaling_works)
	// for more information.
	//
	// Bounds: [0.0, 1.0].
	ScaleDownFactor float64 `protobuf:"fixed64,2,opt,name=scale_down_factor,json=scaleDownFactor,proto3" json:"scale_down_factor,omitempty"`
	// Optional. Minimum scale-up threshold as a fraction of total cluster size
	// before scaling occurs. For example, in a 20-worker cluster, a threshold of
	// 0.1 means the autoscaler must recommend at least a 2-worker scale-up for
	// the cluster to scale. A threshold of 0 means the autoscaler will scale up
	// on any recommended change.
	//
	// Bounds: [0.0, 1.0]. Default: 0.0.
	ScaleUpMinWorkerFraction float64 `` /* 141-byte string literal not displayed */
	// Optional. Minimum scale-down threshold as a fraction of total cluster size
	// before scaling occurs. For example, in a 20-worker cluster, a threshold of
	// 0.1 means the autoscaler must recommend at least a 2 worker scale-down for
	// the cluster to scale. A threshold of 0 means the autoscaler will scale down
	// on any recommended change.
	//
	// Bounds: [0.0, 1.0]. Default: 0.0.
	ScaleDownMinWorkerFraction float64 `` /* 147-byte string literal not displayed */
	// contains filtered or unexported fields
}

Basic autoscaling configurations for YARN.

func (*BasicYarnAutoscalingConfig) Descriptor deprecated

func (*BasicYarnAutoscalingConfig) Descriptor() ([]byte, []int)

Deprecated: Use BasicYarnAutoscalingConfig.ProtoReflect.Descriptor instead.

func (*BasicYarnAutoscalingConfig) GetGracefulDecommissionTimeout

func (x *BasicYarnAutoscalingConfig) GetGracefulDecommissionTimeout() *durationpb.Duration

func (*BasicYarnAutoscalingConfig) GetScaleDownFactor

func (x *BasicYarnAutoscalingConfig) GetScaleDownFactor() float64

func (*BasicYarnAutoscalingConfig) GetScaleDownMinWorkerFraction

func (x *BasicYarnAutoscalingConfig) GetScaleDownMinWorkerFraction() float64

func (*BasicYarnAutoscalingConfig) GetScaleUpFactor

func (x *BasicYarnAutoscalingConfig) GetScaleUpFactor() float64

func (*BasicYarnAutoscalingConfig) GetScaleUpMinWorkerFraction

func (x *BasicYarnAutoscalingConfig) GetScaleUpMinWorkerFraction() float64

func (*BasicYarnAutoscalingConfig) ProtoMessage

func (*BasicYarnAutoscalingConfig) ProtoMessage()

func (*BasicYarnAutoscalingConfig) ProtoReflect

func (*BasicYarnAutoscalingConfig) Reset

func (x *BasicYarnAutoscalingConfig) Reset()

func (*BasicYarnAutoscalingConfig) String

func (x *BasicYarnAutoscalingConfig) String() string

type CancelJobRequest

type CancelJobRequest struct {

	// Required. The ID of the Google Cloud Platform project that the job
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The job ID.
	JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// contains filtered or unexported fields
}

A request to cancel a job.

func (*CancelJobRequest) Descriptor deprecated

func (*CancelJobRequest) Descriptor() ([]byte, []int)

Deprecated: Use CancelJobRequest.ProtoReflect.Descriptor instead.

func (*CancelJobRequest) GetJobId

func (x *CancelJobRequest) GetJobId() string

func (*CancelJobRequest) GetProjectId

func (x *CancelJobRequest) GetProjectId() string

func (*CancelJobRequest) GetRegion

func (x *CancelJobRequest) GetRegion() string

func (*CancelJobRequest) ProtoMessage

func (*CancelJobRequest) ProtoMessage()

func (*CancelJobRequest) ProtoReflect

func (x *CancelJobRequest) ProtoReflect() protoreflect.Message

func (*CancelJobRequest) Reset

func (x *CancelJobRequest) Reset()

func (*CancelJobRequest) String

func (x *CancelJobRequest) String() string

type Cluster

type Cluster struct {

	// Required. The Google Cloud Platform project ID that the cluster belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The cluster name. Cluster names within a project must be
	// unique. Names of deleted clusters can be reused.
	ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Required. The cluster config. Note that Dataproc may set
	// default values, and values may change when clusters are updated.
	Config *ClusterConfig `protobuf:"bytes,3,opt,name=config,proto3" json:"config,omitempty"`
	// Optional. The labels to associate with this cluster.
	// Label **keys** must contain 1 to 63 characters, and must conform to
	// [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
	// Label **values** may be empty, but, if present, must contain 1 to 63
	// characters, and must conform to [RFC
	// 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
	// associated with a cluster.
	Labels map[string]string `` /* 153-byte string literal not displayed */
	// Output only. Cluster status.
	Status *ClusterStatus `protobuf:"bytes,4,opt,name=status,proto3" json:"status,omitempty"`
	// Output only. The previous cluster status.
	StatusHistory []*ClusterStatus `protobuf:"bytes,7,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"`
	// Output only. A cluster UUID (Unique Universal Identifier). Dataproc
	// generates this value when it creates the cluster.
	ClusterUuid string `protobuf:"bytes,6,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"`
	// Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
	//
	// **Beta Feature**: This report is available for testing purposes only. It
	// may be changed before final release.
	Metrics *ClusterMetrics `protobuf:"bytes,9,opt,name=metrics,proto3" json:"metrics,omitempty"`
	// contains filtered or unexported fields
}

Describes the identifying information, config, and status of a cluster of Compute Engine instances.

func (*Cluster) Descriptor deprecated

func (*Cluster) Descriptor() ([]byte, []int)

Deprecated: Use Cluster.ProtoReflect.Descriptor instead.

func (*Cluster) GetClusterName

func (x *Cluster) GetClusterName() string

func (*Cluster) GetClusterUuid

func (x *Cluster) GetClusterUuid() string

func (*Cluster) GetConfig

func (x *Cluster) GetConfig() *ClusterConfig

func (*Cluster) GetLabels

func (x *Cluster) GetLabels() map[string]string

func (*Cluster) GetMetrics

func (x *Cluster) GetMetrics() *ClusterMetrics

func (*Cluster) GetProjectId

func (x *Cluster) GetProjectId() string

func (*Cluster) GetStatus

func (x *Cluster) GetStatus() *ClusterStatus

func (*Cluster) GetStatusHistory

func (x *Cluster) GetStatusHistory() []*ClusterStatus

func (*Cluster) ProtoMessage

func (*Cluster) ProtoMessage()

func (*Cluster) ProtoReflect

func (x *Cluster) ProtoReflect() protoreflect.Message

func (*Cluster) Reset

func (x *Cluster) Reset()

func (*Cluster) String

func (x *Cluster) String() string

type ClusterConfig

type ClusterConfig struct {

	// Optional. A Cloud Storage bucket used to stage job
	// dependencies, config files, and job driver console output.
	// If you do not specify a staging bucket, Cloud
	// Dataproc will determine a Cloud Storage location (US,
	// ASIA, or EU) for your cluster's staging bucket according to the
	// Compute Engine zone where your cluster is deployed, and then create
	// and manage this project-level, per-location bucket (see
	// [Dataproc staging
	// bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
	ConfigBucket string `protobuf:"bytes,1,opt,name=config_bucket,json=configBucket,proto3" json:"config_bucket,omitempty"`
	// Optional. A Cloud Storage bucket used to store ephemeral cluster and jobs data,
	// such as Spark and MapReduce history files.
	// If you do not specify a temp bucket,
	// Dataproc will determine a Cloud Storage location (US,
	// ASIA, or EU) for your cluster's temp bucket according to the
	// Compute Engine zone where your cluster is deployed, and then create
	// and manage this project-level, per-location bucket. The default bucket has
	// a TTL of 90 days, but you can use any TTL (or none) if you specify a
	// bucket.
	TempBucket string `protobuf:"bytes,2,opt,name=temp_bucket,json=tempBucket,proto3" json:"temp_bucket,omitempty"`
	// Optional. The shared Compute Engine config settings for
	// all instances in a cluster.
	GceClusterConfig *GceClusterConfig `protobuf:"bytes,8,opt,name=gce_cluster_config,json=gceClusterConfig,proto3" json:"gce_cluster_config,omitempty"`
	// Optional. The Compute Engine config settings for
	// the master instance in a cluster.
	MasterConfig *InstanceGroupConfig `protobuf:"bytes,9,opt,name=master_config,json=masterConfig,proto3" json:"master_config,omitempty"`
	// Optional. The Compute Engine config settings for
	// worker instances in a cluster.
	WorkerConfig *InstanceGroupConfig `protobuf:"bytes,10,opt,name=worker_config,json=workerConfig,proto3" json:"worker_config,omitempty"`
	// Optional. The Compute Engine config settings for
	// additional worker instances in a cluster.
	SecondaryWorkerConfig *InstanceGroupConfig `` /* 127-byte string literal not displayed */
	// Optional. The config settings for software inside the cluster.
	SoftwareConfig *SoftwareConfig `protobuf:"bytes,13,opt,name=software_config,json=softwareConfig,proto3" json:"software_config,omitempty"`
	// Optional. Commands to execute on each node after config is
	// completed. By default, executables are run on master and all worker nodes.
	// You can test a node's `role` metadata to run an executable on
	// a master or worker node, as shown below using `curl` (you can also use
	// `wget`):
	//
	//     ROLE=$(curl -H Metadata-Flavor:Google
	//     http://metadata/computeMetadata/v1/instance/attributes/dataproc-role)
	//     if [[ "${ROLE}" == 'Master' ]]; then
	//       ... master specific actions ...
	//     else
	//       ... worker specific actions ...
	//     fi
	InitializationActions []*NodeInitializationAction `protobuf:"bytes,11,rep,name=initialization_actions,json=initializationActions,proto3" json:"initialization_actions,omitempty"`
	// Optional. Encryption settings for the cluster.
	EncryptionConfig *EncryptionConfig `protobuf:"bytes,15,opt,name=encryption_config,json=encryptionConfig,proto3" json:"encryption_config,omitempty"`
	// Optional. Autoscaling config for the policy associated with the cluster.
	// Cluster does not autoscale if this field is unset.
	AutoscalingConfig *AutoscalingConfig `protobuf:"bytes,18,opt,name=autoscaling_config,json=autoscalingConfig,proto3" json:"autoscaling_config,omitempty"`
	// Optional. Security settings for the cluster.
	SecurityConfig *SecurityConfig `protobuf:"bytes,16,opt,name=security_config,json=securityConfig,proto3" json:"security_config,omitempty"`
	// Optional. Lifecycle setting for the cluster.
	LifecycleConfig *LifecycleConfig `protobuf:"bytes,17,opt,name=lifecycle_config,json=lifecycleConfig,proto3" json:"lifecycle_config,omitempty"`
	// Optional. Port/endpoint configuration for this cluster
	EndpointConfig *EndpointConfig `protobuf:"bytes,19,opt,name=endpoint_config,json=endpointConfig,proto3" json:"endpoint_config,omitempty"`
	// contains filtered or unexported fields
}

The cluster config.

func (*ClusterConfig) Descriptor deprecated

func (*ClusterConfig) Descriptor() ([]byte, []int)

Deprecated: Use ClusterConfig.ProtoReflect.Descriptor instead.

func (*ClusterConfig) GetAutoscalingConfig

func (x *ClusterConfig) GetAutoscalingConfig() *AutoscalingConfig

func (*ClusterConfig) GetConfigBucket

func (x *ClusterConfig) GetConfigBucket() string

func (*ClusterConfig) GetEncryptionConfig

func (x *ClusterConfig) GetEncryptionConfig() *EncryptionConfig

func (*ClusterConfig) GetEndpointConfig

func (x *ClusterConfig) GetEndpointConfig() *EndpointConfig

func (*ClusterConfig) GetGceClusterConfig

func (x *ClusterConfig) GetGceClusterConfig() *GceClusterConfig

func (*ClusterConfig) GetInitializationActions

func (x *ClusterConfig) GetInitializationActions() []*NodeInitializationAction

func (*ClusterConfig) GetLifecycleConfig

func (x *ClusterConfig) GetLifecycleConfig() *LifecycleConfig

func (*ClusterConfig) GetMasterConfig

func (x *ClusterConfig) GetMasterConfig() *InstanceGroupConfig

func (*ClusterConfig) GetSecondaryWorkerConfig

func (x *ClusterConfig) GetSecondaryWorkerConfig() *InstanceGroupConfig

func (*ClusterConfig) GetSecurityConfig

func (x *ClusterConfig) GetSecurityConfig() *SecurityConfig

func (*ClusterConfig) GetSoftwareConfig

func (x *ClusterConfig) GetSoftwareConfig() *SoftwareConfig

func (*ClusterConfig) GetTempBucket

func (x *ClusterConfig) GetTempBucket() string

func (*ClusterConfig) GetWorkerConfig

func (x *ClusterConfig) GetWorkerConfig() *InstanceGroupConfig

func (*ClusterConfig) ProtoMessage

func (*ClusterConfig) ProtoMessage()

func (*ClusterConfig) ProtoReflect

func (x *ClusterConfig) ProtoReflect() protoreflect.Message

func (*ClusterConfig) Reset

func (x *ClusterConfig) Reset()

func (*ClusterConfig) String

func (x *ClusterConfig) String() string

type ClusterControllerClient

type ClusterControllerClient interface {
	// Creates a cluster in a project. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	CreateCluster(ctx context.Context, in *CreateClusterRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
	// Updates a cluster in a project. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	UpdateCluster(ctx context.Context, in *UpdateClusterRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
	// Deletes a cluster in a project. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	DeleteCluster(ctx context.Context, in *DeleteClusterRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
	// Gets the resource representation for a cluster in a project.
	GetCluster(ctx context.Context, in *GetClusterRequest, opts ...grpc.CallOption) (*Cluster, error)
	// Lists all regions/{region}/clusters in a project alphabetically.
	ListClusters(ctx context.Context, in *ListClustersRequest, opts ...grpc.CallOption) (*ListClustersResponse, error)
	// Gets cluster diagnostic information. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	// After the operation completes,
	// [Operation.response][google.longrunning.Operation.response]
	// contains
	// [DiagnoseClusterResults](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#diagnoseclusterresults).
	DiagnoseCluster(ctx context.Context, in *DiagnoseClusterRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
}

ClusterControllerClient is the client API for ClusterController service.

For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.

type ClusterControllerServer

type ClusterControllerServer interface {
	// Creates a cluster in a project. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	CreateCluster(context.Context, *CreateClusterRequest) (*longrunning.Operation, error)
	// Updates a cluster in a project. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	UpdateCluster(context.Context, *UpdateClusterRequest) (*longrunning.Operation, error)
	// Deletes a cluster in a project. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	DeleteCluster(context.Context, *DeleteClusterRequest) (*longrunning.Operation, error)
	// Gets the resource representation for a cluster in a project.
	GetCluster(context.Context, *GetClusterRequest) (*Cluster, error)
	// Lists all regions/{region}/clusters in a project alphabetically.
	ListClusters(context.Context, *ListClustersRequest) (*ListClustersResponse, error)
	// Gets cluster diagnostic information. The returned
	// [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata).
	// After the operation completes,
	// [Operation.response][google.longrunning.Operation.response]
	// contains
	// [DiagnoseClusterResults](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#diagnoseclusterresults).
	DiagnoseCluster(context.Context, *DiagnoseClusterRequest) (*longrunning.Operation, error)
}

ClusterControllerServer is the server API for ClusterController service.

type ClusterMetrics

type ClusterMetrics struct {

	// The HDFS metrics.
	HdfsMetrics map[string]int64 `` /* 183-byte string literal not displayed */
	// The YARN metrics.
	YarnMetrics map[string]int64 `` /* 183-byte string literal not displayed */
	// contains filtered or unexported fields
}

Contains cluster daemon metrics, such as HDFS and YARN stats.

**Beta Feature**: This report is available for testing purposes only. It may be changed before final release.

func (*ClusterMetrics) Descriptor deprecated

func (*ClusterMetrics) Descriptor() ([]byte, []int)

Deprecated: Use ClusterMetrics.ProtoReflect.Descriptor instead.

func (*ClusterMetrics) GetHdfsMetrics

func (x *ClusterMetrics) GetHdfsMetrics() map[string]int64

func (*ClusterMetrics) GetYarnMetrics

func (x *ClusterMetrics) GetYarnMetrics() map[string]int64

func (*ClusterMetrics) ProtoMessage

func (*ClusterMetrics) ProtoMessage()

func (*ClusterMetrics) ProtoReflect

func (x *ClusterMetrics) ProtoReflect() protoreflect.Message

func (*ClusterMetrics) Reset

func (x *ClusterMetrics) Reset()

func (*ClusterMetrics) String

func (x *ClusterMetrics) String() string

type ClusterOperation

type ClusterOperation struct {

	// Output only. The id of the cluster operation.
	OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"`
	// Output only. Error, if operation failed.
	Error string `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"`
	// Output only. Indicates the operation is done.
	Done bool `protobuf:"varint,3,opt,name=done,proto3" json:"done,omitempty"`
	// contains filtered or unexported fields
}

The cluster operation triggered by a workflow.

func (*ClusterOperation) Descriptor deprecated

func (*ClusterOperation) Descriptor() ([]byte, []int)

Deprecated: Use ClusterOperation.ProtoReflect.Descriptor instead.

func (*ClusterOperation) GetDone

func (x *ClusterOperation) GetDone() bool

func (*ClusterOperation) GetError

func (x *ClusterOperation) GetError() string

func (*ClusterOperation) GetOperationId

func (x *ClusterOperation) GetOperationId() string

func (*ClusterOperation) ProtoMessage

func (*ClusterOperation) ProtoMessage()

func (*ClusterOperation) ProtoReflect

func (x *ClusterOperation) ProtoReflect() protoreflect.Message

func (*ClusterOperation) Reset

func (x *ClusterOperation) Reset()

func (*ClusterOperation) String

func (x *ClusterOperation) String() string

type ClusterOperationMetadata

type ClusterOperationMetadata struct {

	// Output only. Name of the cluster for the operation.
	ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Output only. Cluster UUID for the operation.
	ClusterUuid string `protobuf:"bytes,8,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"`
	// Output only. Current operation status.
	Status *ClusterOperationStatus `protobuf:"bytes,9,opt,name=status,proto3" json:"status,omitempty"`
	// Output only. The previous operation status.
	StatusHistory []*ClusterOperationStatus `protobuf:"bytes,10,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"`
	// Output only. The operation type.
	OperationType string `protobuf:"bytes,11,opt,name=operation_type,json=operationType,proto3" json:"operation_type,omitempty"`
	// Output only. Short description of operation.
	Description string `protobuf:"bytes,12,opt,name=description,proto3" json:"description,omitempty"`
	// Output only. Labels associated with the operation
	Labels map[string]string `` /* 154-byte string literal not displayed */
	// Output only. Errors encountered during operation execution.
	Warnings []string `protobuf:"bytes,14,rep,name=warnings,proto3" json:"warnings,omitempty"`
	// contains filtered or unexported fields
}

Metadata describing the operation.

func (*ClusterOperationMetadata) Descriptor deprecated

func (*ClusterOperationMetadata) Descriptor() ([]byte, []int)

Deprecated: Use ClusterOperationMetadata.ProtoReflect.Descriptor instead.

func (*ClusterOperationMetadata) GetClusterName

func (x *ClusterOperationMetadata) GetClusterName() string

func (*ClusterOperationMetadata) GetClusterUuid

func (x *ClusterOperationMetadata) GetClusterUuid() string

func (*ClusterOperationMetadata) GetDescription

func (x *ClusterOperationMetadata) GetDescription() string

func (*ClusterOperationMetadata) GetLabels

func (x *ClusterOperationMetadata) GetLabels() map[string]string

func (*ClusterOperationMetadata) GetOperationType

func (x *ClusterOperationMetadata) GetOperationType() string

func (*ClusterOperationMetadata) GetStatus

func (*ClusterOperationMetadata) GetStatusHistory

func (x *ClusterOperationMetadata) GetStatusHistory() []*ClusterOperationStatus

func (*ClusterOperationMetadata) GetWarnings

func (x *ClusterOperationMetadata) GetWarnings() []string

func (*ClusterOperationMetadata) ProtoMessage

func (*ClusterOperationMetadata) ProtoMessage()

func (*ClusterOperationMetadata) ProtoReflect

func (x *ClusterOperationMetadata) ProtoReflect() protoreflect.Message

func (*ClusterOperationMetadata) Reset

func (x *ClusterOperationMetadata) Reset()

func (*ClusterOperationMetadata) String

func (x *ClusterOperationMetadata) String() string

type ClusterOperationStatus

type ClusterOperationStatus struct {

	// Output only. A message containing the operation state.
	State ClusterOperationStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1.ClusterOperationStatus_State" json:"state,omitempty"`
	// Output only. A message containing the detailed operation state.
	InnerState string `protobuf:"bytes,2,opt,name=inner_state,json=innerState,proto3" json:"inner_state,omitempty"`
	// Output only. A message containing any operation metadata details.
	Details string `protobuf:"bytes,3,opt,name=details,proto3" json:"details,omitempty"`
	// Output only. The time this state was entered.
	StateStartTime *timestamppb.Timestamp `protobuf:"bytes,4,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"`
	// contains filtered or unexported fields
}

The status of the operation.

func (*ClusterOperationStatus) Descriptor deprecated

func (*ClusterOperationStatus) Descriptor() ([]byte, []int)

Deprecated: Use ClusterOperationStatus.ProtoReflect.Descriptor instead.

func (*ClusterOperationStatus) GetDetails

func (x *ClusterOperationStatus) GetDetails() string

func (*ClusterOperationStatus) GetInnerState

func (x *ClusterOperationStatus) GetInnerState() string

func (*ClusterOperationStatus) GetState

func (*ClusterOperationStatus) GetStateStartTime

func (x *ClusterOperationStatus) GetStateStartTime() *timestamppb.Timestamp

func (*ClusterOperationStatus) ProtoMessage

func (*ClusterOperationStatus) ProtoMessage()

func (*ClusterOperationStatus) ProtoReflect

func (x *ClusterOperationStatus) ProtoReflect() protoreflect.Message

func (*ClusterOperationStatus) Reset

func (x *ClusterOperationStatus) Reset()

func (*ClusterOperationStatus) String

func (x *ClusterOperationStatus) String() string

type ClusterOperationStatus_State

type ClusterOperationStatus_State int32

The operation state.

const (
	// Unused.
	ClusterOperationStatus_UNKNOWN ClusterOperationStatus_State = 0
	// The operation has been created.
	ClusterOperationStatus_PENDING ClusterOperationStatus_State = 1
	// The operation is running.
	ClusterOperationStatus_RUNNING ClusterOperationStatus_State = 2
	// The operation is done; either cancelled or completed.
	ClusterOperationStatus_DONE ClusterOperationStatus_State = 3
)

func (ClusterOperationStatus_State) Descriptor

func (ClusterOperationStatus_State) Enum

func (ClusterOperationStatus_State) EnumDescriptor deprecated

func (ClusterOperationStatus_State) EnumDescriptor() ([]byte, []int)

Deprecated: Use ClusterOperationStatus_State.Descriptor instead.

func (ClusterOperationStatus_State) Number

func (ClusterOperationStatus_State) String

func (ClusterOperationStatus_State) Type

type ClusterSelector

type ClusterSelector struct {

	// Optional. The zone where workflow process executes. This parameter does not
	// affect the selection of the cluster.
	//
	// If unspecified, the zone of the first cluster matching the selector
	// is used.
	Zone string `protobuf:"bytes,1,opt,name=zone,proto3" json:"zone,omitempty"`
	// Required. The cluster labels. Cluster must have all labels
	// to match.
	ClusterLabels map[string]string `` /* 188-byte string literal not displayed */
	// contains filtered or unexported fields
}

A selector that chooses target cluster for jobs based on metadata.

func (*ClusterSelector) Descriptor deprecated

func (*ClusterSelector) Descriptor() ([]byte, []int)

Deprecated: Use ClusterSelector.ProtoReflect.Descriptor instead.

func (*ClusterSelector) GetClusterLabels

func (x *ClusterSelector) GetClusterLabels() map[string]string

func (*ClusterSelector) GetZone

func (x *ClusterSelector) GetZone() string

func (*ClusterSelector) ProtoMessage

func (*ClusterSelector) ProtoMessage()

func (*ClusterSelector) ProtoReflect

func (x *ClusterSelector) ProtoReflect() protoreflect.Message

func (*ClusterSelector) Reset

func (x *ClusterSelector) Reset()

func (*ClusterSelector) String

func (x *ClusterSelector) String() string

type ClusterStatus

type ClusterStatus struct {

	// Output only. The cluster's state.
	State ClusterStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1.ClusterStatus_State" json:"state,omitempty"`
	// Optional. Output only. Details of cluster's state.
	Detail string `protobuf:"bytes,2,opt,name=detail,proto3" json:"detail,omitempty"`
	// Output only. Time when this state was entered (see JSON representation of
	// [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
	StateStartTime *timestamppb.Timestamp `protobuf:"bytes,3,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"`
	// Output only. Additional state information that includes
	// status reported by the agent.
	Substate ClusterStatus_Substate `protobuf:"varint,4,opt,name=substate,proto3,enum=google.cloud.dataproc.v1.ClusterStatus_Substate" json:"substate,omitempty"`
	// contains filtered or unexported fields
}

The status of a cluster and its instances.

func (*ClusterStatus) Descriptor deprecated

func (*ClusterStatus) Descriptor() ([]byte, []int)

Deprecated: Use ClusterStatus.ProtoReflect.Descriptor instead.

func (*ClusterStatus) GetDetail

func (x *ClusterStatus) GetDetail() string

func (*ClusterStatus) GetState

func (x *ClusterStatus) GetState() ClusterStatus_State

func (*ClusterStatus) GetStateStartTime

func (x *ClusterStatus) GetStateStartTime() *timestamppb.Timestamp

func (*ClusterStatus) GetSubstate

func (x *ClusterStatus) GetSubstate() ClusterStatus_Substate

func (*ClusterStatus) ProtoMessage

func (*ClusterStatus) ProtoMessage()

func (*ClusterStatus) ProtoReflect

func (x *ClusterStatus) ProtoReflect() protoreflect.Message

func (*ClusterStatus) Reset

func (x *ClusterStatus) Reset()

func (*ClusterStatus) String

func (x *ClusterStatus) String() string

type ClusterStatus_State

type ClusterStatus_State int32

The cluster state.

const (
	// The cluster state is unknown.
	ClusterStatus_UNKNOWN ClusterStatus_State = 0
	// The cluster is being created and set up. It is not ready for use.
	ClusterStatus_CREATING ClusterStatus_State = 1
	// The cluster is currently running and healthy. It is ready for use.
	ClusterStatus_RUNNING ClusterStatus_State = 2
	// The cluster encountered an error. It is not ready for use.
	ClusterStatus_ERROR ClusterStatus_State = 3
	// The cluster is being deleted. It cannot be used.
	ClusterStatus_DELETING ClusterStatus_State = 4
	// The cluster is being updated. It continues to accept and process jobs.
	ClusterStatus_UPDATING ClusterStatus_State = 5
)

func (ClusterStatus_State) Descriptor

func (ClusterStatus_State) Enum

func (ClusterStatus_State) EnumDescriptor deprecated

func (ClusterStatus_State) EnumDescriptor() ([]byte, []int)

Deprecated: Use ClusterStatus_State.Descriptor instead.

func (ClusterStatus_State) Number

func (ClusterStatus_State) String

func (x ClusterStatus_State) String() string

func (ClusterStatus_State) Type

type ClusterStatus_Substate

type ClusterStatus_Substate int32

The cluster substate.

const (
	// The cluster substate is unknown.
	ClusterStatus_UNSPECIFIED ClusterStatus_Substate = 0
	// The cluster is known to be in an unhealthy state
	// (for example, critical daemons are not running or HDFS capacity is
	// exhausted).
	//
	// Applies to RUNNING state.
	ClusterStatus_UNHEALTHY ClusterStatus_Substate = 1
	// The agent-reported status is out of date (may occur if
	// Dataproc loses communication with Agent).
	//
	// Applies to RUNNING state.
	ClusterStatus_STALE_STATUS ClusterStatus_Substate = 2
)

func (ClusterStatus_Substate) Descriptor

func (ClusterStatus_Substate) Enum

func (ClusterStatus_Substate) EnumDescriptor deprecated

func (ClusterStatus_Substate) EnumDescriptor() ([]byte, []int)

Deprecated: Use ClusterStatus_Substate.Descriptor instead.

func (ClusterStatus_Substate) Number

func (ClusterStatus_Substate) String

func (x ClusterStatus_Substate) String() string

func (ClusterStatus_Substate) Type

type Component

type Component int32

Cluster components that can be activated.

const (
	// Unspecified component. Specifying this will cause Cluster creation to fail.
	Component_COMPONENT_UNSPECIFIED Component = 0
	// The Anaconda python distribution.
	Component_ANACONDA Component = 5
	// The Hive Web HCatalog (the REST service for accessing HCatalog).
	Component_HIVE_WEBHCAT Component = 3
	// The Jupyter Notebook.
	Component_JUPYTER Component = 1
	// The Presto query engine.
	Component_PRESTO Component = 6
	// The Zeppelin notebook.
	Component_ZEPPELIN Component = 4
	// The Zookeeper service.
	Component_ZOOKEEPER Component = 8
)

func (Component) Descriptor

func (Component) Descriptor() protoreflect.EnumDescriptor

func (Component) Enum

func (x Component) Enum() *Component

func (Component) EnumDescriptor deprecated

func (Component) EnumDescriptor() ([]byte, []int)

Deprecated: Use Component.Descriptor instead.

func (Component) Number

func (x Component) Number() protoreflect.EnumNumber

func (Component) String

func (x Component) String() string

func (Component) Type

type CreateAutoscalingPolicyRequest

type CreateAutoscalingPolicyRequest struct {

	// Required. The "resource name" of the region or location, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.autoscalingPolicies.create`, the resource name
	//   of the region has the following format:
	//   `projects/{project_id}/regions/{region}`
	//
	// * For `projects.locations.autoscalingPolicies.create`, the resource name
	//   of the location has the following format:
	//   `projects/{project_id}/locations/{location}`
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Required. The autoscaling policy to create.
	Policy *AutoscalingPolicy `protobuf:"bytes,2,opt,name=policy,proto3" json:"policy,omitempty"`
	// contains filtered or unexported fields
}

A request to create an autoscaling policy.

func (*CreateAutoscalingPolicyRequest) Descriptor deprecated

func (*CreateAutoscalingPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateAutoscalingPolicyRequest.ProtoReflect.Descriptor instead.

func (*CreateAutoscalingPolicyRequest) GetParent

func (x *CreateAutoscalingPolicyRequest) GetParent() string

func (*CreateAutoscalingPolicyRequest) GetPolicy

func (*CreateAutoscalingPolicyRequest) ProtoMessage

func (*CreateAutoscalingPolicyRequest) ProtoMessage()

func (*CreateAutoscalingPolicyRequest) ProtoReflect

func (*CreateAutoscalingPolicyRequest) Reset

func (x *CreateAutoscalingPolicyRequest) Reset()

func (*CreateAutoscalingPolicyRequest) String

type CreateClusterRequest

type CreateClusterRequest struct {

	// Required. The ID of the Google Cloud Platform project that the cluster
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The cluster to create.
	Cluster *Cluster `protobuf:"bytes,2,opt,name=cluster,proto3" json:"cluster,omitempty"`
	// Optional. A unique id used to identify the request. If the server
	// receives two [CreateClusterRequest][google.cloud.dataproc.v1.CreateClusterRequest] requests  with the same
	// id, then the second request will be ignored and the
	// first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the backend
	// is returned.
	//
	// It is recommended to always set this value to a
	// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
	//
	// The id must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). The maximum length is 40 characters.
	RequestId string `protobuf:"bytes,4,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"`
	// contains filtered or unexported fields
}

A request to create a cluster.

func (*CreateClusterRequest) Descriptor deprecated

func (*CreateClusterRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateClusterRequest.ProtoReflect.Descriptor instead.

func (*CreateClusterRequest) GetCluster

func (x *CreateClusterRequest) GetCluster() *Cluster

func (*CreateClusterRequest) GetProjectId

func (x *CreateClusterRequest) GetProjectId() string

func (*CreateClusterRequest) GetRegion

func (x *CreateClusterRequest) GetRegion() string

func (*CreateClusterRequest) GetRequestId

func (x *CreateClusterRequest) GetRequestId() string

func (*CreateClusterRequest) ProtoMessage

func (*CreateClusterRequest) ProtoMessage()

func (*CreateClusterRequest) ProtoReflect

func (x *CreateClusterRequest) ProtoReflect() protoreflect.Message

func (*CreateClusterRequest) Reset

func (x *CreateClusterRequest) Reset()

func (*CreateClusterRequest) String

func (x *CreateClusterRequest) String() string

type CreateWorkflowTemplateRequest

type CreateWorkflowTemplateRequest struct {

	// Required. The resource name of the region or location, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates,create`, the resource name of the
	//   region has the following format:
	//   `projects/{project_id}/regions/{region}`
	//
	// * For `projects.locations.workflowTemplates.create`, the resource name of
	//   the location has the following format:
	//   `projects/{project_id}/locations/{location}`
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Required. The Dataproc workflow template to create.
	Template *WorkflowTemplate `protobuf:"bytes,2,opt,name=template,proto3" json:"template,omitempty"`
	// contains filtered or unexported fields
}

A request to create a workflow template.

func (*CreateWorkflowTemplateRequest) Descriptor deprecated

func (*CreateWorkflowTemplateRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateWorkflowTemplateRequest.ProtoReflect.Descriptor instead.

func (*CreateWorkflowTemplateRequest) GetParent

func (x *CreateWorkflowTemplateRequest) GetParent() string

func (*CreateWorkflowTemplateRequest) GetTemplate

func (*CreateWorkflowTemplateRequest) ProtoMessage

func (*CreateWorkflowTemplateRequest) ProtoMessage()

func (*CreateWorkflowTemplateRequest) ProtoReflect

func (*CreateWorkflowTemplateRequest) Reset

func (x *CreateWorkflowTemplateRequest) Reset()

func (*CreateWorkflowTemplateRequest) String

type DeleteAutoscalingPolicyRequest

type DeleteAutoscalingPolicyRequest struct {

	// Required. The "resource name" of the autoscaling policy, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.autoscalingPolicies.delete`, the resource name
	//   of the policy has the following format:
	//   `projects/{project_id}/regions/{region}/autoscalingPolicies/{policy_id}`
	//
	// * For `projects.locations.autoscalingPolicies.delete`, the resource name
	//   of the policy has the following format:
	//   `projects/{project_id}/locations/{location}/autoscalingPolicies/{policy_id}`
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

A request to delete an autoscaling policy.

Autoscaling policies in use by one or more clusters will not be deleted.

func (*DeleteAutoscalingPolicyRequest) Descriptor deprecated

func (*DeleteAutoscalingPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteAutoscalingPolicyRequest.ProtoReflect.Descriptor instead.

func (*DeleteAutoscalingPolicyRequest) GetName

func (*DeleteAutoscalingPolicyRequest) ProtoMessage

func (*DeleteAutoscalingPolicyRequest) ProtoMessage()

func (*DeleteAutoscalingPolicyRequest) ProtoReflect

func (*DeleteAutoscalingPolicyRequest) Reset

func (x *DeleteAutoscalingPolicyRequest) Reset()

func (*DeleteAutoscalingPolicyRequest) String

type DeleteClusterRequest

type DeleteClusterRequest struct {

	// Required. The ID of the Google Cloud Platform project that the cluster
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The cluster name.
	ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Optional. Specifying the `cluster_uuid` means the RPC should fail
	// (with error NOT_FOUND) if cluster with specified UUID does not exist.
	ClusterUuid string `protobuf:"bytes,4,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"`
	// Optional. A unique id used to identify the request. If the server
	// receives two [DeleteClusterRequest][google.cloud.dataproc.v1.DeleteClusterRequest] requests  with the same
	// id, then the second request will be ignored and the
	// first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the
	// backend is returned.
	//
	// It is recommended to always set this value to a
	// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
	//
	// The id must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). The maximum length is 40 characters.
	RequestId string `protobuf:"bytes,5,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"`
	// contains filtered or unexported fields
}

A request to delete a cluster.

func (*DeleteClusterRequest) Descriptor deprecated

func (*DeleteClusterRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteClusterRequest.ProtoReflect.Descriptor instead.

func (*DeleteClusterRequest) GetClusterName

func (x *DeleteClusterRequest) GetClusterName() string

func (*DeleteClusterRequest) GetClusterUuid

func (x *DeleteClusterRequest) GetClusterUuid() string

func (*DeleteClusterRequest) GetProjectId

func (x *DeleteClusterRequest) GetProjectId() string

func (*DeleteClusterRequest) GetRegion

func (x *DeleteClusterRequest) GetRegion() string

func (*DeleteClusterRequest) GetRequestId

func (x *DeleteClusterRequest) GetRequestId() string

func (*DeleteClusterRequest) ProtoMessage

func (*DeleteClusterRequest) ProtoMessage()

func (*DeleteClusterRequest) ProtoReflect

func (x *DeleteClusterRequest) ProtoReflect() protoreflect.Message

func (*DeleteClusterRequest) Reset

func (x *DeleteClusterRequest) Reset()

func (*DeleteClusterRequest) String

func (x *DeleteClusterRequest) String() string

type DeleteJobRequest

type DeleteJobRequest struct {

	// Required. The ID of the Google Cloud Platform project that the job
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The job ID.
	JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// contains filtered or unexported fields
}

A request to delete a job.

func (*DeleteJobRequest) Descriptor deprecated

func (*DeleteJobRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteJobRequest.ProtoReflect.Descriptor instead.

func (*DeleteJobRequest) GetJobId

func (x *DeleteJobRequest) GetJobId() string

func (*DeleteJobRequest) GetProjectId

func (x *DeleteJobRequest) GetProjectId() string

func (*DeleteJobRequest) GetRegion

func (x *DeleteJobRequest) GetRegion() string

func (*DeleteJobRequest) ProtoMessage

func (*DeleteJobRequest) ProtoMessage()

func (*DeleteJobRequest) ProtoReflect

func (x *DeleteJobRequest) ProtoReflect() protoreflect.Message

func (*DeleteJobRequest) Reset

func (x *DeleteJobRequest) Reset()

func (*DeleteJobRequest) String

func (x *DeleteJobRequest) String() string

type DeleteWorkflowTemplateRequest

type DeleteWorkflowTemplateRequest struct {

	// Required. The resource name of the workflow template, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates.delete`, the resource name
	// of the template has the following format:
	//   `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}`
	//
	// * For `projects.locations.workflowTemplates.instantiate`, the resource name
	//   of the template has the following format:
	//   `projects/{project_id}/locations/{location}/workflowTemplates/{template_id}`
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// Optional. The version of workflow template to delete. If specified,
	// will only delete the template if the current server version matches
	// specified version.
	Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"`
	// contains filtered or unexported fields
}

A request to delete a workflow template.

Currently started workflows will remain running.

func (*DeleteWorkflowTemplateRequest) Descriptor deprecated

func (*DeleteWorkflowTemplateRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteWorkflowTemplateRequest.ProtoReflect.Descriptor instead.

func (*DeleteWorkflowTemplateRequest) GetName

func (*DeleteWorkflowTemplateRequest) GetVersion

func (x *DeleteWorkflowTemplateRequest) GetVersion() int32

func (*DeleteWorkflowTemplateRequest) ProtoMessage

func (*DeleteWorkflowTemplateRequest) ProtoMessage()

func (*DeleteWorkflowTemplateRequest) ProtoReflect

func (*DeleteWorkflowTemplateRequest) Reset

func (x *DeleteWorkflowTemplateRequest) Reset()

func (*DeleteWorkflowTemplateRequest) String

type DiagnoseClusterRequest

type DiagnoseClusterRequest struct {

	// Required. The ID of the Google Cloud Platform project that the cluster
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The cluster name.
	ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// contains filtered or unexported fields
}

A request to collect cluster diagnostic information.

func (*DiagnoseClusterRequest) Descriptor deprecated

func (*DiagnoseClusterRequest) Descriptor() ([]byte, []int)

Deprecated: Use DiagnoseClusterRequest.ProtoReflect.Descriptor instead.

func (*DiagnoseClusterRequest) GetClusterName

func (x *DiagnoseClusterRequest) GetClusterName() string

func (*DiagnoseClusterRequest) GetProjectId

func (x *DiagnoseClusterRequest) GetProjectId() string

func (*DiagnoseClusterRequest) GetRegion

func (x *DiagnoseClusterRequest) GetRegion() string

func (*DiagnoseClusterRequest) ProtoMessage

func (*DiagnoseClusterRequest) ProtoMessage()

func (*DiagnoseClusterRequest) ProtoReflect

func (x *DiagnoseClusterRequest) ProtoReflect() protoreflect.Message

func (*DiagnoseClusterRequest) Reset

func (x *DiagnoseClusterRequest) Reset()

func (*DiagnoseClusterRequest) String

func (x *DiagnoseClusterRequest) String() string

type DiagnoseClusterResults

type DiagnoseClusterResults struct {

	// Output only. The Cloud Storage URI of the diagnostic output.
	// The output report is a plain text file with a summary of collected
	// diagnostics.
	OutputUri string `protobuf:"bytes,1,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"`
	// contains filtered or unexported fields
}

The location of diagnostic output.

func (*DiagnoseClusterResults) Descriptor deprecated

func (*DiagnoseClusterResults) Descriptor() ([]byte, []int)

Deprecated: Use DiagnoseClusterResults.ProtoReflect.Descriptor instead.

func (*DiagnoseClusterResults) GetOutputUri

func (x *DiagnoseClusterResults) GetOutputUri() string

func (*DiagnoseClusterResults) ProtoMessage

func (*DiagnoseClusterResults) ProtoMessage()

func (*DiagnoseClusterResults) ProtoReflect

func (x *DiagnoseClusterResults) ProtoReflect() protoreflect.Message

func (*DiagnoseClusterResults) Reset

func (x *DiagnoseClusterResults) Reset()

func (*DiagnoseClusterResults) String

func (x *DiagnoseClusterResults) String() string

type DiskConfig

type DiskConfig struct {

	// Optional. Type of the boot disk (default is "pd-standard").
	// Valid values: "pd-ssd" (Persistent Disk Solid State Drive) or
	// "pd-standard" (Persistent Disk Hard Disk Drive).
	BootDiskType string `protobuf:"bytes,3,opt,name=boot_disk_type,json=bootDiskType,proto3" json:"boot_disk_type,omitempty"`
	// Optional. Size in GB of the boot disk (default is 500GB).
	BootDiskSizeGb int32 `protobuf:"varint,1,opt,name=boot_disk_size_gb,json=bootDiskSizeGb,proto3" json:"boot_disk_size_gb,omitempty"`
	// Optional. Number of attached SSDs, from 0 to 4 (default is 0).
	// If SSDs are not attached, the boot disk is used to store runtime logs and
	// [HDFS](https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data.
	// If one or more SSDs are attached, this runtime bulk
	// data is spread across them, and the boot disk contains only basic
	// config and installed binaries.
	NumLocalSsds int32 `protobuf:"varint,2,opt,name=num_local_ssds,json=numLocalSsds,proto3" json:"num_local_ssds,omitempty"`
	// contains filtered or unexported fields
}

Specifies the config of disk options for a group of VM instances.

func (*DiskConfig) Descriptor deprecated

func (*DiskConfig) Descriptor() ([]byte, []int)

Deprecated: Use DiskConfig.ProtoReflect.Descriptor instead.

func (*DiskConfig) GetBootDiskSizeGb

func (x *DiskConfig) GetBootDiskSizeGb() int32

func (*DiskConfig) GetBootDiskType

func (x *DiskConfig) GetBootDiskType() string

func (*DiskConfig) GetNumLocalSsds

func (x *DiskConfig) GetNumLocalSsds() int32

func (*DiskConfig) ProtoMessage

func (*DiskConfig) ProtoMessage()

func (*DiskConfig) ProtoReflect

func (x *DiskConfig) ProtoReflect() protoreflect.Message

func (*DiskConfig) Reset

func (x *DiskConfig) Reset()

func (*DiskConfig) String

func (x *DiskConfig) String() string

type EncryptionConfig

type EncryptionConfig struct {

	// Optional. The Cloud KMS key name to use for PD disk encryption for all
	// instances in the cluster.
	GcePdKmsKeyName string `protobuf:"bytes,1,opt,name=gce_pd_kms_key_name,json=gcePdKmsKeyName,proto3" json:"gce_pd_kms_key_name,omitempty"`
	// contains filtered or unexported fields
}

Encryption settings for the cluster.

func (*EncryptionConfig) Descriptor deprecated

func (*EncryptionConfig) Descriptor() ([]byte, []int)

Deprecated: Use EncryptionConfig.ProtoReflect.Descriptor instead.

func (*EncryptionConfig) GetGcePdKmsKeyName

func (x *EncryptionConfig) GetGcePdKmsKeyName() string

func (*EncryptionConfig) ProtoMessage

func (*EncryptionConfig) ProtoMessage()

func (*EncryptionConfig) ProtoReflect

func (x *EncryptionConfig) ProtoReflect() protoreflect.Message

func (*EncryptionConfig) Reset

func (x *EncryptionConfig) Reset()

func (*EncryptionConfig) String

func (x *EncryptionConfig) String() string

type EndpointConfig

type EndpointConfig struct {

	// Output only. The map of port descriptions to URLs. Will only be populated
	// if enable_http_port_access is true.
	HttpPorts map[string]string `` /* 176-byte string literal not displayed */
	// Optional. If true, enable http access to specific ports on the cluster
	// from external sources. Defaults to false.
	EnableHttpPortAccess bool `` /* 126-byte string literal not displayed */
	// contains filtered or unexported fields
}

Endpoint config for this cluster

func (*EndpointConfig) Descriptor deprecated

func (*EndpointConfig) Descriptor() ([]byte, []int)

Deprecated: Use EndpointConfig.ProtoReflect.Descriptor instead.

func (*EndpointConfig) GetEnableHttpPortAccess

func (x *EndpointConfig) GetEnableHttpPortAccess() bool

func (*EndpointConfig) GetHttpPorts

func (x *EndpointConfig) GetHttpPorts() map[string]string

func (*EndpointConfig) ProtoMessage

func (*EndpointConfig) ProtoMessage()

func (*EndpointConfig) ProtoReflect

func (x *EndpointConfig) ProtoReflect() protoreflect.Message

func (*EndpointConfig) Reset

func (x *EndpointConfig) Reset()

func (*EndpointConfig) String

func (x *EndpointConfig) String() string

type GceClusterConfig

type GceClusterConfig struct {

	// Optional. The zone where the Compute Engine cluster will be located.
	// On a create request, it is required in the "global" region. If omitted
	// in a non-global Dataproc region, the service will pick a zone in the
	// corresponding Compute Engine region. On a get request, zone will
	// always be present.
	//
	// A full URL, partial URI, or short name are valid. Examples:
	//
	// * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]`
	// * `projects/[project_id]/zones/[zone]`
	// * `us-central1-f`
	ZoneUri string `protobuf:"bytes,1,opt,name=zone_uri,json=zoneUri,proto3" json:"zone_uri,omitempty"`
	// Optional. The Compute Engine network to be used for machine
	// communications. Cannot be specified with subnetwork_uri. If neither
	// `network_uri` nor `subnetwork_uri` is specified, the "default" network of
	// the project is used, if it exists. Cannot be a "Custom Subnet Network" (see
	// [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for
	// more information).
	//
	// A full URL, partial URI, or short name are valid. Examples:
	//
	// * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default`
	// * `projects/[project_id]/regions/global/default`
	// * `default`
	NetworkUri string `protobuf:"bytes,2,opt,name=network_uri,json=networkUri,proto3" json:"network_uri,omitempty"`
	// Optional. The Compute Engine subnetwork to be used for machine
	// communications. Cannot be specified with network_uri.
	//
	// A full URL, partial URI, or short name are valid. Examples:
	//
	// * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/us-east1/subnetworks/sub0`
	// * `projects/[project_id]/regions/us-east1/subnetworks/sub0`
	// * `sub0`
	SubnetworkUri string `protobuf:"bytes,6,opt,name=subnetwork_uri,json=subnetworkUri,proto3" json:"subnetwork_uri,omitempty"`
	// Optional. If true, all instances in the cluster will only have internal IP
	// addresses. By default, clusters are not restricted to internal IP
	// addresses, and will have ephemeral external IP addresses assigned to each
	// instance. This `internal_ip_only` restriction can only be enabled for
	// subnetwork enabled networks, and all off-cluster dependencies must be
	// configured to be accessible without external IP addresses.
	InternalIpOnly bool `protobuf:"varint,7,opt,name=internal_ip_only,json=internalIpOnly,proto3" json:"internal_ip_only,omitempty"`
	// Optional. The [Dataproc service
	// account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_dataproc)
	// (also see [VM Data Plane
	// identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
	// used by Dataproc cluster VM instances to access Google Cloud Platform
	// services.
	//
	// If not specified, the
	// [Compute Engine default service
	// account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
	// is used.
	ServiceAccount string `protobuf:"bytes,8,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"`
	// Optional. The URIs of service account scopes to be included in
	// Compute Engine instances. The following base set of scopes is always
	// included:
	//
	// * https://www.googleapis.com/auth/cloud.useraccounts.readonly
	// * https://www.googleapis.com/auth/devstorage.read_write
	// * https://www.googleapis.com/auth/logging.write
	//
	// If no scopes are specified, the following defaults are also provided:
	//
	// * https://www.googleapis.com/auth/bigquery
	// * https://www.googleapis.com/auth/bigtable.admin.table
	// * https://www.googleapis.com/auth/bigtable.data
	// * https://www.googleapis.com/auth/devstorage.full_control
	ServiceAccountScopes []string `protobuf:"bytes,3,rep,name=service_account_scopes,json=serviceAccountScopes,proto3" json:"service_account_scopes,omitempty"`
	// The Compute Engine tags to add to all instances (see [Tagging
	// instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
	Tags []string `protobuf:"bytes,4,rep,name=tags,proto3" json:"tags,omitempty"`
	// The Compute Engine metadata entries to add to all instances (see
	// [Project and instance
	// metadata](https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)).
	Metadata map[string]string `` /* 157-byte string literal not displayed */
	// Optional. Reservation Affinity for consuming Zonal reservation.
	ReservationAffinity *ReservationAffinity `protobuf:"bytes,11,opt,name=reservation_affinity,json=reservationAffinity,proto3" json:"reservation_affinity,omitempty"`
	// contains filtered or unexported fields
}

Common config settings for resources of Compute Engine cluster instances, applicable to all instances in the cluster.

func (*GceClusterConfig) Descriptor deprecated

func (*GceClusterConfig) Descriptor() ([]byte, []int)

Deprecated: Use GceClusterConfig.ProtoReflect.Descriptor instead.

func (*GceClusterConfig) GetInternalIpOnly

func (x *GceClusterConfig) GetInternalIpOnly() bool

func (*GceClusterConfig) GetMetadata

func (x *GceClusterConfig) GetMetadata() map[string]string

func (*GceClusterConfig) GetNetworkUri

func (x *GceClusterConfig) GetNetworkUri() string

func (*GceClusterConfig) GetReservationAffinity

func (x *GceClusterConfig) GetReservationAffinity() *ReservationAffinity

func (*GceClusterConfig) GetServiceAccount

func (x *GceClusterConfig) GetServiceAccount() string

func (*GceClusterConfig) GetServiceAccountScopes

func (x *GceClusterConfig) GetServiceAccountScopes() []string

func (*GceClusterConfig) GetSubnetworkUri

func (x *GceClusterConfig) GetSubnetworkUri() string

func (*GceClusterConfig) GetTags

func (x *GceClusterConfig) GetTags() []string

func (*GceClusterConfig) GetZoneUri

func (x *GceClusterConfig) GetZoneUri() string

func (*GceClusterConfig) ProtoMessage

func (*GceClusterConfig) ProtoMessage()

func (*GceClusterConfig) ProtoReflect

func (x *GceClusterConfig) ProtoReflect() protoreflect.Message

func (*GceClusterConfig) Reset

func (x *GceClusterConfig) Reset()

func (*GceClusterConfig) String

func (x *GceClusterConfig) String() string

type GetAutoscalingPolicyRequest

type GetAutoscalingPolicyRequest struct {

	// Required. The "resource name" of the autoscaling policy, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.autoscalingPolicies.get`, the resource name
	//   of the policy has the following format:
	//   `projects/{project_id}/regions/{region}/autoscalingPolicies/{policy_id}`
	//
	// * For `projects.locations.autoscalingPolicies.get`, the resource name
	//   of the policy has the following format:
	//   `projects/{project_id}/locations/{location}/autoscalingPolicies/{policy_id}`
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

A request to fetch an autoscaling policy.

func (*GetAutoscalingPolicyRequest) Descriptor deprecated

func (*GetAutoscalingPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use GetAutoscalingPolicyRequest.ProtoReflect.Descriptor instead.

func (*GetAutoscalingPolicyRequest) GetName

func (x *GetAutoscalingPolicyRequest) GetName() string

func (*GetAutoscalingPolicyRequest) ProtoMessage

func (*GetAutoscalingPolicyRequest) ProtoMessage()

func (*GetAutoscalingPolicyRequest) ProtoReflect

func (*GetAutoscalingPolicyRequest) Reset

func (x *GetAutoscalingPolicyRequest) Reset()

func (*GetAutoscalingPolicyRequest) String

func (x *GetAutoscalingPolicyRequest) String() string

type GetClusterRequest

type GetClusterRequest struct {

	// Required. The ID of the Google Cloud Platform project that the cluster
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The cluster name.
	ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// contains filtered or unexported fields
}

Request to get the resource representation for a cluster in a project.

func (*GetClusterRequest) Descriptor deprecated

func (*GetClusterRequest) Descriptor() ([]byte, []int)

Deprecated: Use GetClusterRequest.ProtoReflect.Descriptor instead.

func (*GetClusterRequest) GetClusterName

func (x *GetClusterRequest) GetClusterName() string

func (*GetClusterRequest) GetProjectId

func (x *GetClusterRequest) GetProjectId() string

func (*GetClusterRequest) GetRegion

func (x *GetClusterRequest) GetRegion() string

func (*GetClusterRequest) ProtoMessage

func (*GetClusterRequest) ProtoMessage()

func (*GetClusterRequest) ProtoReflect

func (x *GetClusterRequest) ProtoReflect() protoreflect.Message

func (*GetClusterRequest) Reset

func (x *GetClusterRequest) Reset()

func (*GetClusterRequest) String

func (x *GetClusterRequest) String() string

type GetJobRequest

type GetJobRequest struct {

	// Required. The ID of the Google Cloud Platform project that the job
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The job ID.
	JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// contains filtered or unexported fields
}

A request to get the resource representation for a job in a project.

func (*GetJobRequest) Descriptor deprecated

func (*GetJobRequest) Descriptor() ([]byte, []int)

Deprecated: Use GetJobRequest.ProtoReflect.Descriptor instead.

func (*GetJobRequest) GetJobId

func (x *GetJobRequest) GetJobId() string

func (*GetJobRequest) GetProjectId

func (x *GetJobRequest) GetProjectId() string

func (*GetJobRequest) GetRegion

func (x *GetJobRequest) GetRegion() string

func (*GetJobRequest) ProtoMessage

func (*GetJobRequest) ProtoMessage()

func (*GetJobRequest) ProtoReflect

func (x *GetJobRequest) ProtoReflect() protoreflect.Message

func (*GetJobRequest) Reset

func (x *GetJobRequest) Reset()

func (*GetJobRequest) String

func (x *GetJobRequest) String() string

type GetWorkflowTemplateRequest

type GetWorkflowTemplateRequest struct {

	// Required. The resource name of the workflow template, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates.get`, the resource name of the
	//   template has the following format:
	//   `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}`
	//
	// * For `projects.locations.workflowTemplates.get`, the resource name of the
	//   template has the following format:
	//   `projects/{project_id}/locations/{location}/workflowTemplates/{template_id}`
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// Optional. The version of workflow template to retrieve. Only previously
	// instantiated versions can be retrieved.
	//
	// If unspecified, retrieves the current version.
	Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"`
	// contains filtered or unexported fields
}

A request to fetch a workflow template.

func (*GetWorkflowTemplateRequest) Descriptor deprecated

func (*GetWorkflowTemplateRequest) Descriptor() ([]byte, []int)

Deprecated: Use GetWorkflowTemplateRequest.ProtoReflect.Descriptor instead.

func (*GetWorkflowTemplateRequest) GetName

func (x *GetWorkflowTemplateRequest) GetName() string

func (*GetWorkflowTemplateRequest) GetVersion

func (x *GetWorkflowTemplateRequest) GetVersion() int32

func (*GetWorkflowTemplateRequest) ProtoMessage

func (*GetWorkflowTemplateRequest) ProtoMessage()

func (*GetWorkflowTemplateRequest) ProtoReflect

func (*GetWorkflowTemplateRequest) Reset

func (x *GetWorkflowTemplateRequest) Reset()

func (*GetWorkflowTemplateRequest) String

func (x *GetWorkflowTemplateRequest) String() string

type HadoopJob

type HadoopJob struct {

	// Required. Indicates the location of the driver's main class. Specify
	// either the jar file that contains the main class or the main class name.
	// To specify both, add the jar file to `jar_file_uris`, and then specify
	// the main class name in this property.
	//
	// Types that are assignable to Driver:
	//	*HadoopJob_MainJarFileUri
	//	*HadoopJob_MainClass
	Driver isHadoopJob_Driver `protobuf_oneof:"driver"`
	// Optional. The arguments to pass to the driver. Do not
	// include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as
	// job properties, since a collision may occur that causes an incorrect job
	// submission.
	Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"`
	// Optional. Jar file URIs to add to the CLASSPATHs of the
	// Hadoop driver and tasks.
	JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"`
	// Optional. HCFS (Hadoop Compatible Filesystem) URIs of files to be copied
	// to the working directory of Hadoop drivers and distributed tasks. Useful
	// for naively parallel tasks.
	FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"`
	// Optional. HCFS URIs of archives to be extracted in the working directory of
	// Hadoop drivers and tasks. Supported file types:
	// .jar, .tar, .tar.gz, .tgz, or .zip.
	ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"`
	// Optional. A mapping of property names to values, used to configure Hadoop.
	// Properties that conflict with values set by the Dataproc API may be
	// overwritten. Can include properties set in /etc/hadoop/conf/*-site and
	// classes in user code.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html).

func (*HadoopJob) Descriptor deprecated

func (*HadoopJob) Descriptor() ([]byte, []int)

Deprecated: Use HadoopJob.ProtoReflect.Descriptor instead.

func (*HadoopJob) GetArchiveUris

func (x *HadoopJob) GetArchiveUris() []string

func (*HadoopJob) GetArgs

func (x *HadoopJob) GetArgs() []string

func (*HadoopJob) GetDriver

func (m *HadoopJob) GetDriver() isHadoopJob_Driver

func (*HadoopJob) GetFileUris

func (x *HadoopJob) GetFileUris() []string

func (*HadoopJob) GetJarFileUris

func (x *HadoopJob) GetJarFileUris() []string

func (*HadoopJob) GetLoggingConfig

func (x *HadoopJob) GetLoggingConfig() *LoggingConfig

func (*HadoopJob) GetMainClass

func (x *HadoopJob) GetMainClass() string

func (*HadoopJob) GetMainJarFileUri

func (x *HadoopJob) GetMainJarFileUri() string

func (*HadoopJob) GetProperties

func (x *HadoopJob) GetProperties() map[string]string

func (*HadoopJob) ProtoMessage

func (*HadoopJob) ProtoMessage()

func (*HadoopJob) ProtoReflect

func (x *HadoopJob) ProtoReflect() protoreflect.Message

func (*HadoopJob) Reset

func (x *HadoopJob) Reset()

func (*HadoopJob) String

func (x *HadoopJob) String() string

type HadoopJob_MainClass

type HadoopJob_MainClass struct {
	// The name of the driver's main class. The jar file containing the class
	// must be in the default CLASSPATH or specified in `jar_file_uris`.
	MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,proto3,oneof"`
}

type HadoopJob_MainJarFileUri

type HadoopJob_MainJarFileUri struct {
	// The HCFS URI of the jar file containing the main class.
	// Examples:
	//     'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar'
	//     'hdfs:/tmp/test-samples/custom-wordcount.jar'
	//     'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
	MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,proto3,oneof"`
}

type HiveJob

type HiveJob struct {

	// Required. The sequence of Hive queries to execute, specified as either
	// an HCFS file URI or a list of queries.
	//
	// Types that are assignable to Queries:
	//	*HiveJob_QueryFileUri
	//	*HiveJob_QueryList
	Queries isHiveJob_Queries `protobuf_oneof:"queries"`
	// Optional. Whether to continue executing queries if a query fails.
	// The default value is `false`. Setting to `true` can be useful when
	// executing independent parallel queries.
	ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"`
	// Optional. Mapping of query variable names to values (equivalent to the
	// Hive command: `SET name="value";`).
	ScriptVariables map[string]string `` /* 194-byte string literal not displayed */
	// Optional. A mapping of property names and values, used to configure Hive.
	// Properties that conflict with values set by the Dataproc API may be
	// overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
	// /etc/hive/conf/hive-site.xml, and classes in user code.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. HCFS URIs of jar files to add to the CLASSPATH of the
	// Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes
	// and UDFs.
	JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache Hive](https://hive.apache.org/) queries on YARN.

func (*HiveJob) Descriptor deprecated

func (*HiveJob) Descriptor() ([]byte, []int)

Deprecated: Use HiveJob.ProtoReflect.Descriptor instead.

func (*HiveJob) GetContinueOnFailure

func (x *HiveJob) GetContinueOnFailure() bool

func (*HiveJob) GetJarFileUris

func (x *HiveJob) GetJarFileUris() []string

func (*HiveJob) GetProperties

func (x *HiveJob) GetProperties() map[string]string

func (*HiveJob) GetQueries

func (m *HiveJob) GetQueries() isHiveJob_Queries

func (*HiveJob) GetQueryFileUri

func (x *HiveJob) GetQueryFileUri() string

func (*HiveJob) GetQueryList

func (x *HiveJob) GetQueryList() *QueryList

func (*HiveJob) GetScriptVariables

func (x *HiveJob) GetScriptVariables() map[string]string

func (*HiveJob) ProtoMessage

func (*HiveJob) ProtoMessage()

func (*HiveJob) ProtoReflect

func (x *HiveJob) ProtoReflect() protoreflect.Message

func (*HiveJob) Reset

func (x *HiveJob) Reset()

func (*HiveJob) String

func (x *HiveJob) String() string

type HiveJob_QueryFileUri

type HiveJob_QueryFileUri struct {
	// The HCFS URI of the script that contains Hive queries.
	QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"`
}

type HiveJob_QueryList

type HiveJob_QueryList struct {
	// A list of queries.
	QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"`
}

type InstanceGroupAutoscalingPolicyConfig

type InstanceGroupAutoscalingPolicyConfig struct {

	// Optional. Minimum number of instances for this group.
	//
	// Primary workers - Bounds: [2, max_instances]. Default: 2.
	// Secondary workers - Bounds: [0, max_instances]. Default: 0.
	MinInstances int32 `protobuf:"varint,1,opt,name=min_instances,json=minInstances,proto3" json:"min_instances,omitempty"`
	// Required. Maximum number of instances for this group. Required for primary
	// workers. Note that by default, clusters will not use secondary workers.
	// Required for secondary workers if the minimum secondary instances is set.
	//
	// Primary workers - Bounds: [min_instances, ).
	// Secondary workers - Bounds: [min_instances, ). Default: 0.
	MaxInstances int32 `protobuf:"varint,2,opt,name=max_instances,json=maxInstances,proto3" json:"max_instances,omitempty"`
	// Optional. Weight for the instance group, which is used to determine the
	// fraction of total workers in the cluster from this instance group.
	// For example, if primary workers have weight 2, and secondary workers have
	// weight 1, the cluster will have approximately 2 primary workers for each
	// secondary worker.
	//
	// The cluster may not reach the specified balance if constrained
	// by min/max bounds or other autoscaling settings. For example, if
	// `max_instances` for secondary workers is 0, then only primary workers will
	// be added. The cluster can also be out of balance when created.
	//
	// If weight is not set on any instance group, the cluster will default to
	// equal weight for all groups: the cluster will attempt to maintain an equal
	// number of workers in each group within the configured size bounds for each
	// group. If weight is set for one group only, the cluster will default to
	// zero weight on the unset group. For example if weight is set only on
	// primary workers, the cluster will use primary workers only and no
	// secondary workers.
	Weight int32 `protobuf:"varint,3,opt,name=weight,proto3" json:"weight,omitempty"`
	// contains filtered or unexported fields
}

Configuration for the size bounds of an instance group, including its proportional size to other groups.

func (*InstanceGroupAutoscalingPolicyConfig) Descriptor deprecated

func (*InstanceGroupAutoscalingPolicyConfig) Descriptor() ([]byte, []int)

Deprecated: Use InstanceGroupAutoscalingPolicyConfig.ProtoReflect.Descriptor instead.

func (*InstanceGroupAutoscalingPolicyConfig) GetMaxInstances

func (x *InstanceGroupAutoscalingPolicyConfig) GetMaxInstances() int32

func (*InstanceGroupAutoscalingPolicyConfig) GetMinInstances

func (x *InstanceGroupAutoscalingPolicyConfig) GetMinInstances() int32

func (*InstanceGroupAutoscalingPolicyConfig) GetWeight

func (*InstanceGroupAutoscalingPolicyConfig) ProtoMessage

func (*InstanceGroupAutoscalingPolicyConfig) ProtoMessage()

func (*InstanceGroupAutoscalingPolicyConfig) ProtoReflect

func (*InstanceGroupAutoscalingPolicyConfig) Reset

func (*InstanceGroupAutoscalingPolicyConfig) String

type InstanceGroupConfig

type InstanceGroupConfig struct {

	// Optional. The number of VM instances in the instance group.
	// For master instance groups, must be set to 1.
	NumInstances int32 `protobuf:"varint,1,opt,name=num_instances,json=numInstances,proto3" json:"num_instances,omitempty"`
	// Output only. The list of instance names. Dataproc derives the names
	// from `cluster_name`, `num_instances`, and the instance group.
	InstanceNames []string `protobuf:"bytes,2,rep,name=instance_names,json=instanceNames,proto3" json:"instance_names,omitempty"`
	// Optional. The Compute Engine image resource used for cluster instances.
	//
	// The URI can represent an image or image family.
	//
	// Image examples:
	//
	// * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
	// * `projects/[project_id]/global/images/[image-id]`
	// * `image-id`
	//
	// Image family examples. Dataproc will use the most recent
	// image from the family:
	//
	// * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
	// * `projects/[project_id]/global/images/family/[custom-image-family-name]`
	//
	// If the URI is unspecified, it will be inferred from
	// `SoftwareConfig.image_version` or the system default.
	ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"`
	// Optional. The Compute Engine machine type used for cluster instances.
	//
	// A full URL, partial URI, or short name are valid. Examples:
	//
	// * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2`
	// * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2`
	// * `n1-standard-2`
	//
	// **Auto Zone Exception**: If you are using the Dataproc
	// [Auto Zone
	// Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
	// feature, you must use the short name of the machine type
	// resource, for example, `n1-standard-2`.
	MachineTypeUri string `protobuf:"bytes,4,opt,name=machine_type_uri,json=machineTypeUri,proto3" json:"machine_type_uri,omitempty"`
	// Optional. Disk option config settings.
	DiskConfig *DiskConfig `protobuf:"bytes,5,opt,name=disk_config,json=diskConfig,proto3" json:"disk_config,omitempty"`
	// Output only. Specifies that this instance group contains preemptible
	// instances.
	IsPreemptible bool `protobuf:"varint,6,opt,name=is_preemptible,json=isPreemptible,proto3" json:"is_preemptible,omitempty"`
	// Optional. Specifies the preemptibility of the instance group.
	//
	// The default value for master and worker groups is
	// `NON_PREEMPTIBLE`. This default cannot be changed.
	//
	// The default value for secondary instances is
	// `PREEMPTIBLE`.
	Preemptibility InstanceGroupConfig_Preemptibility `` /* 148-byte string literal not displayed */
	// Output only. The config for Compute Engine Instance Group
	// Manager that manages this group.
	// This is only used for preemptible instance groups.
	ManagedGroupConfig *ManagedGroupConfig `protobuf:"bytes,7,opt,name=managed_group_config,json=managedGroupConfig,proto3" json:"managed_group_config,omitempty"`
	// Optional. The Compute Engine accelerator configuration for these
	// instances.
	Accelerators []*AcceleratorConfig `protobuf:"bytes,8,rep,name=accelerators,proto3" json:"accelerators,omitempty"`
	// Optional. Specifies the minimum cpu platform for the Instance Group.
	// See [Dataproc -> Minimum CPU
	// Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
	MinCpuPlatform string `protobuf:"bytes,9,opt,name=min_cpu_platform,json=minCpuPlatform,proto3" json:"min_cpu_platform,omitempty"`
	// contains filtered or unexported fields
}

The config settings for Compute Engine resources in an instance group, such as a master or worker group.

func (*InstanceGroupConfig) Descriptor deprecated

func (*InstanceGroupConfig) Descriptor() ([]byte, []int)

Deprecated: Use InstanceGroupConfig.ProtoReflect.Descriptor instead.

func (*InstanceGroupConfig) GetAccelerators

func (x *InstanceGroupConfig) GetAccelerators() []*AcceleratorConfig

func (*InstanceGroupConfig) GetDiskConfig

func (x *InstanceGroupConfig) GetDiskConfig() *DiskConfig

func (*InstanceGroupConfig) GetImageUri

func (x *InstanceGroupConfig) GetImageUri() string

func (*InstanceGroupConfig) GetInstanceNames

func (x *InstanceGroupConfig) GetInstanceNames() []string

func (*InstanceGroupConfig) GetIsPreemptible

func (x *InstanceGroupConfig) GetIsPreemptible() bool

func (*InstanceGroupConfig) GetMachineTypeUri

func (x *InstanceGroupConfig) GetMachineTypeUri() string

func (*InstanceGroupConfig) GetManagedGroupConfig

func (x *InstanceGroupConfig) GetManagedGroupConfig() *ManagedGroupConfig

func (*InstanceGroupConfig) GetMinCpuPlatform

func (x *InstanceGroupConfig) GetMinCpuPlatform() string

func (*InstanceGroupConfig) GetNumInstances

func (x *InstanceGroupConfig) GetNumInstances() int32

func (*InstanceGroupConfig) GetPreemptibility

func (*InstanceGroupConfig) ProtoMessage

func (*InstanceGroupConfig) ProtoMessage()

func (*InstanceGroupConfig) ProtoReflect

func (x *InstanceGroupConfig) ProtoReflect() protoreflect.Message

func (*InstanceGroupConfig) Reset

func (x *InstanceGroupConfig) Reset()

func (*InstanceGroupConfig) String

func (x *InstanceGroupConfig) String() string

type InstanceGroupConfig_Preemptibility

type InstanceGroupConfig_Preemptibility int32

Controls the use of [preemptible instances] (https://cloud.google.com/compute/docs/instances/preemptible) within the group.

const (
	// Preemptibility is unspecified, the system will choose the
	// appropriate setting for each instance group.
	InstanceGroupConfig_PREEMPTIBILITY_UNSPECIFIED InstanceGroupConfig_Preemptibility = 0
	// Instances are non-preemptible.
	//
	// This option is allowed for all instance groups and is the only valid
	// value for Master and Worker instance groups.
	InstanceGroupConfig_NON_PREEMPTIBLE InstanceGroupConfig_Preemptibility = 1
	// Instances are preemptible.
	//
	// This option is allowed only for secondary worker groups.
	InstanceGroupConfig_PREEMPTIBLE InstanceGroupConfig_Preemptibility = 2
)

func (InstanceGroupConfig_Preemptibility) Descriptor

func (InstanceGroupConfig_Preemptibility) Enum

func (InstanceGroupConfig_Preemptibility) EnumDescriptor deprecated

func (InstanceGroupConfig_Preemptibility) EnumDescriptor() ([]byte, []int)

Deprecated: Use InstanceGroupConfig_Preemptibility.Descriptor instead.

func (InstanceGroupConfig_Preemptibility) Number

func (InstanceGroupConfig_Preemptibility) String

func (InstanceGroupConfig_Preemptibility) Type

type InstantiateInlineWorkflowTemplateRequest

type InstantiateInlineWorkflowTemplateRequest struct {

	// Required. The resource name of the region or location, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates,instantiateinline`, the resource
	//   name of the region has the following format:
	//   `projects/{project_id}/regions/{region}`
	//
	// * For `projects.locations.workflowTemplates.instantiateinline`, the
	//   resource name of the location has the following format:
	//   `projects/{project_id}/locations/{location}`
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Required. The workflow template to instantiate.
	Template *WorkflowTemplate `protobuf:"bytes,2,opt,name=template,proto3" json:"template,omitempty"`
	// Optional. A tag that prevents multiple concurrent workflow
	// instances with the same tag from running. This mitigates risk of
	// concurrent instances started due to retries.
	//
	// It is recommended to always set this value to a
	// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
	//
	// The tag must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). The maximum length is 40 characters.
	RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"`
	// contains filtered or unexported fields
}

A request to instantiate an inline workflow template.

func (*InstantiateInlineWorkflowTemplateRequest) Descriptor deprecated

func (*InstantiateInlineWorkflowTemplateRequest) Descriptor() ([]byte, []int)

Deprecated: Use InstantiateInlineWorkflowTemplateRequest.ProtoReflect.Descriptor instead.

func (*InstantiateInlineWorkflowTemplateRequest) GetParent

func (*InstantiateInlineWorkflowTemplateRequest) GetRequestId

func (*InstantiateInlineWorkflowTemplateRequest) GetTemplate

func (*InstantiateInlineWorkflowTemplateRequest) ProtoMessage

func (*InstantiateInlineWorkflowTemplateRequest) ProtoReflect

func (*InstantiateInlineWorkflowTemplateRequest) Reset

func (*InstantiateInlineWorkflowTemplateRequest) String

type InstantiateWorkflowTemplateRequest

type InstantiateWorkflowTemplateRequest struct {

	// Required. The resource name of the workflow template, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates.instantiate`, the resource name
	// of the template has the following format:
	//   `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}`
	//
	// * For `projects.locations.workflowTemplates.instantiate`, the resource name
	//   of the template has the following format:
	//   `projects/{project_id}/locations/{location}/workflowTemplates/{template_id}`
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// Optional. The version of workflow template to instantiate. If specified,
	// the workflow will be instantiated only if the current version of
	// the workflow template has the supplied version.
	//
	// This option cannot be used to instantiate a previous version of
	// workflow template.
	Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"`
	// Optional. A tag that prevents multiple concurrent workflow
	// instances with the same tag from running. This mitigates risk of
	// concurrent instances started due to retries.
	//
	// It is recommended to always set this value to a
	// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
	//
	// The tag must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). The maximum length is 40 characters.
	RequestId string `protobuf:"bytes,5,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"`
	// Optional. Map from parameter names to values that should be used for those
	// parameters. Values may not exceed 100 characters.
	Parameters map[string]string `` /* 161-byte string literal not displayed */
	// contains filtered or unexported fields
}

A request to instantiate a workflow template.

func (*InstantiateWorkflowTemplateRequest) Descriptor deprecated

func (*InstantiateWorkflowTemplateRequest) Descriptor() ([]byte, []int)

Deprecated: Use InstantiateWorkflowTemplateRequest.ProtoReflect.Descriptor instead.

func (*InstantiateWorkflowTemplateRequest) GetName

func (*InstantiateWorkflowTemplateRequest) GetParameters

func (x *InstantiateWorkflowTemplateRequest) GetParameters() map[string]string

func (*InstantiateWorkflowTemplateRequest) GetRequestId

func (x *InstantiateWorkflowTemplateRequest) GetRequestId() string

func (*InstantiateWorkflowTemplateRequest) GetVersion

func (x *InstantiateWorkflowTemplateRequest) GetVersion() int32

func (*InstantiateWorkflowTemplateRequest) ProtoMessage

func (*InstantiateWorkflowTemplateRequest) ProtoMessage()

func (*InstantiateWorkflowTemplateRequest) ProtoReflect

func (*InstantiateWorkflowTemplateRequest) Reset

func (*InstantiateWorkflowTemplateRequest) String

type Job

type Job struct {

	// Optional. The fully qualified reference to the job, which can be used to
	// obtain the equivalent REST path of the job resource. If this property
	// is not specified when a job is created, the server generates a
	// <code>job_id</code>.
	Reference *JobReference `protobuf:"bytes,1,opt,name=reference,proto3" json:"reference,omitempty"`
	// Required. Job information, including how, when, and where to
	// run the job.
	Placement *JobPlacement `protobuf:"bytes,2,opt,name=placement,proto3" json:"placement,omitempty"`
	// Required. The application/framework-specific portion of the job.
	//
	// Types that are assignable to TypeJob:
	//	*Job_HadoopJob
	//	*Job_SparkJob
	//	*Job_PysparkJob
	//	*Job_HiveJob
	//	*Job_PigJob
	//	*Job_SparkRJob
	//	*Job_SparkSqlJob
	//	*Job_PrestoJob
	TypeJob isJob_TypeJob `protobuf_oneof:"type_job"`
	// Output only. The job status. Additional application-specific
	// status information may be contained in the <code>type_job</code>
	// and <code>yarn_applications</code> fields.
	Status *JobStatus `protobuf:"bytes,8,opt,name=status,proto3" json:"status,omitempty"`
	// Output only. The previous job status.
	StatusHistory []*JobStatus `protobuf:"bytes,13,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"`
	// Output only. The collection of YARN applications spun up by this job.
	//
	// **Beta** Feature: This report is available for testing purposes only. It
	// may be changed before final release.
	YarnApplications []*YarnApplication `protobuf:"bytes,9,rep,name=yarn_applications,json=yarnApplications,proto3" json:"yarn_applications,omitempty"`
	// Output only. A URI pointing to the location of the stdout of the job's
	// driver program.
	DriverOutputResourceUri string `` /* 135-byte string literal not displayed */
	// Output only. If present, the location of miscellaneous control files
	// which may be used as part of job setup and handling. If not present,
	// control files may be placed in the same location as `driver_output_uri`.
	DriverControlFilesUri string `` /* 129-byte string literal not displayed */
	// Optional. The labels to associate with this job.
	// Label **keys** must contain 1 to 63 characters, and must conform to
	// [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
	// Label **values** may be empty, but, if present, must contain 1 to 63
	// characters, and must conform to [RFC
	// 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
	// associated with a job.
	Labels map[string]string `` /* 154-byte string literal not displayed */
	// Optional. Job scheduling configuration.
	Scheduling *JobScheduling `protobuf:"bytes,20,opt,name=scheduling,proto3" json:"scheduling,omitempty"`
	// Output only. A UUID that uniquely identifies a job within the project
	// over time. This is in contrast to a user-settable reference.job_id that
	// may be reused over time.
	JobUuid string `protobuf:"bytes,22,opt,name=job_uuid,json=jobUuid,proto3" json:"job_uuid,omitempty"`
	// Output only. Indicates whether the job is completed. If the value is `false`,
	// the job is still in progress. If `true`, the job is completed, and
	// `status.state` field will indicate if it was successful, failed,
	// or cancelled.
	Done bool `protobuf:"varint,24,opt,name=done,proto3" json:"done,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job resource.

func (*Job) Descriptor deprecated

func (*Job) Descriptor() ([]byte, []int)

Deprecated: Use Job.ProtoReflect.Descriptor instead.

func (*Job) GetDone

func (x *Job) GetDone() bool

func (*Job) GetDriverControlFilesUri

func (x *Job) GetDriverControlFilesUri() string

func (*Job) GetDriverOutputResourceUri

func (x *Job) GetDriverOutputResourceUri() string

func (*Job) GetHadoopJob

func (x *Job) GetHadoopJob() *HadoopJob

func (*Job) GetHiveJob

func (x *Job) GetHiveJob() *HiveJob

func (*Job) GetJobUuid

func (x *Job) GetJobUuid() string

func (*Job) GetLabels

func (x *Job) GetLabels() map[string]string

func (*Job) GetPigJob

func (x *Job) GetPigJob() *PigJob

func (*Job) GetPlacement

func (x *Job) GetPlacement() *JobPlacement

func (*Job) GetPrestoJob

func (x *Job) GetPrestoJob() *PrestoJob

func (*Job) GetPysparkJob

func (x *Job) GetPysparkJob() *PySparkJob

func (*Job) GetReference

func (x *Job) GetReference() *JobReference

func (*Job) GetScheduling

func (x *Job) GetScheduling() *JobScheduling

func (*Job) GetSparkJob

func (x *Job) GetSparkJob() *SparkJob

func (*Job) GetSparkRJob

func (x *Job) GetSparkRJob() *SparkRJob

func (*Job) GetSparkSqlJob

func (x *Job) GetSparkSqlJob() *SparkSqlJob

func (*Job) GetStatus

func (x *Job) GetStatus() *JobStatus

func (*Job) GetStatusHistory

func (x *Job) GetStatusHistory() []*JobStatus

func (*Job) GetTypeJob

func (m *Job) GetTypeJob() isJob_TypeJob

func (*Job) GetYarnApplications

func (x *Job) GetYarnApplications() []*YarnApplication

func (*Job) ProtoMessage

func (*Job) ProtoMessage()

func (*Job) ProtoReflect

func (x *Job) ProtoReflect() protoreflect.Message

func (*Job) Reset

func (x *Job) Reset()

func (*Job) String

func (x *Job) String() string

type JobControllerClient

type JobControllerClient interface {
	// Submits a job to a cluster.
	SubmitJob(ctx context.Context, in *SubmitJobRequest, opts ...grpc.CallOption) (*Job, error)
	// Submits job to a cluster.
	SubmitJobAsOperation(ctx context.Context, in *SubmitJobRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
	// Gets the resource representation for a job in a project.
	GetJob(ctx context.Context, in *GetJobRequest, opts ...grpc.CallOption) (*Job, error)
	// Lists regions/{region}/jobs in a project.
	ListJobs(ctx context.Context, in *ListJobsRequest, opts ...grpc.CallOption) (*ListJobsResponse, error)
	// Updates a job in a project.
	UpdateJob(ctx context.Context, in *UpdateJobRequest, opts ...grpc.CallOption) (*Job, error)
	// Starts a job cancellation request. To access the job resource
	// after cancellation, call
	// [regions/{region}/jobs.list](https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.jobs/list)
	// or
	// [regions/{region}/jobs.get](https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.jobs/get).
	CancelJob(ctx context.Context, in *CancelJobRequest, opts ...grpc.CallOption) (*Job, error)
	// Deletes the job from the project. If the job is active, the delete fails,
	// and the response returns `FAILED_PRECONDITION`.
	DeleteJob(ctx context.Context, in *DeleteJobRequest, opts ...grpc.CallOption) (*emptypb.Empty, error)
}

JobControllerClient is the client API for JobController service.

For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.

type JobControllerServer

type JobControllerServer interface {
	// Submits a job to a cluster.
	SubmitJob(context.Context, *SubmitJobRequest) (*Job, error)
	// Submits job to a cluster.
	SubmitJobAsOperation(context.Context, *SubmitJobRequest) (*longrunning.Operation, error)
	// Gets the resource representation for a job in a project.
	GetJob(context.Context, *GetJobRequest) (*Job, error)
	// Lists regions/{region}/jobs in a project.
	ListJobs(context.Context, *ListJobsRequest) (*ListJobsResponse, error)
	// Updates a job in a project.
	UpdateJob(context.Context, *UpdateJobRequest) (*Job, error)
	// Starts a job cancellation request. To access the job resource
	// after cancellation, call
	// [regions/{region}/jobs.list](https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.jobs/list)
	// or
	// [regions/{region}/jobs.get](https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.jobs/get).
	CancelJob(context.Context, *CancelJobRequest) (*Job, error)
	// Deletes the job from the project. If the job is active, the delete fails,
	// and the response returns `FAILED_PRECONDITION`.
	DeleteJob(context.Context, *DeleteJobRequest) (*emptypb.Empty, error)
}

JobControllerServer is the server API for JobController service.

type JobMetadata

type JobMetadata struct {

	// Output only. The job id.
	JobId string `protobuf:"bytes,1,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// Output only. Most recent job status.
	Status *JobStatus `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"`
	// Output only. Operation type.
	OperationType string `protobuf:"bytes,3,opt,name=operation_type,json=operationType,proto3" json:"operation_type,omitempty"`
	// Output only. Job submission time.
	StartTime *timestamppb.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
	// contains filtered or unexported fields
}

Job Operation metadata.

func (*JobMetadata) Descriptor deprecated

func (*JobMetadata) Descriptor() ([]byte, []int)

Deprecated: Use JobMetadata.ProtoReflect.Descriptor instead.

func (*JobMetadata) GetJobId

func (x *JobMetadata) GetJobId() string

func (*JobMetadata) GetOperationType

func (x *JobMetadata) GetOperationType() string

func (*JobMetadata) GetStartTime

func (x *JobMetadata) GetStartTime() *timestamppb.Timestamp

func (*JobMetadata) GetStatus

func (x *JobMetadata) GetStatus() *JobStatus

func (*JobMetadata) ProtoMessage

func (*JobMetadata) ProtoMessage()

func (*JobMetadata) ProtoReflect

func (x *JobMetadata) ProtoReflect() protoreflect.Message

func (*JobMetadata) Reset

func (x *JobMetadata) Reset()

func (*JobMetadata) String

func (x *JobMetadata) String() string

type JobPlacement

type JobPlacement struct {

	// Required. The name of the cluster where the job will be submitted.
	ClusterName string `protobuf:"bytes,1,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Output only. A cluster UUID generated by the Dataproc service when
	// the job is submitted.
	ClusterUuid string `protobuf:"bytes,2,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"`
	// contains filtered or unexported fields
}

Dataproc job config.

func (*JobPlacement) Descriptor deprecated

func (*JobPlacement) Descriptor() ([]byte, []int)

Deprecated: Use JobPlacement.ProtoReflect.Descriptor instead.

func (*JobPlacement) GetClusterName

func (x *JobPlacement) GetClusterName() string

func (*JobPlacement) GetClusterUuid

func (x *JobPlacement) GetClusterUuid() string

func (*JobPlacement) ProtoMessage

func (*JobPlacement) ProtoMessage()

func (*JobPlacement) ProtoReflect

func (x *JobPlacement) ProtoReflect() protoreflect.Message

func (*JobPlacement) Reset

func (x *JobPlacement) Reset()

func (*JobPlacement) String

func (x *JobPlacement) String() string

type JobReference

type JobReference struct {

	// Optional. The ID of the Google Cloud Platform project that the job belongs to. If
	// specified, must match the request project ID.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Optional. The job ID, which must be unique within the project.
	//
	// The ID must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), or hyphens (-). The maximum length is 100 characters.
	//
	// If not specified by the caller, the job ID will be provided by the server.
	JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// contains filtered or unexported fields
}

Encapsulates the full scoping used to reference a job.

func (*JobReference) Descriptor deprecated

func (*JobReference) Descriptor() ([]byte, []int)

Deprecated: Use JobReference.ProtoReflect.Descriptor instead.

func (*JobReference) GetJobId

func (x *JobReference) GetJobId() string

func (*JobReference) GetProjectId

func (x *JobReference) GetProjectId() string

func (*JobReference) ProtoMessage

func (*JobReference) ProtoMessage()

func (*JobReference) ProtoReflect

func (x *JobReference) ProtoReflect() protoreflect.Message

func (*JobReference) Reset

func (x *JobReference) Reset()

func (*JobReference) String

func (x *JobReference) String() string

type JobScheduling

type JobScheduling struct {

	// Optional. Maximum number of times per hour a driver may be restarted as
	// a result of driver terminating with non-zero code before job is
	// reported failed.
	//
	// A job may be reported as thrashing if driver exits with non-zero code
	// 4 times within 10 minute window.
	//
	// Maximum value is 10.
	MaxFailuresPerHour int32 `protobuf:"varint,1,opt,name=max_failures_per_hour,json=maxFailuresPerHour,proto3" json:"max_failures_per_hour,omitempty"`
	// contains filtered or unexported fields
}

Job scheduling options.

func (*JobScheduling) Descriptor deprecated

func (*JobScheduling) Descriptor() ([]byte, []int)

Deprecated: Use JobScheduling.ProtoReflect.Descriptor instead.

func (*JobScheduling) GetMaxFailuresPerHour

func (x *JobScheduling) GetMaxFailuresPerHour() int32

func (*JobScheduling) ProtoMessage

func (*JobScheduling) ProtoMessage()

func (*JobScheduling) ProtoReflect

func (x *JobScheduling) ProtoReflect() protoreflect.Message

func (*JobScheduling) Reset

func (x *JobScheduling) Reset()

func (*JobScheduling) String

func (x *JobScheduling) String() string

type JobStatus

type JobStatus struct {

	// Output only. A state message specifying the overall job state.
	State JobStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1.JobStatus_State" json:"state,omitempty"`
	// Optional. Output only. Job state details, such as an error
	// description if the state is <code>ERROR</code>.
	Details string `protobuf:"bytes,2,opt,name=details,proto3" json:"details,omitempty"`
	// Output only. The time when this state was entered.
	StateStartTime *timestamppb.Timestamp `protobuf:"bytes,6,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"`
	// Output only. Additional state information, which includes
	// status reported by the agent.
	Substate JobStatus_Substate `protobuf:"varint,7,opt,name=substate,proto3,enum=google.cloud.dataproc.v1.JobStatus_Substate" json:"substate,omitempty"`
	// contains filtered or unexported fields
}

Dataproc job status.

func (*JobStatus) Descriptor deprecated

func (*JobStatus) Descriptor() ([]byte, []int)

Deprecated: Use JobStatus.ProtoReflect.Descriptor instead.

func (*JobStatus) GetDetails

func (x *JobStatus) GetDetails() string

func (*JobStatus) GetState

func (x *JobStatus) GetState() JobStatus_State

func (*JobStatus) GetStateStartTime

func (x *JobStatus) GetStateStartTime() *timestamppb.Timestamp

func (*JobStatus) GetSubstate

func (x *JobStatus) GetSubstate() JobStatus_Substate

func (*JobStatus) ProtoMessage

func (*JobStatus) ProtoMessage()

func (*JobStatus) ProtoReflect

func (x *JobStatus) ProtoReflect() protoreflect.Message

func (*JobStatus) Reset

func (x *JobStatus) Reset()

func (*JobStatus) String

func (x *JobStatus) String() string

type JobStatus_State

type JobStatus_State int32

The job state.

const (
	// The job state is unknown.
	JobStatus_STATE_UNSPECIFIED JobStatus_State = 0
	// The job is pending; it has been submitted, but is not yet running.
	JobStatus_PENDING JobStatus_State = 1
	// Job has been received by the service and completed initial setup;
	// it will soon be submitted to the cluster.
	JobStatus_SETUP_DONE JobStatus_State = 8
	// The job is running on the cluster.
	JobStatus_RUNNING JobStatus_State = 2
	// A CancelJob request has been received, but is pending.
	JobStatus_CANCEL_PENDING JobStatus_State = 3
	// Transient in-flight resources have been canceled, and the request to
	// cancel the running job has been issued to the cluster.
	JobStatus_CANCEL_STARTED JobStatus_State = 7
	// The job cancellation was successful.
	JobStatus_CANCELLED JobStatus_State = 4
	// The job has completed successfully.
	JobStatus_DONE JobStatus_State = 5
	// The job has completed, but encountered an error.
	JobStatus_ERROR JobStatus_State = 6
	// Job attempt has failed. The detail field contains failure details for
	// this attempt.
	//
	// Applies to restartable jobs only.
	JobStatus_ATTEMPT_FAILURE JobStatus_State = 9
)

func (JobStatus_State) Descriptor

func (JobStatus_State) Enum

func (x JobStatus_State) Enum() *JobStatus_State

func (JobStatus_State) EnumDescriptor deprecated

func (JobStatus_State) EnumDescriptor() ([]byte, []int)

Deprecated: Use JobStatus_State.Descriptor instead.

func (JobStatus_State) Number

func (JobStatus_State) String

func (x JobStatus_State) String() string

func (JobStatus_State) Type

type JobStatus_Substate

type JobStatus_Substate int32

The job substate.

const (
	// The job substate is unknown.
	JobStatus_UNSPECIFIED JobStatus_Substate = 0
	// The Job is submitted to the agent.
	//
	// Applies to RUNNING state.
	JobStatus_SUBMITTED JobStatus_Substate = 1
	// The Job has been received and is awaiting execution (it may be waiting
	// for a condition to be met). See the "details" field for the reason for
	// the delay.
	//
	// Applies to RUNNING state.
	JobStatus_QUEUED JobStatus_Substate = 2
	// The agent-reported status is out of date, which may be caused by a
	// loss of communication between the agent and Dataproc. If the
	// agent does not send a timely update, the job will fail.
	//
	// Applies to RUNNING state.
	JobStatus_STALE_STATUS JobStatus_Substate = 3
)

func (JobStatus_Substate) Descriptor

func (JobStatus_Substate) Enum

func (JobStatus_Substate) EnumDescriptor deprecated

func (JobStatus_Substate) EnumDescriptor() ([]byte, []int)

Deprecated: Use JobStatus_Substate.Descriptor instead.

func (JobStatus_Substate) Number

func (JobStatus_Substate) String

func (x JobStatus_Substate) String() string

func (JobStatus_Substate) Type

type Job_HadoopJob

type Job_HadoopJob struct {
	// Optional. Job is a Hadoop job.
	HadoopJob *HadoopJob `protobuf:"bytes,3,opt,name=hadoop_job,json=hadoopJob,proto3,oneof"`
}

type Job_HiveJob

type Job_HiveJob struct {
	// Optional. Job is a Hive job.
	HiveJob *HiveJob `protobuf:"bytes,6,opt,name=hive_job,json=hiveJob,proto3,oneof"`
}

type Job_PigJob

type Job_PigJob struct {
	// Optional. Job is a Pig job.
	PigJob *PigJob `protobuf:"bytes,7,opt,name=pig_job,json=pigJob,proto3,oneof"`
}

type Job_PrestoJob

type Job_PrestoJob struct {
	// Optional. Job is a Presto job.
	PrestoJob *PrestoJob `protobuf:"bytes,23,opt,name=presto_job,json=prestoJob,proto3,oneof"`
}

type Job_PysparkJob

type Job_PysparkJob struct {
	// Optional. Job is a PySpark job.
	PysparkJob *PySparkJob `protobuf:"bytes,5,opt,name=pyspark_job,json=pysparkJob,proto3,oneof"`
}

type Job_SparkJob

type Job_SparkJob struct {
	// Optional. Job is a Spark job.
	SparkJob *SparkJob `protobuf:"bytes,4,opt,name=spark_job,json=sparkJob,proto3,oneof"`
}

type Job_SparkRJob

type Job_SparkRJob struct {
	// Optional. Job is a SparkR job.
	SparkRJob *SparkRJob `protobuf:"bytes,21,opt,name=spark_r_job,json=sparkRJob,proto3,oneof"`
}

type Job_SparkSqlJob

type Job_SparkSqlJob struct {
	// Optional. Job is a SparkSql job.
	SparkSqlJob *SparkSqlJob `protobuf:"bytes,12,opt,name=spark_sql_job,json=sparkSqlJob,proto3,oneof"`
}

type KerberosConfig

type KerberosConfig struct {

	// Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
	// this field to true to enable Kerberos on a cluster.
	EnableKerberos bool `protobuf:"varint,1,opt,name=enable_kerberos,json=enableKerberos,proto3" json:"enable_kerberos,omitempty"`
	// Required. The Cloud Storage URI of a KMS encrypted file containing the root
	// principal password.
	RootPrincipalPasswordUri string `` /* 137-byte string literal not displayed */
	// Required. The uri of the KMS key used to encrypt various sensitive
	// files.
	KmsKeyUri string `protobuf:"bytes,3,opt,name=kms_key_uri,json=kmsKeyUri,proto3" json:"kms_key_uri,omitempty"`
	// Optional. The Cloud Storage URI of the keystore file used for SSL
	// encryption. If not provided, Dataproc will provide a self-signed
	// certificate.
	KeystoreUri string `protobuf:"bytes,4,opt,name=keystore_uri,json=keystoreUri,proto3" json:"keystore_uri,omitempty"`
	// Optional. The Cloud Storage URI of the truststore file used for SSL
	// encryption. If not provided, Dataproc will provide a self-signed
	// certificate.
	TruststoreUri string `protobuf:"bytes,5,opt,name=truststore_uri,json=truststoreUri,proto3" json:"truststore_uri,omitempty"`
	// Optional. The Cloud Storage URI of a KMS encrypted file containing the
	// password to the user provided keystore. For the self-signed certificate,
	// this password is generated by Dataproc.
	KeystorePasswordUri string `protobuf:"bytes,6,opt,name=keystore_password_uri,json=keystorePasswordUri,proto3" json:"keystore_password_uri,omitempty"`
	// Optional. The Cloud Storage URI of a KMS encrypted file containing the
	// password to the user provided key. For the self-signed certificate, this
	// password is generated by Dataproc.
	KeyPasswordUri string `protobuf:"bytes,7,opt,name=key_password_uri,json=keyPasswordUri,proto3" json:"key_password_uri,omitempty"`
	// Optional. The Cloud Storage URI of a KMS encrypted file containing the
	// password to the user provided truststore. For the self-signed certificate,
	// this password is generated by Dataproc.
	TruststorePasswordUri string `` /* 126-byte string literal not displayed */
	// Optional. The remote realm the Dataproc on-cluster KDC will trust, should
	// the user enable cross realm trust.
	CrossRealmTrustRealm string `protobuf:"bytes,9,opt,name=cross_realm_trust_realm,json=crossRealmTrustRealm,proto3" json:"cross_realm_trust_realm,omitempty"`
	// Optional. The KDC (IP or hostname) for the remote trusted realm in a cross
	// realm trust relationship.
	CrossRealmTrustKdc string `protobuf:"bytes,10,opt,name=cross_realm_trust_kdc,json=crossRealmTrustKdc,proto3" json:"cross_realm_trust_kdc,omitempty"`
	// Optional. The admin server (IP or hostname) for the remote trusted realm in
	// a cross realm trust relationship.
	CrossRealmTrustAdminServer string `` /* 146-byte string literal not displayed */
	// Optional. The Cloud Storage URI of a KMS encrypted file containing the
	// shared password between the on-cluster Kerberos realm and the remote
	// trusted realm, in a cross realm trust relationship.
	CrossRealmTrustSharedPasswordUri string `` /* 166-byte string literal not displayed */
	// Optional. The Cloud Storage URI of a KMS encrypted file containing the
	// master key of the KDC database.
	KdcDbKeyUri string `protobuf:"bytes,13,opt,name=kdc_db_key_uri,json=kdcDbKeyUri,proto3" json:"kdc_db_key_uri,omitempty"`
	// Optional. The lifetime of the ticket granting ticket, in hours.
	// If not specified, or user specifies 0, then default value 10
	// will be used.
	TgtLifetimeHours int32 `protobuf:"varint,14,opt,name=tgt_lifetime_hours,json=tgtLifetimeHours,proto3" json:"tgt_lifetime_hours,omitempty"`
	// Optional. The name of the on-cluster Kerberos realm.
	// If not specified, the uppercased domain of hostnames will be the realm.
	Realm string `protobuf:"bytes,15,opt,name=realm,proto3" json:"realm,omitempty"`
	// contains filtered or unexported fields
}

Specifies Kerberos related configuration.

func (*KerberosConfig) Descriptor deprecated

func (*KerberosConfig) Descriptor() ([]byte, []int)

Deprecated: Use KerberosConfig.ProtoReflect.Descriptor instead.

func (*KerberosConfig) GetCrossRealmTrustAdminServer

func (x *KerberosConfig) GetCrossRealmTrustAdminServer() string

func (*KerberosConfig) GetCrossRealmTrustKdc

func (x *KerberosConfig) GetCrossRealmTrustKdc() string

func (*KerberosConfig) GetCrossRealmTrustRealm

func (x *KerberosConfig) GetCrossRealmTrustRealm() string

func (*KerberosConfig) GetCrossRealmTrustSharedPasswordUri

func (x *KerberosConfig) GetCrossRealmTrustSharedPasswordUri() string

func (*KerberosConfig) GetEnableKerberos

func (x *KerberosConfig) GetEnableKerberos() bool

func (*KerberosConfig) GetKdcDbKeyUri

func (x *KerberosConfig) GetKdcDbKeyUri() string

func (*KerberosConfig) GetKeyPasswordUri

func (x *KerberosConfig) GetKeyPasswordUri() string

func (*KerberosConfig) GetKeystorePasswordUri

func (x *KerberosConfig) GetKeystorePasswordUri() string

func (*KerberosConfig) GetKeystoreUri

func (x *KerberosConfig) GetKeystoreUri() string

func (*KerberosConfig) GetKmsKeyUri

func (x *KerberosConfig) GetKmsKeyUri() string

func (*KerberosConfig) GetRealm

func (x *KerberosConfig) GetRealm() string

func (*KerberosConfig) GetRootPrincipalPasswordUri

func (x *KerberosConfig) GetRootPrincipalPasswordUri() string

func (*KerberosConfig) GetTgtLifetimeHours

func (x *KerberosConfig) GetTgtLifetimeHours() int32

func (*KerberosConfig) GetTruststorePasswordUri

func (x *KerberosConfig) GetTruststorePasswordUri() string

func (*KerberosConfig) GetTruststoreUri

func (x *KerberosConfig) GetTruststoreUri() string

func (*KerberosConfig) ProtoMessage

func (*KerberosConfig) ProtoMessage()

func (*KerberosConfig) ProtoReflect

func (x *KerberosConfig) ProtoReflect() protoreflect.Message

func (*KerberosConfig) Reset

func (x *KerberosConfig) Reset()

func (*KerberosConfig) String

func (x *KerberosConfig) String() string

type LifecycleConfig

type LifecycleConfig struct {

	// Optional. The duration to keep the cluster alive while idling (when no jobs
	// are running). Passing this threshold will cause the cluster to be
	// deleted. Minimum value is 10 minutes; maximum value is 14 days (see JSON
	// representation of
	// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json).
	IdleDeleteTtl *durationpb.Duration `protobuf:"bytes,1,opt,name=idle_delete_ttl,json=idleDeleteTtl,proto3" json:"idle_delete_ttl,omitempty"`
	// Either the exact time the cluster should be deleted at or
	// the cluster maximum age.
	//
	// Types that are assignable to Ttl:
	//	*LifecycleConfig_AutoDeleteTime
	//	*LifecycleConfig_AutoDeleteTtl
	Ttl isLifecycleConfig_Ttl `protobuf_oneof:"ttl"`
	// Output only. The time when cluster became idle (most recent job finished)
	// and became eligible for deletion due to idleness (see JSON representation
	// of
	// [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
	IdleStartTime *timestamppb.Timestamp `protobuf:"bytes,4,opt,name=idle_start_time,json=idleStartTime,proto3" json:"idle_start_time,omitempty"`
	// contains filtered or unexported fields
}

Specifies the cluster auto-delete schedule configuration.

func (*LifecycleConfig) Descriptor deprecated

func (*LifecycleConfig) Descriptor() ([]byte, []int)

Deprecated: Use LifecycleConfig.ProtoReflect.Descriptor instead.

func (*LifecycleConfig) GetAutoDeleteTime

func (x *LifecycleConfig) GetAutoDeleteTime() *timestamppb.Timestamp

func (*LifecycleConfig) GetAutoDeleteTtl

func (x *LifecycleConfig) GetAutoDeleteTtl() *durationpb.Duration

func (*LifecycleConfig) GetIdleDeleteTtl

func (x *LifecycleConfig) GetIdleDeleteTtl() *durationpb.Duration

func (*LifecycleConfig) GetIdleStartTime

func (x *LifecycleConfig) GetIdleStartTime() *timestamppb.Timestamp

func (*LifecycleConfig) GetTtl

func (m *LifecycleConfig) GetTtl() isLifecycleConfig_Ttl

func (*LifecycleConfig) ProtoMessage

func (*LifecycleConfig) ProtoMessage()

func (*LifecycleConfig) ProtoReflect

func (x *LifecycleConfig) ProtoReflect() protoreflect.Message

func (*LifecycleConfig) Reset

func (x *LifecycleConfig) Reset()

func (*LifecycleConfig) String

func (x *LifecycleConfig) String() string

type LifecycleConfig_AutoDeleteTime

type LifecycleConfig_AutoDeleteTime struct {
	// Optional. The time when cluster will be auto-deleted (see JSON representation of
	// [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
	AutoDeleteTime *timestamppb.Timestamp `protobuf:"bytes,2,opt,name=auto_delete_time,json=autoDeleteTime,proto3,oneof"`
}

type LifecycleConfig_AutoDeleteTtl

type LifecycleConfig_AutoDeleteTtl struct {
	// Optional. The lifetime duration of cluster. The cluster will be
	// auto-deleted at the end of this period. Minimum value is 10 minutes;
	// maximum value is 14 days (see JSON representation of
	// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
	AutoDeleteTtl *durationpb.Duration `protobuf:"bytes,3,opt,name=auto_delete_ttl,json=autoDeleteTtl,proto3,oneof"`
}

type ListAutoscalingPoliciesRequest

type ListAutoscalingPoliciesRequest struct {

	// Required. The "resource name" of the region or location, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.autoscalingPolicies.list`, the resource name
	//   of the region has the following format:
	//   `projects/{project_id}/regions/{region}`
	//
	// * For `projects.locations.autoscalingPolicies.list`, the resource name
	//   of the location has the following format:
	//   `projects/{project_id}/locations/{location}`
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Optional. The maximum number of results to return in each response.
	// Must be less than or equal to 1000. Defaults to 100.
	PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"`
	// Optional. The page token, returned by a previous call, to request the
	// next page of results.
	PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"`
	// contains filtered or unexported fields
}

A request to list autoscaling policies in a project.

func (*ListAutoscalingPoliciesRequest) Descriptor deprecated

func (*ListAutoscalingPoliciesRequest) Descriptor() ([]byte, []int)

Deprecated: Use ListAutoscalingPoliciesRequest.ProtoReflect.Descriptor instead.

func (*ListAutoscalingPoliciesRequest) GetPageSize

func (x *ListAutoscalingPoliciesRequest) GetPageSize() int32

func (*ListAutoscalingPoliciesRequest) GetPageToken

func (x *ListAutoscalingPoliciesRequest) GetPageToken() string

func (*ListAutoscalingPoliciesRequest) GetParent

func (x *ListAutoscalingPoliciesRequest) GetParent() string

func (*ListAutoscalingPoliciesRequest) ProtoMessage

func (*ListAutoscalingPoliciesRequest) ProtoMessage()

func (*ListAutoscalingPoliciesRequest) ProtoReflect

func (*ListAutoscalingPoliciesRequest) Reset

func (x *ListAutoscalingPoliciesRequest) Reset()

func (*ListAutoscalingPoliciesRequest) String

type ListAutoscalingPoliciesResponse

type ListAutoscalingPoliciesResponse struct {

	// Output only. Autoscaling policies list.
	Policies []*AutoscalingPolicy `protobuf:"bytes,1,rep,name=policies,proto3" json:"policies,omitempty"`
	// Output only. This token is included in the response if there are more
	// results to fetch.
	NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"`
	// contains filtered or unexported fields
}

A response to a request to list autoscaling policies in a project.

func (*ListAutoscalingPoliciesResponse) Descriptor deprecated

func (*ListAutoscalingPoliciesResponse) Descriptor() ([]byte, []int)

Deprecated: Use ListAutoscalingPoliciesResponse.ProtoReflect.Descriptor instead.

func (*ListAutoscalingPoliciesResponse) GetNextPageToken

func (x *ListAutoscalingPoliciesResponse) GetNextPageToken() string

func (*ListAutoscalingPoliciesResponse) GetPolicies

func (*ListAutoscalingPoliciesResponse) ProtoMessage

func (*ListAutoscalingPoliciesResponse) ProtoMessage()

func (*ListAutoscalingPoliciesResponse) ProtoReflect

func (*ListAutoscalingPoliciesResponse) Reset

func (*ListAutoscalingPoliciesResponse) String

type ListClustersRequest

type ListClustersRequest struct {

	// Required. The ID of the Google Cloud Platform project that the cluster
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,4,opt,name=region,proto3" json:"region,omitempty"`
	// Optional. A filter constraining the clusters to list. Filters are
	// case-sensitive and have the following syntax:
	//
	// field = value [AND [field = value]] ...
	//
	// where **field** is one of `status.state`, `clusterName`, or `labels.[KEY]`,
	// and `[KEY]` is a label key. **value** can be `*` to match all values.
	// `status.state` can be one of the following: `ACTIVE`, `INACTIVE`,
	// `CREATING`, `RUNNING`, `ERROR`, `DELETING`, or `UPDATING`. `ACTIVE`
	// contains the `CREATING`, `UPDATING`, and `RUNNING` states. `INACTIVE`
	// contains the `DELETING` and `ERROR` states.
	// `clusterName` is the name of the cluster provided at creation time.
	// Only the logical `AND` operator is supported; space-separated items are
	// treated as having an implicit `AND` operator.
	//
	// Example filter:
	//
	// status.state = ACTIVE AND clusterName = mycluster
	// AND labels.env = staging AND labels.starred = *
	Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"`
	// Optional. The standard List page size.
	PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"`
	// Optional. The standard List page token.
	PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"`
	// contains filtered or unexported fields
}

A request to list the clusters in a project.

func (*ListClustersRequest) Descriptor deprecated

func (*ListClustersRequest) Descriptor() ([]byte, []int)

Deprecated: Use ListClustersRequest.ProtoReflect.Descriptor instead.

func (*ListClustersRequest) GetFilter

func (x *ListClustersRequest) GetFilter() string

func (*ListClustersRequest) GetPageSize

func (x *ListClustersRequest) GetPageSize() int32

func (*ListClustersRequest) GetPageToken

func (x *ListClustersRequest) GetPageToken() string

func (*ListClustersRequest) GetProjectId

func (x *ListClustersRequest) GetProjectId() string

func (*ListClustersRequest) GetRegion

func (x *ListClustersRequest) GetRegion() string

func (*ListClustersRequest) ProtoMessage

func (*ListClustersRequest) ProtoMessage()

func (*ListClustersRequest) ProtoReflect

func (x *ListClustersRequest) ProtoReflect() protoreflect.Message

func (*ListClustersRequest) Reset

func (x *ListClustersRequest) Reset()

func (*ListClustersRequest) String

func (x *ListClustersRequest) String() string

type ListClustersResponse

type ListClustersResponse struct {

	// Output only. The clusters in the project.
	Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"`
	// Output only. This token is included in the response if there are more
	// results to fetch. To fetch additional results, provide this value as the
	// `page_token` in a subsequent `ListClustersRequest`.
	NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"`
	// contains filtered or unexported fields
}

The list of all clusters in a project.

func (*ListClustersResponse) Descriptor deprecated

func (*ListClustersResponse) Descriptor() ([]byte, []int)

Deprecated: Use ListClustersResponse.ProtoReflect.Descriptor instead.

func (*ListClustersResponse) GetClusters

func (x *ListClustersResponse) GetClusters() []*Cluster

func (*ListClustersResponse) GetNextPageToken

func (x *ListClustersResponse) GetNextPageToken() string

func (*ListClustersResponse) ProtoMessage

func (*ListClustersResponse) ProtoMessage()

func (*ListClustersResponse) ProtoReflect

func (x *ListClustersResponse) ProtoReflect() protoreflect.Message

func (*ListClustersResponse) Reset

func (x *ListClustersResponse) Reset()

func (*ListClustersResponse) String

func (x *ListClustersResponse) String() string

type ListJobsRequest

type ListJobsRequest struct {

	// Required. The ID of the Google Cloud Platform project that the job
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,6,opt,name=region,proto3" json:"region,omitempty"`
	// Optional. The number of results to return in each response.
	PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"`
	// Optional. The page token, returned by a previous call, to request the
	// next page of results.
	PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"`
	// Optional. If set, the returned jobs list includes only jobs that were
	// submitted to the named cluster.
	ClusterName string `protobuf:"bytes,4,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Optional. Specifies enumerated categories of jobs to list.
	// (default = match ALL jobs).
	//
	// If `filter` is provided, `jobStateMatcher` will be ignored.
	JobStateMatcher ListJobsRequest_JobStateMatcher `` /* 171-byte string literal not displayed */
	// Optional. A filter constraining the jobs to list. Filters are
	// case-sensitive and have the following syntax:
	//
	// [field = value] AND [field [= value]] ...
	//
	// where **field** is `status.state` or `labels.[KEY]`, and `[KEY]` is a label
	// key. **value** can be `*` to match all values.
	// `status.state` can be either `ACTIVE` or `NON_ACTIVE`.
	// Only the logical `AND` operator is supported; space-separated items are
	// treated as having an implicit `AND` operator.
	//
	// Example filter:
	//
	// status.state = ACTIVE AND labels.env = staging AND labels.starred = *
	Filter string `protobuf:"bytes,7,opt,name=filter,proto3" json:"filter,omitempty"`
	// contains filtered or unexported fields
}

A request to list jobs in a project.

func (*ListJobsRequest) Descriptor deprecated

func (*ListJobsRequest) Descriptor() ([]byte, []int)

Deprecated: Use ListJobsRequest.ProtoReflect.Descriptor instead.

func (*ListJobsRequest) GetClusterName

func (x *ListJobsRequest) GetClusterName() string

func (*ListJobsRequest) GetFilter

func (x *ListJobsRequest) GetFilter() string

func (*ListJobsRequest) GetJobStateMatcher

func (x *ListJobsRequest) GetJobStateMatcher() ListJobsRequest_JobStateMatcher

func (*ListJobsRequest) GetPageSize

func (x *ListJobsRequest) GetPageSize() int32

func (*ListJobsRequest) GetPageToken

func (x *ListJobsRequest) GetPageToken() string

func (*ListJobsRequest) GetProjectId

func (x *ListJobsRequest) GetProjectId() string

func (*ListJobsRequest) GetRegion

func (x *ListJobsRequest) GetRegion() string

func (*ListJobsRequest) ProtoMessage

func (*ListJobsRequest) ProtoMessage()

func (*ListJobsRequest) ProtoReflect

func (x *ListJobsRequest) ProtoReflect() protoreflect.Message

func (*ListJobsRequest) Reset

func (x *ListJobsRequest) Reset()

func (*ListJobsRequest) String

func (x *ListJobsRequest) String() string

type ListJobsRequest_JobStateMatcher

type ListJobsRequest_JobStateMatcher int32

A matcher that specifies categories of job states.

const (
	// Match all jobs, regardless of state.
	ListJobsRequest_ALL ListJobsRequest_JobStateMatcher = 0
	// Only match jobs in non-terminal states: PENDING, RUNNING, or
	// CANCEL_PENDING.
	ListJobsRequest_ACTIVE ListJobsRequest_JobStateMatcher = 1
	// Only match jobs in terminal states: CANCELLED, DONE, or ERROR.
	ListJobsRequest_NON_ACTIVE ListJobsRequest_JobStateMatcher = 2
)

func (ListJobsRequest_JobStateMatcher) Descriptor

func (ListJobsRequest_JobStateMatcher) Enum

func (ListJobsRequest_JobStateMatcher) EnumDescriptor deprecated

func (ListJobsRequest_JobStateMatcher) EnumDescriptor() ([]byte, []int)

Deprecated: Use ListJobsRequest_JobStateMatcher.Descriptor instead.

func (ListJobsRequest_JobStateMatcher) Number

func (ListJobsRequest_JobStateMatcher) String

func (ListJobsRequest_JobStateMatcher) Type

type ListJobsResponse

type ListJobsResponse struct {

	// Output only. Jobs list.
	Jobs []*Job `protobuf:"bytes,1,rep,name=jobs,proto3" json:"jobs,omitempty"`
	// Optional. This token is included in the response if there are more results
	// to fetch. To fetch additional results, provide this value as the
	// `page_token` in a subsequent <code>ListJobsRequest</code>.
	NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"`
	// contains filtered or unexported fields
}

A list of jobs in a project.

func (*ListJobsResponse) Descriptor deprecated

func (*ListJobsResponse) Descriptor() ([]byte, []int)

Deprecated: Use ListJobsResponse.ProtoReflect.Descriptor instead.

func (*ListJobsResponse) GetJobs

func (x *ListJobsResponse) GetJobs() []*Job

func (*ListJobsResponse) GetNextPageToken

func (x *ListJobsResponse) GetNextPageToken() string

func (*ListJobsResponse) ProtoMessage

func (*ListJobsResponse) ProtoMessage()

func (*ListJobsResponse) ProtoReflect

func (x *ListJobsResponse) ProtoReflect() protoreflect.Message

func (*ListJobsResponse) Reset

func (x *ListJobsResponse) Reset()

func (*ListJobsResponse) String

func (x *ListJobsResponse) String() string

type ListWorkflowTemplatesRequest

type ListWorkflowTemplatesRequest struct {

	// Required. The resource name of the region or location, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates,list`, the resource
	//   name of the region has the following format:
	//   `projects/{project_id}/regions/{region}`
	//
	// * For `projects.locations.workflowTemplates.list`, the
	//   resource name of the location has the following format:
	//   `projects/{project_id}/locations/{location}`
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Optional. The maximum number of results to return in each response.
	PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"`
	// Optional. The page token, returned by a previous call, to request the
	// next page of results.
	PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"`
	// contains filtered or unexported fields
}

A request to list workflow templates in a project.

func (*ListWorkflowTemplatesRequest) Descriptor deprecated

func (*ListWorkflowTemplatesRequest) Descriptor() ([]byte, []int)

Deprecated: Use ListWorkflowTemplatesRequest.ProtoReflect.Descriptor instead.

func (*ListWorkflowTemplatesRequest) GetPageSize

func (x *ListWorkflowTemplatesRequest) GetPageSize() int32

func (*ListWorkflowTemplatesRequest) GetPageToken

func (x *ListWorkflowTemplatesRequest) GetPageToken() string

func (*ListWorkflowTemplatesRequest) GetParent

func (x *ListWorkflowTemplatesRequest) GetParent() string

func (*ListWorkflowTemplatesRequest) ProtoMessage

func (*ListWorkflowTemplatesRequest) ProtoMessage()

func (*ListWorkflowTemplatesRequest) ProtoReflect

func (*ListWorkflowTemplatesRequest) Reset

func (x *ListWorkflowTemplatesRequest) Reset()

func (*ListWorkflowTemplatesRequest) String

type ListWorkflowTemplatesResponse

type ListWorkflowTemplatesResponse struct {

	// Output only. WorkflowTemplates list.
	Templates []*WorkflowTemplate `protobuf:"bytes,1,rep,name=templates,proto3" json:"templates,omitempty"`
	// Output only. This token is included in the response if there are more
	// results to fetch. To fetch additional results, provide this value as the
	// page_token in a subsequent <code>ListWorkflowTemplatesRequest</code>.
	NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"`
	// contains filtered or unexported fields
}

A response to a request to list workflow templates in a project.

func (*ListWorkflowTemplatesResponse) Descriptor deprecated

func (*ListWorkflowTemplatesResponse) Descriptor() ([]byte, []int)

Deprecated: Use ListWorkflowTemplatesResponse.ProtoReflect.Descriptor instead.

func (*ListWorkflowTemplatesResponse) GetNextPageToken

func (x *ListWorkflowTemplatesResponse) GetNextPageToken() string

func (*ListWorkflowTemplatesResponse) GetTemplates

func (x *ListWorkflowTemplatesResponse) GetTemplates() []*WorkflowTemplate

func (*ListWorkflowTemplatesResponse) ProtoMessage

func (*ListWorkflowTemplatesResponse) ProtoMessage()

func (*ListWorkflowTemplatesResponse) ProtoReflect

func (*ListWorkflowTemplatesResponse) Reset

func (x *ListWorkflowTemplatesResponse) Reset()

func (*ListWorkflowTemplatesResponse) String

type LoggingConfig

type LoggingConfig struct {

	// The per-package log levels for the driver. This may include
	// "root" package name to configure rootLogger.
	// Examples:
	//   'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
	DriverLogLevels map[string]LoggingConfig_Level `` /* 247-byte string literal not displayed */
	// contains filtered or unexported fields
}

The runtime logging config of the job.

func (*LoggingConfig) Descriptor deprecated

func (*LoggingConfig) Descriptor() ([]byte, []int)

Deprecated: Use LoggingConfig.ProtoReflect.Descriptor instead.

func (*LoggingConfig) GetDriverLogLevels

func (x *LoggingConfig) GetDriverLogLevels() map[string]LoggingConfig_Level

func (*LoggingConfig) ProtoMessage

func (*LoggingConfig) ProtoMessage()

func (*LoggingConfig) ProtoReflect

func (x *LoggingConfig) ProtoReflect() protoreflect.Message

func (*LoggingConfig) Reset

func (x *LoggingConfig) Reset()

func (*LoggingConfig) String

func (x *LoggingConfig) String() string

type LoggingConfig_Level

type LoggingConfig_Level int32

The Log4j level for job execution. When running an [Apache Hive](http://hive.apache.org/) job, Cloud Dataproc configures the Hive client to an equivalent verbosity level.

const (
	// Level is unspecified. Use default level for log4j.
	LoggingConfig_LEVEL_UNSPECIFIED LoggingConfig_Level = 0
	// Use ALL level for log4j.
	LoggingConfig_ALL LoggingConfig_Level = 1
	// Use TRACE level for log4j.
	LoggingConfig_TRACE LoggingConfig_Level = 2
	// Use DEBUG level for log4j.
	LoggingConfig_DEBUG LoggingConfig_Level = 3
	// Use INFO level for log4j.
	LoggingConfig_INFO LoggingConfig_Level = 4
	// Use WARN level for log4j.
	LoggingConfig_WARN LoggingConfig_Level = 5
	// Use ERROR level for log4j.
	LoggingConfig_ERROR LoggingConfig_Level = 6
	// Use FATAL level for log4j.
	LoggingConfig_FATAL LoggingConfig_Level = 7
	// Turn off log4j.
	LoggingConfig_OFF LoggingConfig_Level = 8
)

func (LoggingConfig_Level) Descriptor

func (LoggingConfig_Level) Enum

func (LoggingConfig_Level) EnumDescriptor deprecated

func (LoggingConfig_Level) EnumDescriptor() ([]byte, []int)

Deprecated: Use LoggingConfig_Level.Descriptor instead.

func (LoggingConfig_Level) Number

func (LoggingConfig_Level) String

func (x LoggingConfig_Level) String() string

func (LoggingConfig_Level) Type

type ManagedCluster

type ManagedCluster struct {

	// Required. The cluster name prefix. A unique cluster name will be formed by
	// appending a random suffix.
	//
	// The name must contain only lower-case letters (a-z), numbers (0-9),
	// and hyphens (-). Must begin with a letter. Cannot begin or end with
	// hyphen. Must consist of between 2 and 35 characters.
	ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Required. The cluster configuration.
	Config *ClusterConfig `protobuf:"bytes,3,opt,name=config,proto3" json:"config,omitempty"`
	// Optional. The labels to associate with this cluster.
	//
	// Label keys must be between 1 and 63 characters long, and must conform to
	// the following PCRE regular expression:
	// [\p{Ll}\p{Lo}][\p{Ll}\p{Lo}\p{N}_-]{0,62}
	//
	// Label values must be between 1 and 63 characters long, and must conform to
	// the following PCRE regular expression: [\p{Ll}\p{Lo}\p{N}_-]{0,63}
	//
	// No more than 32 labels can be associated with a given cluster.
	Labels map[string]string `` /* 153-byte string literal not displayed */
	// contains filtered or unexported fields
}

Cluster that is managed by the workflow.

func (*ManagedCluster) Descriptor deprecated

func (*ManagedCluster) Descriptor() ([]byte, []int)

Deprecated: Use ManagedCluster.ProtoReflect.Descriptor instead.

func (*ManagedCluster) GetClusterName

func (x *ManagedCluster) GetClusterName() string

func (*ManagedCluster) GetConfig

func (x *ManagedCluster) GetConfig() *ClusterConfig

func (*ManagedCluster) GetLabels

func (x *ManagedCluster) GetLabels() map[string]string

func (*ManagedCluster) ProtoMessage

func (*ManagedCluster) ProtoMessage()

func (*ManagedCluster) ProtoReflect

func (x *ManagedCluster) ProtoReflect() protoreflect.Message

func (*ManagedCluster) Reset

func (x *ManagedCluster) Reset()

func (*ManagedCluster) String

func (x *ManagedCluster) String() string

type ManagedGroupConfig

type ManagedGroupConfig struct {

	// Output only. The name of the Instance Template used for the Managed
	// Instance Group.
	InstanceTemplateName string `protobuf:"bytes,1,opt,name=instance_template_name,json=instanceTemplateName,proto3" json:"instance_template_name,omitempty"`
	// Output only. The name of the Instance Group Manager for this group.
	InstanceGroupManagerName string `` /* 137-byte string literal not displayed */
	// contains filtered or unexported fields
}

Specifies the resources used to actively manage an instance group.

func (*ManagedGroupConfig) Descriptor deprecated

func (*ManagedGroupConfig) Descriptor() ([]byte, []int)

Deprecated: Use ManagedGroupConfig.ProtoReflect.Descriptor instead.

func (*ManagedGroupConfig) GetInstanceGroupManagerName

func (x *ManagedGroupConfig) GetInstanceGroupManagerName() string

func (*ManagedGroupConfig) GetInstanceTemplateName

func (x *ManagedGroupConfig) GetInstanceTemplateName() string

func (*ManagedGroupConfig) ProtoMessage

func (*ManagedGroupConfig) ProtoMessage()

func (*ManagedGroupConfig) ProtoReflect

func (x *ManagedGroupConfig) ProtoReflect() protoreflect.Message

func (*ManagedGroupConfig) Reset

func (x *ManagedGroupConfig) Reset()

func (*ManagedGroupConfig) String

func (x *ManagedGroupConfig) String() string

type NodeInitializationAction

type NodeInitializationAction struct {

	// Required. Cloud Storage URI of executable file.
	ExecutableFile string `protobuf:"bytes,1,opt,name=executable_file,json=executableFile,proto3" json:"executable_file,omitempty"`
	// Optional. Amount of time executable has to complete. Default is
	// 10 minutes (see JSON representation of
	// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
	//
	// Cluster creation fails with an explanatory error message (the
	// name of the executable that caused the error and the exceeded timeout
	// period) if the executable is not completed at end of the timeout period.
	ExecutionTimeout *durationpb.Duration `protobuf:"bytes,2,opt,name=execution_timeout,json=executionTimeout,proto3" json:"execution_timeout,omitempty"`
	// contains filtered or unexported fields
}

Specifies an executable to run on a fully configured node and a timeout period for executable completion.

func (*NodeInitializationAction) Descriptor deprecated

func (*NodeInitializationAction) Descriptor() ([]byte, []int)

Deprecated: Use NodeInitializationAction.ProtoReflect.Descriptor instead.

func (*NodeInitializationAction) GetExecutableFile

func (x *NodeInitializationAction) GetExecutableFile() string

func (*NodeInitializationAction) GetExecutionTimeout

func (x *NodeInitializationAction) GetExecutionTimeout() *durationpb.Duration

func (*NodeInitializationAction) ProtoMessage

func (*NodeInitializationAction) ProtoMessage()

func (*NodeInitializationAction) ProtoReflect

func (x *NodeInitializationAction) ProtoReflect() protoreflect.Message

func (*NodeInitializationAction) Reset

func (x *NodeInitializationAction) Reset()

func (*NodeInitializationAction) String

func (x *NodeInitializationAction) String() string

type OrderedJob

type OrderedJob struct {

	// Required. The step id. The id must be unique among all jobs
	// within the template.
	//
	// The step id is used as prefix for job id, as job
	// `goog-dataproc-workflow-step-id` label, and in
	// [prerequisiteStepIds][google.cloud.dataproc.v1.OrderedJob.prerequisite_step_ids] field from other
	// steps.
	//
	// The id must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). Cannot begin or end with underscore
	// or hyphen. Must consist of between 3 and 50 characters.
	StepId string `protobuf:"bytes,1,opt,name=step_id,json=stepId,proto3" json:"step_id,omitempty"`
	// Required. The job definition.
	//
	// Types that are assignable to JobType:
	//	*OrderedJob_HadoopJob
	//	*OrderedJob_SparkJob
	//	*OrderedJob_PysparkJob
	//	*OrderedJob_HiveJob
	//	*OrderedJob_PigJob
	//	*OrderedJob_SparkRJob
	//	*OrderedJob_SparkSqlJob
	//	*OrderedJob_PrestoJob
	JobType isOrderedJob_JobType `protobuf_oneof:"job_type"`
	// Optional. The labels to associate with this job.
	//
	// Label keys must be between 1 and 63 characters long, and must conform to
	// the following regular expression:
	// [\p{Ll}\p{Lo}][\p{Ll}\p{Lo}\p{N}_-]{0,62}
	//
	// Label values must be between 1 and 63 characters long, and must conform to
	// the following regular expression: [\p{Ll}\p{Lo}\p{N}_-]{0,63}
	//
	// No more than 32 labels can be associated with a given job.
	Labels map[string]string `` /* 153-byte string literal not displayed */
	// Optional. Job scheduling configuration.
	Scheduling *JobScheduling `protobuf:"bytes,9,opt,name=scheduling,proto3" json:"scheduling,omitempty"`
	// Optional. The optional list of prerequisite job step_ids.
	// If not specified, the job will start at the beginning of workflow.
	PrerequisiteStepIds []string `protobuf:"bytes,10,rep,name=prerequisite_step_ids,json=prerequisiteStepIds,proto3" json:"prerequisite_step_ids,omitempty"`
	// contains filtered or unexported fields
}

A job executed by the workflow.

func (*OrderedJob) Descriptor deprecated

func (*OrderedJob) Descriptor() ([]byte, []int)

Deprecated: Use OrderedJob.ProtoReflect.Descriptor instead.

func (*OrderedJob) GetHadoopJob

func (x *OrderedJob) GetHadoopJob() *HadoopJob

func (*OrderedJob) GetHiveJob

func (x *OrderedJob) GetHiveJob() *HiveJob

func (*OrderedJob) GetJobType

func (m *OrderedJob) GetJobType() isOrderedJob_JobType

func (*OrderedJob) GetLabels

func (x *OrderedJob) GetLabels() map[string]string

func (*OrderedJob) GetPigJob

func (x *OrderedJob) GetPigJob() *PigJob

func (*OrderedJob) GetPrerequisiteStepIds

func (x *OrderedJob) GetPrerequisiteStepIds() []string

func (*OrderedJob) GetPrestoJob

func (x *OrderedJob) GetPrestoJob() *PrestoJob

func (*OrderedJob) GetPysparkJob

func (x *OrderedJob) GetPysparkJob() *PySparkJob

func (*OrderedJob) GetScheduling

func (x *OrderedJob) GetScheduling() *JobScheduling

func (*OrderedJob) GetSparkJob

func (x *OrderedJob) GetSparkJob() *SparkJob

func (*OrderedJob) GetSparkRJob

func (x *OrderedJob) GetSparkRJob() *SparkRJob

func (*OrderedJob) GetSparkSqlJob

func (x *OrderedJob) GetSparkSqlJob() *SparkSqlJob

func (*OrderedJob) GetStepId

func (x *OrderedJob) GetStepId() string

func (*OrderedJob) ProtoMessage

func (*OrderedJob) ProtoMessage()

func (*OrderedJob) ProtoReflect

func (x *OrderedJob) ProtoReflect() protoreflect.Message

func (*OrderedJob) Reset

func (x *OrderedJob) Reset()

func (*OrderedJob) String

func (x *OrderedJob) String() string

type OrderedJob_HadoopJob

type OrderedJob_HadoopJob struct {
	// Optional. Job is a Hadoop job.
	HadoopJob *HadoopJob `protobuf:"bytes,2,opt,name=hadoop_job,json=hadoopJob,proto3,oneof"`
}

type OrderedJob_HiveJob

type OrderedJob_HiveJob struct {
	// Optional. Job is a Hive job.
	HiveJob *HiveJob `protobuf:"bytes,5,opt,name=hive_job,json=hiveJob,proto3,oneof"`
}

type OrderedJob_PigJob

type OrderedJob_PigJob struct {
	// Optional. Job is a Pig job.
	PigJob *PigJob `protobuf:"bytes,6,opt,name=pig_job,json=pigJob,proto3,oneof"`
}

type OrderedJob_PrestoJob

type OrderedJob_PrestoJob struct {
	// Optional. Job is a Presto job.
	PrestoJob *PrestoJob `protobuf:"bytes,12,opt,name=presto_job,json=prestoJob,proto3,oneof"`
}

type OrderedJob_PysparkJob

type OrderedJob_PysparkJob struct {
	// Optional. Job is a PySpark job.
	PysparkJob *PySparkJob `protobuf:"bytes,4,opt,name=pyspark_job,json=pysparkJob,proto3,oneof"`
}

type OrderedJob_SparkJob

type OrderedJob_SparkJob struct {
	// Optional. Job is a Spark job.
	SparkJob *SparkJob `protobuf:"bytes,3,opt,name=spark_job,json=sparkJob,proto3,oneof"`
}

type OrderedJob_SparkRJob

type OrderedJob_SparkRJob struct {
	// Optional. Job is a SparkR job.
	SparkRJob *SparkRJob `protobuf:"bytes,11,opt,name=spark_r_job,json=sparkRJob,proto3,oneof"`
}

type OrderedJob_SparkSqlJob

type OrderedJob_SparkSqlJob struct {
	// Optional. Job is a SparkSql job.
	SparkSqlJob *SparkSqlJob `protobuf:"bytes,7,opt,name=spark_sql_job,json=sparkSqlJob,proto3,oneof"`
}

type ParameterValidation

type ParameterValidation struct {

	// Required. The type of validation to be performed.
	//
	// Types that are assignable to ValidationType:
	//	*ParameterValidation_Regex
	//	*ParameterValidation_Values
	ValidationType isParameterValidation_ValidationType `protobuf_oneof:"validation_type"`
	// contains filtered or unexported fields
}

Configuration for parameter validation.

func (*ParameterValidation) Descriptor deprecated

func (*ParameterValidation) Descriptor() ([]byte, []int)

Deprecated: Use ParameterValidation.ProtoReflect.Descriptor instead.

func (*ParameterValidation) GetRegex

func (x *ParameterValidation) GetRegex() *RegexValidation

func (*ParameterValidation) GetValidationType

func (m *ParameterValidation) GetValidationType() isParameterValidation_ValidationType

func (*ParameterValidation) GetValues

func (x *ParameterValidation) GetValues() *ValueValidation

func (*ParameterValidation) ProtoMessage

func (*ParameterValidation) ProtoMessage()

func (*ParameterValidation) ProtoReflect

func (x *ParameterValidation) ProtoReflect() protoreflect.Message

func (*ParameterValidation) Reset

func (x *ParameterValidation) Reset()

func (*ParameterValidation) String

func (x *ParameterValidation) String() string

type ParameterValidation_Regex

type ParameterValidation_Regex struct {
	// Validation based on regular expressions.
	Regex *RegexValidation `protobuf:"bytes,1,opt,name=regex,proto3,oneof"`
}

type ParameterValidation_Values

type ParameterValidation_Values struct {
	// Validation based on a list of allowed values.
	Values *ValueValidation `protobuf:"bytes,2,opt,name=values,proto3,oneof"`
}

type PigJob

type PigJob struct {

	// Required. The sequence of Pig queries to execute, specified as an HCFS
	// file URI or a list of queries.
	//
	// Types that are assignable to Queries:
	//	*PigJob_QueryFileUri
	//	*PigJob_QueryList
	Queries isPigJob_Queries `protobuf_oneof:"queries"`
	// Optional. Whether to continue executing queries if a query fails.
	// The default value is `false`. Setting to `true` can be useful when
	// executing independent parallel queries.
	ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"`
	// Optional. Mapping of query variable names to values (equivalent to the Pig
	// command: `name=[value]`).
	ScriptVariables map[string]string `` /* 194-byte string literal not displayed */
	// Optional. A mapping of property names to values, used to configure Pig.
	// Properties that conflict with values set by the Dataproc API may be
	// overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
	// /etc/pig/conf/pig.properties, and classes in user code.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. HCFS URIs of jar files to add to the CLASSPATH of
	// the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
	JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"`
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,7,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache Pig](https://pig.apache.org/) queries on YARN.

func (*PigJob) Descriptor deprecated

func (*PigJob) Descriptor() ([]byte, []int)

Deprecated: Use PigJob.ProtoReflect.Descriptor instead.

func (*PigJob) GetContinueOnFailure

func (x *PigJob) GetContinueOnFailure() bool

func (*PigJob) GetJarFileUris

func (x *PigJob) GetJarFileUris() []string

func (*PigJob) GetLoggingConfig

func (x *PigJob) GetLoggingConfig() *LoggingConfig

func (*PigJob) GetProperties

func (x *PigJob) GetProperties() map[string]string

func (*PigJob) GetQueries

func (m *PigJob) GetQueries() isPigJob_Queries

func (*PigJob) GetQueryFileUri

func (x *PigJob) GetQueryFileUri() string

func (*PigJob) GetQueryList

func (x *PigJob) GetQueryList() *QueryList

func (*PigJob) GetScriptVariables

func (x *PigJob) GetScriptVariables() map[string]string

func (*PigJob) ProtoMessage

func (*PigJob) ProtoMessage()

func (*PigJob) ProtoReflect

func (x *PigJob) ProtoReflect() protoreflect.Message

func (*PigJob) Reset

func (x *PigJob) Reset()

func (*PigJob) String

func (x *PigJob) String() string

type PigJob_QueryFileUri

type PigJob_QueryFileUri struct {
	// The HCFS URI of the script that contains the Pig queries.
	QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"`
}

type PigJob_QueryList

type PigJob_QueryList struct {
	// A list of queries.
	QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"`
}

type PrestoJob

type PrestoJob struct {

	// Required. The sequence of Presto queries to execute, specified as
	// either an HCFS file URI or as a list of queries.
	//
	// Types that are assignable to Queries:
	//	*PrestoJob_QueryFileUri
	//	*PrestoJob_QueryList
	Queries isPrestoJob_Queries `protobuf_oneof:"queries"`
	// Optional. Whether to continue executing queries if a query fails.
	// The default value is `false`. Setting to `true` can be useful when
	// executing independent parallel queries.
	ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"`
	// Optional. The format in which query output will be displayed. See the
	// Presto documentation for supported output formats
	OutputFormat string `protobuf:"bytes,4,opt,name=output_format,json=outputFormat,proto3" json:"output_format,omitempty"`
	// Optional. Presto client tags to attach to this query
	ClientTags []string `protobuf:"bytes,5,rep,name=client_tags,json=clientTags,proto3" json:"client_tags,omitempty"`
	// Optional. A mapping of property names to values. Used to set Presto
	// [session properties](https://prestodb.io/docs/current/sql/set-session.html)
	// Equivalent to using the --session flag in the Presto CLI
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,7,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Presto](https://prestosql.io/) queries. **IMPORTANT**: The [Dataproc Presto Optional Component](https://cloud.google.com/dataproc/docs/concepts/components/presto) must be enabled when the cluster is created to submit a Presto job to the cluster.

func (*PrestoJob) Descriptor deprecated

func (*PrestoJob) Descriptor() ([]byte, []int)

Deprecated: Use PrestoJob.ProtoReflect.Descriptor instead.

func (*PrestoJob) GetClientTags

func (x *PrestoJob) GetClientTags() []string

func (*PrestoJob) GetContinueOnFailure

func (x *PrestoJob) GetContinueOnFailure() bool

func (*PrestoJob) GetLoggingConfig

func (x *PrestoJob) GetLoggingConfig() *LoggingConfig

func (*PrestoJob) GetOutputFormat

func (x *PrestoJob) GetOutputFormat() string

func (*PrestoJob) GetProperties

func (x *PrestoJob) GetProperties() map[string]string

func (*PrestoJob) GetQueries

func (m *PrestoJob) GetQueries() isPrestoJob_Queries

func (*PrestoJob) GetQueryFileUri

func (x *PrestoJob) GetQueryFileUri() string

func (*PrestoJob) GetQueryList

func (x *PrestoJob) GetQueryList() *QueryList

func (*PrestoJob) ProtoMessage

func (*PrestoJob) ProtoMessage()

func (*PrestoJob) ProtoReflect

func (x *PrestoJob) ProtoReflect() protoreflect.Message

func (*PrestoJob) Reset

func (x *PrestoJob) Reset()

func (*PrestoJob) String

func (x *PrestoJob) String() string

type PrestoJob_QueryFileUri

type PrestoJob_QueryFileUri struct {
	// The HCFS URI of the script that contains SQL queries.
	QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"`
}

type PrestoJob_QueryList

type PrestoJob_QueryList struct {
	// A list of queries.
	QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"`
}

type PySparkJob

type PySparkJob struct {

	// Required. The HCFS URI of the main Python file to use as the driver. Must
	// be a .py file.
	MainPythonFileUri string `protobuf:"bytes,1,opt,name=main_python_file_uri,json=mainPythonFileUri,proto3" json:"main_python_file_uri,omitempty"`
	// Optional. The arguments to pass to the driver.  Do not include arguments,
	// such as `--conf`, that can be set as job properties, since a collision may
	// occur that causes an incorrect job submission.
	Args []string `protobuf:"bytes,2,rep,name=args,proto3" json:"args,omitempty"`
	// Optional. HCFS file URIs of Python files to pass to the PySpark
	// framework. Supported file types: .py, .egg, and .zip.
	PythonFileUris []string `protobuf:"bytes,3,rep,name=python_file_uris,json=pythonFileUris,proto3" json:"python_file_uris,omitempty"`
	// Optional. HCFS URIs of jar files to add to the CLASSPATHs of the
	// Python driver and tasks.
	JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"`
	// Optional. HCFS URIs of files to be placed in the working directory of
	// each executor. Useful for naively parallel tasks.
	FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"`
	// Optional. HCFS URIs of archives to be extracted into the working directory
	// of each executor. Supported file types:
	// .jar, .tar, .tar.gz, .tgz, and .zip.
	ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"`
	// Optional. A mapping of property names to values, used to configure PySpark.
	// Properties that conflict with values set by the Dataproc API may be
	// overwritten. Can include properties set in
	// /etc/spark/conf/spark-defaults.conf and classes in user code.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN.

func (*PySparkJob) Descriptor deprecated

func (*PySparkJob) Descriptor() ([]byte, []int)

Deprecated: Use PySparkJob.ProtoReflect.Descriptor instead.

func (*PySparkJob) GetArchiveUris

func (x *PySparkJob) GetArchiveUris() []string

func (*PySparkJob) GetArgs

func (x *PySparkJob) GetArgs() []string

func (*PySparkJob) GetFileUris

func (x *PySparkJob) GetFileUris() []string

func (*PySparkJob) GetJarFileUris

func (x *PySparkJob) GetJarFileUris() []string

func (*PySparkJob) GetLoggingConfig

func (x *PySparkJob) GetLoggingConfig() *LoggingConfig

func (*PySparkJob) GetMainPythonFileUri

func (x *PySparkJob) GetMainPythonFileUri() string

func (*PySparkJob) GetProperties

func (x *PySparkJob) GetProperties() map[string]string

func (*PySparkJob) GetPythonFileUris

func (x *PySparkJob) GetPythonFileUris() []string

func (*PySparkJob) ProtoMessage

func (*PySparkJob) ProtoMessage()

func (*PySparkJob) ProtoReflect

func (x *PySparkJob) ProtoReflect() protoreflect.Message

func (*PySparkJob) Reset

func (x *PySparkJob) Reset()

func (*PySparkJob) String

func (x *PySparkJob) String() string

type QueryList

type QueryList struct {

	// Required. The queries to execute. You do not need to terminate a query
	// with a semicolon. Multiple queries can be specified in one string
	// by separating each with a semicolon. Here is an example of an Cloud
	// Dataproc API snippet that uses a QueryList to specify a HiveJob:
	//
	//     "hiveJob": {
	//       "queryList": {
	//         "queries": [
	//           "query1",
	//           "query2",
	//           "query3;query4",
	//         ]
	//       }
	//     }
	Queries []string `protobuf:"bytes,1,rep,name=queries,proto3" json:"queries,omitempty"`
	// contains filtered or unexported fields
}

A list of queries to run on a cluster.

func (*QueryList) Descriptor deprecated

func (*QueryList) Descriptor() ([]byte, []int)

Deprecated: Use QueryList.ProtoReflect.Descriptor instead.

func (*QueryList) GetQueries

func (x *QueryList) GetQueries() []string

func (*QueryList) ProtoMessage

func (*QueryList) ProtoMessage()

func (*QueryList) ProtoReflect

func (x *QueryList) ProtoReflect() protoreflect.Message

func (*QueryList) Reset

func (x *QueryList) Reset()

func (*QueryList) String

func (x *QueryList) String() string

type RegexValidation

type RegexValidation struct {

	// Required. RE2 regular expressions used to validate the parameter's value.
	// The value must match the regex in its entirety (substring
	// matches are not sufficient).
	Regexes []string `protobuf:"bytes,1,rep,name=regexes,proto3" json:"regexes,omitempty"`
	// contains filtered or unexported fields
}

Validation based on regular expressions.

func (*RegexValidation) Descriptor deprecated

func (*RegexValidation) Descriptor() ([]byte, []int)

Deprecated: Use RegexValidation.ProtoReflect.Descriptor instead.

func (*RegexValidation) GetRegexes

func (x *RegexValidation) GetRegexes() []string

func (*RegexValidation) ProtoMessage

func (*RegexValidation) ProtoMessage()

func (*RegexValidation) ProtoReflect

func (x *RegexValidation) ProtoReflect() protoreflect.Message

func (*RegexValidation) Reset

func (x *RegexValidation) Reset()

func (*RegexValidation) String

func (x *RegexValidation) String() string

type ReservationAffinity

type ReservationAffinity struct {

	// Optional. Type of reservation to consume
	ConsumeReservationType ReservationAffinity_Type `` /* 185-byte string literal not displayed */
	// Optional. Corresponds to the label key of reservation resource.
	Key string `protobuf:"bytes,2,opt,name=key,proto3" json:"key,omitempty"`
	// Optional. Corresponds to the label values of reservation resource.
	Values []string `protobuf:"bytes,3,rep,name=values,proto3" json:"values,omitempty"`
	// contains filtered or unexported fields
}

Reservation Affinity for consuming Zonal reservation.

func (*ReservationAffinity) Descriptor deprecated

func (*ReservationAffinity) Descriptor() ([]byte, []int)

Deprecated: Use ReservationAffinity.ProtoReflect.Descriptor instead.

func (*ReservationAffinity) GetConsumeReservationType

func (x *ReservationAffinity) GetConsumeReservationType() ReservationAffinity_Type

func (*ReservationAffinity) GetKey

func (x *ReservationAffinity) GetKey() string

func (*ReservationAffinity) GetValues

func (x *ReservationAffinity) GetValues() []string

func (*ReservationAffinity) ProtoMessage

func (*ReservationAffinity) ProtoMessage()

func (*ReservationAffinity) ProtoReflect

func (x *ReservationAffinity) ProtoReflect() protoreflect.Message

func (*ReservationAffinity) Reset

func (x *ReservationAffinity) Reset()

func (*ReservationAffinity) String

func (x *ReservationAffinity) String() string

type ReservationAffinity_Type

type ReservationAffinity_Type int32

Indicates whether to consume capacity from an reservation or not.

const (
	ReservationAffinity_TYPE_UNSPECIFIED ReservationAffinity_Type = 0
	// Do not consume from any allocated capacity.
	ReservationAffinity_NO_RESERVATION ReservationAffinity_Type = 1
	// Consume any reservation available.
	ReservationAffinity_ANY_RESERVATION ReservationAffinity_Type = 2
	// Must consume from a specific reservation. Must specify key value fields
	// for specifying the reservations.
	ReservationAffinity_SPECIFIC_RESERVATION ReservationAffinity_Type = 3
)

func (ReservationAffinity_Type) Descriptor

func (ReservationAffinity_Type) Enum

func (ReservationAffinity_Type) EnumDescriptor deprecated

func (ReservationAffinity_Type) EnumDescriptor() ([]byte, []int)

Deprecated: Use ReservationAffinity_Type.Descriptor instead.

func (ReservationAffinity_Type) Number

func (ReservationAffinity_Type) String

func (x ReservationAffinity_Type) String() string

func (ReservationAffinity_Type) Type

type SecurityConfig

type SecurityConfig struct {

	// Kerberos related configuration.
	KerberosConfig *KerberosConfig `protobuf:"bytes,1,opt,name=kerberos_config,json=kerberosConfig,proto3" json:"kerberos_config,omitempty"`
	// contains filtered or unexported fields
}

Security related configuration, including Kerberos.

func (*SecurityConfig) Descriptor deprecated

func (*SecurityConfig) Descriptor() ([]byte, []int)

Deprecated: Use SecurityConfig.ProtoReflect.Descriptor instead.

func (*SecurityConfig) GetKerberosConfig

func (x *SecurityConfig) GetKerberosConfig() *KerberosConfig

func (*SecurityConfig) ProtoMessage

func (*SecurityConfig) ProtoMessage()

func (*SecurityConfig) ProtoReflect

func (x *SecurityConfig) ProtoReflect() protoreflect.Message

func (*SecurityConfig) Reset

func (x *SecurityConfig) Reset()

func (*SecurityConfig) String

func (x *SecurityConfig) String() string

type SoftwareConfig

type SoftwareConfig struct {

	// Optional. The version of software inside the cluster. It must be one of the
	// supported [Dataproc
	// Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions),
	// such as "1.2" (including a subminor version, such as "1.2.29"), or the
	// ["preview"
	// version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
	// If unspecified, it defaults to the latest Debian version.
	ImageVersion string `protobuf:"bytes,1,opt,name=image_version,json=imageVersion,proto3" json:"image_version,omitempty"`
	// Optional. The properties to set on daemon config files.
	//
	// Property keys are specified in `prefix:property` format, for example
	// `core:hadoop.tmp.dir`. The following are supported prefixes
	// and their mappings:
	//
	// * capacity-scheduler: `capacity-scheduler.xml`
	// * core:   `core-site.xml`
	// * distcp: `distcp-default.xml`
	// * hdfs:   `hdfs-site.xml`
	// * hive:   `hive-site.xml`
	// * mapred: `mapred-site.xml`
	// * pig:    `pig.properties`
	// * spark:  `spark-defaults.conf`
	// * yarn:   `yarn-site.xml`
	//
	// For more information, see [Cluster
	// properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties).
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. The set of components to activate on the cluster.
	OptionalComponents []Component `` /* 163-byte string literal not displayed */
	// contains filtered or unexported fields
}

Specifies the selection and config of software inside the cluster.

func (*SoftwareConfig) Descriptor deprecated

func (*SoftwareConfig) Descriptor() ([]byte, []int)

Deprecated: Use SoftwareConfig.ProtoReflect.Descriptor instead.

func (*SoftwareConfig) GetImageVersion

func (x *SoftwareConfig) GetImageVersion() string

func (*SoftwareConfig) GetOptionalComponents

func (x *SoftwareConfig) GetOptionalComponents() []Component

func (*SoftwareConfig) GetProperties

func (x *SoftwareConfig) GetProperties() map[string]string

func (*SoftwareConfig) ProtoMessage

func (*SoftwareConfig) ProtoMessage()

func (*SoftwareConfig) ProtoReflect

func (x *SoftwareConfig) ProtoReflect() protoreflect.Message

func (*SoftwareConfig) Reset

func (x *SoftwareConfig) Reset()

func (*SoftwareConfig) String

func (x *SoftwareConfig) String() string

type SparkJob

type SparkJob struct {

	// Required. The specification of the main method to call to drive the job.
	// Specify either the jar file that contains the main class or the main class
	// name. To pass both a main jar and a main class in that jar, add the jar to
	// `CommonJob.jar_file_uris`, and then specify the main class name in
	// `main_class`.
	//
	// Types that are assignable to Driver:
	//	*SparkJob_MainJarFileUri
	//	*SparkJob_MainClass
	Driver isSparkJob_Driver `protobuf_oneof:"driver"`
	// Optional. The arguments to pass to the driver. Do not include arguments,
	// such as `--conf`, that can be set as job properties, since a collision may
	// occur that causes an incorrect job submission.
	Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"`
	// Optional. HCFS URIs of jar files to add to the CLASSPATHs of the
	// Spark driver and tasks.
	JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"`
	// Optional. HCFS URIs of files to be placed in the working directory of
	// each executor. Useful for naively parallel tasks.
	FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"`
	// Optional. HCFS URIs of archives to be extracted into the working directory
	// of each executor. Supported file types:
	// .jar, .tar, .tar.gz, .tgz, and .zip.
	ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"`
	// Optional. A mapping of property names to values, used to configure Spark.
	// Properties that conflict with values set by the Dataproc API may be
	// overwritten. Can include properties set in
	// /etc/spark/conf/spark-defaults.conf and classes in user code.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache Spark](http://spark.apache.org/) applications on YARN.

func (*SparkJob) Descriptor deprecated

func (*SparkJob) Descriptor() ([]byte, []int)

Deprecated: Use SparkJob.ProtoReflect.Descriptor instead.

func (*SparkJob) GetArchiveUris

func (x *SparkJob) GetArchiveUris() []string

func (*SparkJob) GetArgs

func (x *SparkJob) GetArgs() []string

func (*SparkJob) GetDriver

func (m *SparkJob) GetDriver() isSparkJob_Driver

func (*SparkJob) GetFileUris

func (x *SparkJob) GetFileUris() []string

func (*SparkJob) GetJarFileUris

func (x *SparkJob) GetJarFileUris() []string

func (*SparkJob) GetLoggingConfig

func (x *SparkJob) GetLoggingConfig() *LoggingConfig

func (*SparkJob) GetMainClass

func (x *SparkJob) GetMainClass() string

func (*SparkJob) GetMainJarFileUri

func (x *SparkJob) GetMainJarFileUri() string

func (*SparkJob) GetProperties

func (x *SparkJob) GetProperties() map[string]string

func (*SparkJob) ProtoMessage

func (*SparkJob) ProtoMessage()

func (*SparkJob) ProtoReflect

func (x *SparkJob) ProtoReflect() protoreflect.Message

func (*SparkJob) Reset

func (x *SparkJob) Reset()

func (*SparkJob) String

func (x *SparkJob) String() string

type SparkJob_MainClass

type SparkJob_MainClass struct {
	// The name of the driver's main class. The jar file that contains the class
	// must be in the default CLASSPATH or specified in `jar_file_uris`.
	MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,proto3,oneof"`
}

type SparkJob_MainJarFileUri

type SparkJob_MainJarFileUri struct {
	// The HCFS URI of the jar file that contains the main class.
	MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,proto3,oneof"`
}

type SparkRJob

type SparkRJob struct {

	// Required. The HCFS URI of the main R file to use as the driver.
	// Must be a .R file.
	MainRFileUri string `protobuf:"bytes,1,opt,name=main_r_file_uri,json=mainRFileUri,proto3" json:"main_r_file_uri,omitempty"`
	// Optional. The arguments to pass to the driver.  Do not include arguments,
	// such as `--conf`, that can be set as job properties, since a collision may
	// occur that causes an incorrect job submission.
	Args []string `protobuf:"bytes,2,rep,name=args,proto3" json:"args,omitempty"`
	// Optional. HCFS URIs of files to be placed in the working directory of
	// each executor. Useful for naively parallel tasks.
	FileUris []string `protobuf:"bytes,3,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"`
	// Optional. HCFS URIs of archives to be extracted into the working directory
	// of each executor. Supported file types:
	// .jar, .tar, .tar.gz, .tgz, and .zip.
	ArchiveUris []string `protobuf:"bytes,4,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"`
	// Optional. A mapping of property names to values, used to configure SparkR.
	// Properties that conflict with values set by the Dataproc API may be
	// overwritten. Can include properties set in
	// /etc/spark/conf/spark-defaults.conf and classes in user code.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,6,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html) applications on YARN.

func (*SparkRJob) Descriptor deprecated

func (*SparkRJob) Descriptor() ([]byte, []int)

Deprecated: Use SparkRJob.ProtoReflect.Descriptor instead.

func (*SparkRJob) GetArchiveUris

func (x *SparkRJob) GetArchiveUris() []string

func (*SparkRJob) GetArgs

func (x *SparkRJob) GetArgs() []string

func (*SparkRJob) GetFileUris

func (x *SparkRJob) GetFileUris() []string

func (*SparkRJob) GetLoggingConfig

func (x *SparkRJob) GetLoggingConfig() *LoggingConfig

func (*SparkRJob) GetMainRFileUri

func (x *SparkRJob) GetMainRFileUri() string

func (*SparkRJob) GetProperties

func (x *SparkRJob) GetProperties() map[string]string

func (*SparkRJob) ProtoMessage

func (*SparkRJob) ProtoMessage()

func (*SparkRJob) ProtoReflect

func (x *SparkRJob) ProtoReflect() protoreflect.Message

func (*SparkRJob) Reset

func (x *SparkRJob) Reset()

func (*SparkRJob) String

func (x *SparkRJob) String() string

type SparkSqlJob

type SparkSqlJob struct {

	// Required. The sequence of Spark SQL queries to execute, specified as
	// either an HCFS file URI or as a list of queries.
	//
	// Types that are assignable to Queries:
	//	*SparkSqlJob_QueryFileUri
	//	*SparkSqlJob_QueryList
	Queries isSparkSqlJob_Queries `protobuf_oneof:"queries"`
	// Optional. Mapping of query variable names to values (equivalent to the
	// Spark SQL command: SET `name="value";`).
	ScriptVariables map[string]string `` /* 194-byte string literal not displayed */
	// Optional. A mapping of property names to values, used to configure
	// Spark SQL's SparkConf. Properties that conflict with values set by the
	// Dataproc API may be overwritten.
	Properties map[string]string `` /* 161-byte string literal not displayed */
	// Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
	JarFileUris []string `protobuf:"bytes,56,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"`
	// Optional. The runtime log config for job execution.
	LoggingConfig *LoggingConfig `protobuf:"bytes,6,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/) queries.

func (*SparkSqlJob) Descriptor deprecated

func (*SparkSqlJob) Descriptor() ([]byte, []int)

Deprecated: Use SparkSqlJob.ProtoReflect.Descriptor instead.

func (*SparkSqlJob) GetJarFileUris

func (x *SparkSqlJob) GetJarFileUris() []string

func (*SparkSqlJob) GetLoggingConfig

func (x *SparkSqlJob) GetLoggingConfig() *LoggingConfig

func (*SparkSqlJob) GetProperties

func (x *SparkSqlJob) GetProperties() map[string]string

func (*SparkSqlJob) GetQueries

func (m *SparkSqlJob) GetQueries() isSparkSqlJob_Queries

func (*SparkSqlJob) GetQueryFileUri

func (x *SparkSqlJob) GetQueryFileUri() string

func (*SparkSqlJob) GetQueryList

func (x *SparkSqlJob) GetQueryList() *QueryList

func (*SparkSqlJob) GetScriptVariables

func (x *SparkSqlJob) GetScriptVariables() map[string]string

func (*SparkSqlJob) ProtoMessage

func (*SparkSqlJob) ProtoMessage()

func (*SparkSqlJob) ProtoReflect

func (x *SparkSqlJob) ProtoReflect() protoreflect.Message

func (*SparkSqlJob) Reset

func (x *SparkSqlJob) Reset()

func (*SparkSqlJob) String

func (x *SparkSqlJob) String() string

type SparkSqlJob_QueryFileUri

type SparkSqlJob_QueryFileUri struct {
	// The HCFS URI of the script that contains SQL queries.
	QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"`
}

type SparkSqlJob_QueryList

type SparkSqlJob_QueryList struct {
	// A list of queries.
	QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"`
}

type SubmitJobRequest

type SubmitJobRequest struct {

	// Required. The ID of the Google Cloud Platform project that the job
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The job resource.
	Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"`
	// Optional. A unique id used to identify the request. If the server
	// receives two [SubmitJobRequest][google.cloud.dataproc.v1.SubmitJobRequest] requests  with the same
	// id, then the second request will be ignored and the
	// first [Job][google.cloud.dataproc.v1.Job] created and stored in the backend
	// is returned.
	//
	// It is recommended to always set this value to a
	// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
	//
	// The id must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). The maximum length is 40 characters.
	RequestId string `protobuf:"bytes,4,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"`
	// contains filtered or unexported fields
}

A request to submit a job.

func (*SubmitJobRequest) Descriptor deprecated

func (*SubmitJobRequest) Descriptor() ([]byte, []int)

Deprecated: Use SubmitJobRequest.ProtoReflect.Descriptor instead.

func (*SubmitJobRequest) GetJob

func (x *SubmitJobRequest) GetJob() *Job

func (*SubmitJobRequest) GetProjectId

func (x *SubmitJobRequest) GetProjectId() string

func (*SubmitJobRequest) GetRegion

func (x *SubmitJobRequest) GetRegion() string

func (*SubmitJobRequest) GetRequestId

func (x *SubmitJobRequest) GetRequestId() string

func (*SubmitJobRequest) ProtoMessage

func (*SubmitJobRequest) ProtoMessage()

func (*SubmitJobRequest) ProtoReflect

func (x *SubmitJobRequest) ProtoReflect() protoreflect.Message

func (*SubmitJobRequest) Reset

func (x *SubmitJobRequest) Reset()

func (*SubmitJobRequest) String

func (x *SubmitJobRequest) String() string

type TemplateParameter

type TemplateParameter struct {

	// Required. Parameter name.
	// The parameter name is used as the key, and paired with the
	// parameter value, which are passed to the template when the template
	// is instantiated.
	// The name must contain only capital letters (A-Z), numbers (0-9), and
	// underscores (_), and must not start with a number. The maximum length is
	// 40 characters.
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// Required. Paths to all fields that the parameter replaces.
	// A field is allowed to appear in at most one parameter's list of field
	// paths.
	//
	// A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask].
	// For example, a field path that references the zone field of a workflow
	// template's cluster selector would be specified as
	// `placement.clusterSelector.zone`.
	//
	// Also, field paths can reference fields using the following syntax:
	//
	// * Values in maps can be referenced by key:
	//     * labels['key']
	//     * placement.clusterSelector.clusterLabels['key']
	//     * placement.managedCluster.labels['key']
	//     * placement.clusterSelector.clusterLabels['key']
	//     * jobs['step-id'].labels['key']
	//
	// * Jobs in the jobs list can be referenced by step-id:
	//     * jobs['step-id'].hadoopJob.mainJarFileUri
	//     * jobs['step-id'].hiveJob.queryFileUri
	//     * jobs['step-id'].pySparkJob.mainPythonFileUri
	//     * jobs['step-id'].hadoopJob.jarFileUris[0]
	//     * jobs['step-id'].hadoopJob.archiveUris[0]
	//     * jobs['step-id'].hadoopJob.fileUris[0]
	//     * jobs['step-id'].pySparkJob.pythonFileUris[0]
	//
	// * Items in repeated fields can be referenced by a zero-based index:
	//     * jobs['step-id'].sparkJob.args[0]
	//
	// * Other examples:
	//     * jobs['step-id'].hadoopJob.properties['key']
	//     * jobs['step-id'].hadoopJob.args[0]
	//     * jobs['step-id'].hiveJob.scriptVariables['key']
	//     * jobs['step-id'].hadoopJob.mainJarFileUri
	//     * placement.clusterSelector.zone
	//
	// It may not be possible to parameterize maps and repeated fields in their
	// entirety since only individual map values and individual items in repeated
	// fields can be referenced. For example, the following field paths are
	// invalid:
	//
	// - placement.clusterSelector.clusterLabels
	// - jobs['step-id'].sparkJob.args
	Fields []string `protobuf:"bytes,2,rep,name=fields,proto3" json:"fields,omitempty"`
	// Optional. Brief description of the parameter.
	// Must not exceed 1024 characters.
	Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"`
	// Optional. Validation rules to be applied to this parameter's value.
	Validation *ParameterValidation `protobuf:"bytes,4,opt,name=validation,proto3" json:"validation,omitempty"`
	// contains filtered or unexported fields
}

A configurable parameter that replaces one or more fields in the template. Parameterizable fields: - Labels - File uris - Job properties - Job arguments - Script variables - Main class (in HadoopJob and SparkJob) - Zone (in ClusterSelector)

func (*TemplateParameter) Descriptor deprecated

func (*TemplateParameter) Descriptor() ([]byte, []int)

Deprecated: Use TemplateParameter.ProtoReflect.Descriptor instead.

func (*TemplateParameter) GetDescription

func (x *TemplateParameter) GetDescription() string

func (*TemplateParameter) GetFields

func (x *TemplateParameter) GetFields() []string

func (*TemplateParameter) GetName

func (x *TemplateParameter) GetName() string

func (*TemplateParameter) GetValidation

func (x *TemplateParameter) GetValidation() *ParameterValidation

func (*TemplateParameter) ProtoMessage

func (*TemplateParameter) ProtoMessage()

func (*TemplateParameter) ProtoReflect

func (x *TemplateParameter) ProtoReflect() protoreflect.Message

func (*TemplateParameter) Reset

func (x *TemplateParameter) Reset()

func (*TemplateParameter) String

func (x *TemplateParameter) String() string

type UnimplementedAutoscalingPolicyServiceServer

type UnimplementedAutoscalingPolicyServiceServer struct {
}

UnimplementedAutoscalingPolicyServiceServer can be embedded to have forward compatible implementations.

func (*UnimplementedAutoscalingPolicyServiceServer) CreateAutoscalingPolicy

func (*UnimplementedAutoscalingPolicyServiceServer) DeleteAutoscalingPolicy

func (*UnimplementedAutoscalingPolicyServiceServer) GetAutoscalingPolicy

func (*UnimplementedAutoscalingPolicyServiceServer) ListAutoscalingPolicies

func (*UnimplementedAutoscalingPolicyServiceServer) UpdateAutoscalingPolicy

type UnimplementedClusterControllerServer

type UnimplementedClusterControllerServer struct {
}

UnimplementedClusterControllerServer can be embedded to have forward compatible implementations.

func (*UnimplementedClusterControllerServer) CreateCluster

func (*UnimplementedClusterControllerServer) DeleteCluster

func (*UnimplementedClusterControllerServer) DiagnoseCluster

func (*UnimplementedClusterControllerServer) GetCluster

func (*UnimplementedClusterControllerServer) ListClusters

func (*UnimplementedClusterControllerServer) UpdateCluster

type UnimplementedJobControllerServer

type UnimplementedJobControllerServer struct {
}

UnimplementedJobControllerServer can be embedded to have forward compatible implementations.

func (*UnimplementedJobControllerServer) CancelJob

func (*UnimplementedJobControllerServer) DeleteJob

func (*UnimplementedJobControllerServer) GetJob

func (*UnimplementedJobControllerServer) ListJobs

func (*UnimplementedJobControllerServer) SubmitJob

func (*UnimplementedJobControllerServer) SubmitJobAsOperation

func (*UnimplementedJobControllerServer) UpdateJob

type UnimplementedWorkflowTemplateServiceServer

type UnimplementedWorkflowTemplateServiceServer struct {
}

UnimplementedWorkflowTemplateServiceServer can be embedded to have forward compatible implementations.

func (*UnimplementedWorkflowTemplateServiceServer) CreateWorkflowTemplate

func (*UnimplementedWorkflowTemplateServiceServer) DeleteWorkflowTemplate

func (*UnimplementedWorkflowTemplateServiceServer) GetWorkflowTemplate

func (*UnimplementedWorkflowTemplateServiceServer) InstantiateInlineWorkflowTemplate

func (*UnimplementedWorkflowTemplateServiceServer) InstantiateWorkflowTemplate

func (*UnimplementedWorkflowTemplateServiceServer) ListWorkflowTemplates

func (*UnimplementedWorkflowTemplateServiceServer) UpdateWorkflowTemplate

type UpdateAutoscalingPolicyRequest

type UpdateAutoscalingPolicyRequest struct {

	// Required. The updated autoscaling policy.
	Policy *AutoscalingPolicy `protobuf:"bytes,1,opt,name=policy,proto3" json:"policy,omitempty"`
	// contains filtered or unexported fields
}

A request to update an autoscaling policy.

func (*UpdateAutoscalingPolicyRequest) Descriptor deprecated

func (*UpdateAutoscalingPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use UpdateAutoscalingPolicyRequest.ProtoReflect.Descriptor instead.

func (*UpdateAutoscalingPolicyRequest) GetPolicy

func (*UpdateAutoscalingPolicyRequest) ProtoMessage

func (*UpdateAutoscalingPolicyRequest) ProtoMessage()

func (*UpdateAutoscalingPolicyRequest) ProtoReflect

func (*UpdateAutoscalingPolicyRequest) Reset

func (x *UpdateAutoscalingPolicyRequest) Reset()

func (*UpdateAutoscalingPolicyRequest) String

type UpdateClusterRequest

type UpdateClusterRequest struct {

	// Required. The ID of the Google Cloud Platform project the
	// cluster belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,5,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The cluster name.
	ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Required. The changes to the cluster.
	Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"`
	// Optional. Timeout for graceful YARN decomissioning. Graceful
	// decommissioning allows removing nodes from the cluster without
	// interrupting jobs in progress. Timeout specifies how long to wait for jobs
	// in progress to finish before forcefully removing nodes (and potentially
	// interrupting jobs). Default timeout is 0 (for forceful decommission), and
	// the maximum allowed timeout is 1 day. (see JSON representation of
	// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
	//
	// Only supported on Dataproc image versions 1.2 and higher.
	GracefulDecommissionTimeout *durationpb.Duration `` /* 144-byte string literal not displayed */
	// Required. Specifies the path, relative to `Cluster`, of
	// the field to update. For example, to change the number of workers
	// in a cluster to 5, the `update_mask` parameter would be
	// specified as `config.worker_config.num_instances`,
	// and the `PATCH` request body would specify the new value, as follows:
	//
	//     {
	//       "config":{
	//         "workerConfig":{
	//           "numInstances":"5"
	//         }
	//       }
	//     }
	// Similarly, to change the number of preemptible workers in a cluster to 5,
	// the `update_mask` parameter would be
	// `config.secondary_worker_config.num_instances`, and the `PATCH` request
	// body would be set as follows:
	//
	//     {
	//       "config":{
	//         "secondaryWorkerConfig":{
	//           "numInstances":"5"
	//         }
	//       }
	//     }
	// <strong>Note:</strong> Currently, only the following fields can be updated:
	//
	//  <table>
	//  <tbody>
	//  <tr>
	//  <td><strong>Mask</strong></td>
	//  <td><strong>Purpose</strong></td>
	//  </tr>
	//  <tr>
	//  <td><strong><em>labels</em></strong></td>
	//  <td>Update labels</td>
	//  </tr>
	//  <tr>
	//  <td><strong><em>config.worker_config.num_instances</em></strong></td>
	//  <td>Resize primary worker group</td>
	//  </tr>
	//  <tr>
	//  <td><strong><em>config.secondary_worker_config.num_instances</em></strong></td>
	//  <td>Resize secondary worker group</td>
	//  </tr>
	//  <tr>
	//  <td>config.autoscaling_config.policy_uri</td><td>Use, stop using, or
	//  change autoscaling policies</td>
	//  </tr>
	//  </tbody>
	//  </table>
	UpdateMask *fieldmaskpb.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"`
	// Optional. A unique id used to identify the request. If the server
	// receives two [UpdateClusterRequest][google.cloud.dataproc.v1.UpdateClusterRequest] requests  with the same
	// id, then the second request will be ignored and the
	// first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the
	// backend is returned.
	//
	// It is recommended to always set this value to a
	// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
	//
	// The id must contain only letters (a-z, A-Z), numbers (0-9),
	// underscores (_), and hyphens (-). The maximum length is 40 characters.
	RequestId string `protobuf:"bytes,7,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"`
	// contains filtered or unexported fields
}

A request to update a cluster.

func (*UpdateClusterRequest) Descriptor deprecated

func (*UpdateClusterRequest) Descriptor() ([]byte, []int)

Deprecated: Use UpdateClusterRequest.ProtoReflect.Descriptor instead.

func (*UpdateClusterRequest) GetCluster

func (x *UpdateClusterRequest) GetCluster() *Cluster

func (*UpdateClusterRequest) GetClusterName

func (x *UpdateClusterRequest) GetClusterName() string

func (*UpdateClusterRequest) GetGracefulDecommissionTimeout

func (x *UpdateClusterRequest) GetGracefulDecommissionTimeout() *durationpb.Duration

func (*UpdateClusterRequest) GetProjectId

func (x *UpdateClusterRequest) GetProjectId() string

func (*UpdateClusterRequest) GetRegion

func (x *UpdateClusterRequest) GetRegion() string

func (*UpdateClusterRequest) GetRequestId

func (x *UpdateClusterRequest) GetRequestId() string

func (*UpdateClusterRequest) GetUpdateMask

func (x *UpdateClusterRequest) GetUpdateMask() *fieldmaskpb.FieldMask

func (*UpdateClusterRequest) ProtoMessage

func (*UpdateClusterRequest) ProtoMessage()

func (*UpdateClusterRequest) ProtoReflect

func (x *UpdateClusterRequest) ProtoReflect() protoreflect.Message

func (*UpdateClusterRequest) Reset

func (x *UpdateClusterRequest) Reset()

func (*UpdateClusterRequest) String

func (x *UpdateClusterRequest) String() string

type UpdateJobRequest

type UpdateJobRequest struct {

	// Required. The ID of the Google Cloud Platform project that the job
	// belongs to.
	ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
	// Required. The Dataproc region in which to handle the request.
	Region string `protobuf:"bytes,2,opt,name=region,proto3" json:"region,omitempty"`
	// Required. The job ID.
	JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// Required. The changes to the job.
	Job *Job `protobuf:"bytes,4,opt,name=job,proto3" json:"job,omitempty"`
	// Required. Specifies the path, relative to <code>Job</code>, of
	// the field to update. For example, to update the labels of a Job the
	// <code>update_mask</code> parameter would be specified as
	// <code>labels</code>, and the `PATCH` request body would specify the new
	// value. <strong>Note:</strong> Currently, <code>labels</code> is the only
	// field that can be updated.
	UpdateMask *fieldmaskpb.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"`
	// contains filtered or unexported fields
}

A request to update a job.

func (*UpdateJobRequest) Descriptor deprecated

func (*UpdateJobRequest) Descriptor() ([]byte, []int)

Deprecated: Use UpdateJobRequest.ProtoReflect.Descriptor instead.

func (*UpdateJobRequest) GetJob

func (x *UpdateJobRequest) GetJob() *Job

func (*UpdateJobRequest) GetJobId

func (x *UpdateJobRequest) GetJobId() string

func (*UpdateJobRequest) GetProjectId

func (x *UpdateJobRequest) GetProjectId() string

func (*UpdateJobRequest) GetRegion

func (x *UpdateJobRequest) GetRegion() string

func (*UpdateJobRequest) GetUpdateMask

func (x *UpdateJobRequest) GetUpdateMask() *fieldmaskpb.FieldMask

func (*UpdateJobRequest) ProtoMessage

func (*UpdateJobRequest) ProtoMessage()

func (*UpdateJobRequest) ProtoReflect

func (x *UpdateJobRequest) ProtoReflect() protoreflect.Message

func (*UpdateJobRequest) Reset

func (x *UpdateJobRequest) Reset()

func (*UpdateJobRequest) String

func (x *UpdateJobRequest) String() string

type UpdateWorkflowTemplateRequest

type UpdateWorkflowTemplateRequest struct {

	// Required. The updated workflow template.
	//
	// The `template.version` field must match the current version.
	Template *WorkflowTemplate `protobuf:"bytes,1,opt,name=template,proto3" json:"template,omitempty"`
	// contains filtered or unexported fields
}

A request to update a workflow template.

func (*UpdateWorkflowTemplateRequest) Descriptor deprecated

func (*UpdateWorkflowTemplateRequest) Descriptor() ([]byte, []int)

Deprecated: Use UpdateWorkflowTemplateRequest.ProtoReflect.Descriptor instead.

func (*UpdateWorkflowTemplateRequest) GetTemplate

func (*UpdateWorkflowTemplateRequest) ProtoMessage

func (*UpdateWorkflowTemplateRequest) ProtoMessage()

func (*UpdateWorkflowTemplateRequest) ProtoReflect

func (*UpdateWorkflowTemplateRequest) Reset

func (x *UpdateWorkflowTemplateRequest) Reset()

func (*UpdateWorkflowTemplateRequest) String

type ValueValidation

type ValueValidation struct {

	// Required. List of allowed values for the parameter.
	Values []string `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"`
	// contains filtered or unexported fields
}

Validation based on a list of allowed values.

func (*ValueValidation) Descriptor deprecated

func (*ValueValidation) Descriptor() ([]byte, []int)

Deprecated: Use ValueValidation.ProtoReflect.Descriptor instead.

func (*ValueValidation) GetValues

func (x *ValueValidation) GetValues() []string

func (*ValueValidation) ProtoMessage

func (*ValueValidation) ProtoMessage()

func (*ValueValidation) ProtoReflect

func (x *ValueValidation) ProtoReflect() protoreflect.Message

func (*ValueValidation) Reset

func (x *ValueValidation) Reset()

func (*ValueValidation) String

func (x *ValueValidation) String() string

type WorkflowGraph

type WorkflowGraph struct {

	// Output only. The workflow nodes.
	Nodes []*WorkflowNode `protobuf:"bytes,1,rep,name=nodes,proto3" json:"nodes,omitempty"`
	// contains filtered or unexported fields
}

The workflow graph.

func (*WorkflowGraph) Descriptor deprecated

func (*WorkflowGraph) Descriptor() ([]byte, []int)

Deprecated: Use WorkflowGraph.ProtoReflect.Descriptor instead.

func (*WorkflowGraph) GetNodes

func (x *WorkflowGraph) GetNodes() []*WorkflowNode

func (*WorkflowGraph) ProtoMessage

func (*WorkflowGraph) ProtoMessage()

func (*WorkflowGraph) ProtoReflect

func (x *WorkflowGraph) ProtoReflect() protoreflect.Message

func (*WorkflowGraph) Reset

func (x *WorkflowGraph) Reset()

func (*WorkflowGraph) String

func (x *WorkflowGraph) String() string

type WorkflowMetadata

type WorkflowMetadata struct {

	// Output only. The resource name of the workflow template as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates`, the resource name of the
	//   template has the following format:
	//   `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}`
	//
	// * For `projects.locations.workflowTemplates`, the resource name of the
	//   template has the following format:
	//   `projects/{project_id}/locations/{location}/workflowTemplates/{template_id}`
	Template string `protobuf:"bytes,1,opt,name=template,proto3" json:"template,omitempty"`
	// Output only. The version of template at the time of
	// workflow instantiation.
	Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"`
	// Output only. The create cluster operation metadata.
	CreateCluster *ClusterOperation `protobuf:"bytes,3,opt,name=create_cluster,json=createCluster,proto3" json:"create_cluster,omitempty"`
	// Output only. The workflow graph.
	Graph *WorkflowGraph `protobuf:"bytes,4,opt,name=graph,proto3" json:"graph,omitempty"`
	// Output only. The delete cluster operation metadata.
	DeleteCluster *ClusterOperation `protobuf:"bytes,5,opt,name=delete_cluster,json=deleteCluster,proto3" json:"delete_cluster,omitempty"`
	// Output only. The workflow state.
	State WorkflowMetadata_State `protobuf:"varint,6,opt,name=state,proto3,enum=google.cloud.dataproc.v1.WorkflowMetadata_State" json:"state,omitempty"`
	// Output only. The name of the target cluster.
	ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"`
	// Map from parameter names to values that were used for those parameters.
	Parameters map[string]string `` /* 161-byte string literal not displayed */
	// Output only. Workflow start time.
	StartTime *timestamppb.Timestamp `protobuf:"bytes,9,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
	// Output only. Workflow end time.
	EndTime *timestamppb.Timestamp `protobuf:"bytes,10,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"`
	// Output only. The UUID of target cluster.
	ClusterUuid string `protobuf:"bytes,11,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc workflow template resource.

func (*WorkflowMetadata) Descriptor deprecated

func (*WorkflowMetadata) Descriptor() ([]byte, []int)

Deprecated: Use WorkflowMetadata.ProtoReflect.Descriptor instead.

func (*WorkflowMetadata) GetClusterName

func (x *WorkflowMetadata) GetClusterName() string

func (*WorkflowMetadata) GetClusterUuid

func (x *WorkflowMetadata) GetClusterUuid() string

func (*WorkflowMetadata) GetCreateCluster

func (x *WorkflowMetadata) GetCreateCluster() *ClusterOperation

func (*WorkflowMetadata) GetDeleteCluster

func (x *WorkflowMetadata) GetDeleteCluster() *ClusterOperation

func (*WorkflowMetadata) GetEndTime

func (x *WorkflowMetadata) GetEndTime() *timestamppb.Timestamp

func (*WorkflowMetadata) GetGraph

func (x *WorkflowMetadata) GetGraph() *WorkflowGraph

func (*WorkflowMetadata) GetParameters

func (x *WorkflowMetadata) GetParameters() map[string]string

func (*WorkflowMetadata) GetStartTime

func (x *WorkflowMetadata) GetStartTime() *timestamppb.Timestamp

func (*WorkflowMetadata) GetState

func (*WorkflowMetadata) GetTemplate

func (x *WorkflowMetadata) GetTemplate() string

func (*WorkflowMetadata) GetVersion

func (x *WorkflowMetadata) GetVersion() int32

func (*WorkflowMetadata) ProtoMessage

func (*WorkflowMetadata) ProtoMessage()

func (*WorkflowMetadata) ProtoReflect

func (x *WorkflowMetadata) ProtoReflect() protoreflect.Message

func (*WorkflowMetadata) Reset

func (x *WorkflowMetadata) Reset()

func (*WorkflowMetadata) String

func (x *WorkflowMetadata) String() string

type WorkflowMetadata_State

type WorkflowMetadata_State int32

The operation state.

const (
	// Unused.
	WorkflowMetadata_UNKNOWN WorkflowMetadata_State = 0
	// The operation has been created.
	WorkflowMetadata_PENDING WorkflowMetadata_State = 1
	// The operation is running.
	WorkflowMetadata_RUNNING WorkflowMetadata_State = 2
	// The operation is done; either cancelled or completed.
	WorkflowMetadata_DONE WorkflowMetadata_State = 3
)

func (WorkflowMetadata_State) Descriptor

func (WorkflowMetadata_State) Enum

func (WorkflowMetadata_State) EnumDescriptor deprecated

func (WorkflowMetadata_State) EnumDescriptor() ([]byte, []int)

Deprecated: Use WorkflowMetadata_State.Descriptor instead.

func (WorkflowMetadata_State) Number

func (WorkflowMetadata_State) String

func (x WorkflowMetadata_State) String() string

func (WorkflowMetadata_State) Type

type WorkflowNode

type WorkflowNode struct {

	// Output only. The name of the node.
	StepId string `protobuf:"bytes,1,opt,name=step_id,json=stepId,proto3" json:"step_id,omitempty"`
	// Output only. Node's prerequisite nodes.
	PrerequisiteStepIds []string `protobuf:"bytes,2,rep,name=prerequisite_step_ids,json=prerequisiteStepIds,proto3" json:"prerequisite_step_ids,omitempty"`
	// Output only. The job id; populated after the node enters RUNNING state.
	JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"`
	// Output only. The node state.
	State WorkflowNode_NodeState `protobuf:"varint,5,opt,name=state,proto3,enum=google.cloud.dataproc.v1.WorkflowNode_NodeState" json:"state,omitempty"`
	// Output only. The error detail.
	Error string `protobuf:"bytes,6,opt,name=error,proto3" json:"error,omitempty"`
	// contains filtered or unexported fields
}

The workflow node.

func (*WorkflowNode) Descriptor deprecated

func (*WorkflowNode) Descriptor() ([]byte, []int)

Deprecated: Use WorkflowNode.ProtoReflect.Descriptor instead.

func (*WorkflowNode) GetError

func (x *WorkflowNode) GetError() string

func (*WorkflowNode) GetJobId

func (x *WorkflowNode) GetJobId() string

func (*WorkflowNode) GetPrerequisiteStepIds

func (x *WorkflowNode) GetPrerequisiteStepIds() []string

func (*WorkflowNode) GetState

func (x *WorkflowNode) GetState() WorkflowNode_NodeState

func (*WorkflowNode) GetStepId

func (x *WorkflowNode) GetStepId() string

func (*WorkflowNode) ProtoMessage

func (*WorkflowNode) ProtoMessage()

func (*WorkflowNode) ProtoReflect

func (x *WorkflowNode) ProtoReflect() protoreflect.Message

func (*WorkflowNode) Reset

func (x *WorkflowNode) Reset()

func (*WorkflowNode) String

func (x *WorkflowNode) String() string

type WorkflowNode_NodeState

type WorkflowNode_NodeState int32

The workflow node state.

const (
	// State is unspecified.
	WorkflowNode_NODE_STATE_UNSPECIFIED WorkflowNode_NodeState = 0
	// The node is awaiting prerequisite node to finish.
	WorkflowNode_BLOCKED WorkflowNode_NodeState = 1
	// The node is runnable but not running.
	WorkflowNode_RUNNABLE WorkflowNode_NodeState = 2
	// The node is running.
	WorkflowNode_RUNNING WorkflowNode_NodeState = 3
	// The node completed successfully.
	WorkflowNode_COMPLETED WorkflowNode_NodeState = 4
	// The node failed. A node can be marked FAILED because
	// its ancestor or peer failed.
	WorkflowNode_FAILED WorkflowNode_NodeState = 5
)

func (WorkflowNode_NodeState) Descriptor

func (WorkflowNode_NodeState) Enum

func (WorkflowNode_NodeState) EnumDescriptor deprecated

func (WorkflowNode_NodeState) EnumDescriptor() ([]byte, []int)

Deprecated: Use WorkflowNode_NodeState.Descriptor instead.

func (WorkflowNode_NodeState) Number

func (WorkflowNode_NodeState) String

func (x WorkflowNode_NodeState) String() string

func (WorkflowNode_NodeState) Type

type WorkflowTemplate

type WorkflowTemplate struct {
	Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"`
	// Output only. The resource name of the workflow template, as described
	// in https://cloud.google.com/apis/design/resource_names.
	//
	// * For `projects.regions.workflowTemplates`, the resource name of the
	//   template has the following format:
	//   `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}`
	//
	// * For `projects.locations.workflowTemplates`, the resource name of the
	//   template has the following format:
	//   `projects/{project_id}/locations/{location}/workflowTemplates/{template_id}`
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// Optional. Used to perform a consistent read-modify-write.
	//
	// This field should be left blank for a `CreateWorkflowTemplate` request. It
	// is required for an `UpdateWorkflowTemplate` request, and must match the
	// current server version. A typical update template flow would fetch the
	// current template with a `GetWorkflowTemplate` request, which will return
	// the current template with the `version` field filled in with the
	// current server version. The user updates other fields in the template,
	// then returns it as part of the `UpdateWorkflowTemplate` request.
	Version int32 `protobuf:"varint,3,opt,name=version,proto3" json:"version,omitempty"`
	// Output only. The time template was created.
	CreateTime *timestamppb.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"`
	// Output only. The time template was last updated.
	UpdateTime *timestamppb.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
	// Optional. The labels to associate with this template. These labels
	// will be propagated to all jobs and clusters created by the workflow
	// instance.
	//
	// Label **keys** must contain 1 to 63 characters, and must conform to
	// [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
	//
	// Label **values** may be empty, but, if present, must contain 1 to 63
	// characters, and must conform to
	// [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
	//
	// No more than 32 labels can be associated with a template.
	Labels map[string]string `` /* 153-byte string literal not displayed */
	// Required. WorkflowTemplate scheduling information.
	Placement *WorkflowTemplatePlacement `protobuf:"bytes,7,opt,name=placement,proto3" json:"placement,omitempty"`
	// Required. The Directed Acyclic Graph of Jobs to submit.
	Jobs []*OrderedJob `protobuf:"bytes,8,rep,name=jobs,proto3" json:"jobs,omitempty"`
	// Optional. Template parameters whose values are substituted into the
	// template. Values for parameters must be provided when the template is
	// instantiated.
	Parameters []*TemplateParameter `protobuf:"bytes,9,rep,name=parameters,proto3" json:"parameters,omitempty"`
	// contains filtered or unexported fields
}

A Dataproc workflow template resource.

func (*WorkflowTemplate) Descriptor deprecated

func (*WorkflowTemplate) Descriptor() ([]byte, []int)

Deprecated: Use WorkflowTemplate.ProtoReflect.Descriptor instead.

func (*WorkflowTemplate) GetCreateTime

func (x *WorkflowTemplate) GetCreateTime() *timestamppb.Timestamp

func (*WorkflowTemplate) GetId

func (x *WorkflowTemplate) GetId() string

func (*WorkflowTemplate) GetJobs

func (x *WorkflowTemplate) GetJobs() []*OrderedJob

func (*WorkflowTemplate) GetLabels

func (x *WorkflowTemplate) GetLabels() map[string]string

func (*WorkflowTemplate) GetName

func (x *WorkflowTemplate) GetName() string

func (*WorkflowTemplate) GetParameters

func (x *WorkflowTemplate) GetParameters() []*TemplateParameter

func (*WorkflowTemplate) GetPlacement

func (x *WorkflowTemplate) GetPlacement() *WorkflowTemplatePlacement

func (*WorkflowTemplate) GetUpdateTime

func (x *WorkflowTemplate) GetUpdateTime() *timestamppb.Timestamp

func (*WorkflowTemplate) GetVersion

func (x *WorkflowTemplate) GetVersion() int32

func (*WorkflowTemplate) ProtoMessage

func (*WorkflowTemplate) ProtoMessage()

func (*WorkflowTemplate) ProtoReflect

func (x *WorkflowTemplate) ProtoReflect() protoreflect.Message

func (*WorkflowTemplate) Reset

func (x *WorkflowTemplate) Reset()

func (*WorkflowTemplate) String

func (x *WorkflowTemplate) String() string

type WorkflowTemplatePlacement

type WorkflowTemplatePlacement struct {

	// Required. Specifies where workflow executes; either on a managed
	// cluster or an existing cluster chosen by labels.
	//
	// Types that are assignable to Placement:
	//	*WorkflowTemplatePlacement_ManagedCluster
	//	*WorkflowTemplatePlacement_ClusterSelector
	Placement isWorkflowTemplatePlacement_Placement `protobuf_oneof:"placement"`
	// contains filtered or unexported fields
}

Specifies workflow execution target.

Either `managed_cluster` or `cluster_selector` is required.

func (*WorkflowTemplatePlacement) Descriptor deprecated

func (*WorkflowTemplatePlacement) Descriptor() ([]byte, []int)

Deprecated: Use WorkflowTemplatePlacement.ProtoReflect.Descriptor instead.

func (*WorkflowTemplatePlacement) GetClusterSelector

func (x *WorkflowTemplatePlacement) GetClusterSelector() *ClusterSelector

func (*WorkflowTemplatePlacement) GetManagedCluster

func (x *WorkflowTemplatePlacement) GetManagedCluster() *ManagedCluster

func (*WorkflowTemplatePlacement) GetPlacement

func (m *WorkflowTemplatePlacement) GetPlacement() isWorkflowTemplatePlacement_Placement

func (*WorkflowTemplatePlacement) ProtoMessage

func (*WorkflowTemplatePlacement) ProtoMessage()

func (*WorkflowTemplatePlacement) ProtoReflect

func (*WorkflowTemplatePlacement) Reset

func (x *WorkflowTemplatePlacement) Reset()

func (*WorkflowTemplatePlacement) String

func (x *WorkflowTemplatePlacement) String() string

type WorkflowTemplatePlacement_ClusterSelector

type WorkflowTemplatePlacement_ClusterSelector struct {
	// Optional. A selector that chooses target cluster for jobs based
	// on metadata.
	//
	// The selector is evaluated at the time each job is submitted.
	ClusterSelector *ClusterSelector `protobuf:"bytes,2,opt,name=cluster_selector,json=clusterSelector,proto3,oneof"`
}

type WorkflowTemplatePlacement_ManagedCluster

type WorkflowTemplatePlacement_ManagedCluster struct {
	// A cluster that is managed by the workflow.
	ManagedCluster *ManagedCluster `protobuf:"bytes,1,opt,name=managed_cluster,json=managedCluster,proto3,oneof"`
}

type WorkflowTemplateServiceClient

type WorkflowTemplateServiceClient interface {
	// Creates new workflow template.
	CreateWorkflowTemplate(ctx context.Context, in *CreateWorkflowTemplateRequest, opts ...grpc.CallOption) (*WorkflowTemplate, error)
	// Retrieves the latest workflow template.
	//
	// Can retrieve previously instantiated template by specifying optional
	// version parameter.
	GetWorkflowTemplate(ctx context.Context, in *GetWorkflowTemplateRequest, opts ...grpc.CallOption) (*WorkflowTemplate, error)
	// Instantiates a template and begins execution.
	//
	// The returned Operation can be used to track execution of
	// workflow by polling
	// [operations.get][google.longrunning.Operations.GetOperation].
	// The Operation will complete when entire workflow is finished.
	//
	// The running workflow can be aborted via
	// [operations.cancel][google.longrunning.Operations.CancelOperation].
	// This will cause any inflight jobs to be cancelled and workflow-owned
	// clusters to be deleted.
	//
	// The [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
	// Also see [Using
	// WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
	//
	// On successful completion,
	// [Operation.response][google.longrunning.Operation.response] will be
	// [Empty][google.protobuf.Empty].
	InstantiateWorkflowTemplate(ctx context.Context, in *InstantiateWorkflowTemplateRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
	// Instantiates a template and begins execution.
	//
	// This method is equivalent to executing the sequence
	// [CreateWorkflowTemplate][google.cloud.dataproc.v1.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1.WorkflowTemplateService.InstantiateWorkflowTemplate],
	// [DeleteWorkflowTemplate][google.cloud.dataproc.v1.WorkflowTemplateService.DeleteWorkflowTemplate].
	//
	// The returned Operation can be used to track execution of
	// workflow by polling
	// [operations.get][google.longrunning.Operations.GetOperation].
	// The Operation will complete when entire workflow is finished.
	//
	// The running workflow can be aborted via
	// [operations.cancel][google.longrunning.Operations.CancelOperation].
	// This will cause any inflight jobs to be cancelled and workflow-owned
	// clusters to be deleted.
	//
	// The [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
	// Also see [Using
	// WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
	//
	// On successful completion,
	// [Operation.response][google.longrunning.Operation.response] will be
	// [Empty][google.protobuf.Empty].
	InstantiateInlineWorkflowTemplate(ctx context.Context, in *InstantiateInlineWorkflowTemplateRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
	// Updates (replaces) workflow template. The updated template
	// must contain version that matches the current server version.
	UpdateWorkflowTemplate(ctx context.Context, in *UpdateWorkflowTemplateRequest, opts ...grpc.CallOption) (*WorkflowTemplate, error)
	// Lists workflows that match the specified filter in the request.
	ListWorkflowTemplates(ctx context.Context, in *ListWorkflowTemplatesRequest, opts ...grpc.CallOption) (*ListWorkflowTemplatesResponse, error)
	// Deletes a workflow template. It does not cancel in-progress workflows.
	DeleteWorkflowTemplate(ctx context.Context, in *DeleteWorkflowTemplateRequest, opts ...grpc.CallOption) (*emptypb.Empty, error)
}

WorkflowTemplateServiceClient is the client API for WorkflowTemplateService service.

For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.

type WorkflowTemplateServiceServer

type WorkflowTemplateServiceServer interface {
	// Creates new workflow template.
	CreateWorkflowTemplate(context.Context, *CreateWorkflowTemplateRequest) (*WorkflowTemplate, error)
	// Retrieves the latest workflow template.
	//
	// Can retrieve previously instantiated template by specifying optional
	// version parameter.
	GetWorkflowTemplate(context.Context, *GetWorkflowTemplateRequest) (*WorkflowTemplate, error)
	// Instantiates a template and begins execution.
	//
	// The returned Operation can be used to track execution of
	// workflow by polling
	// [operations.get][google.longrunning.Operations.GetOperation].
	// The Operation will complete when entire workflow is finished.
	//
	// The running workflow can be aborted via
	// [operations.cancel][google.longrunning.Operations.CancelOperation].
	// This will cause any inflight jobs to be cancelled and workflow-owned
	// clusters to be deleted.
	//
	// The [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
	// Also see [Using
	// WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
	//
	// On successful completion,
	// [Operation.response][google.longrunning.Operation.response] will be
	// [Empty][google.protobuf.Empty].
	InstantiateWorkflowTemplate(context.Context, *InstantiateWorkflowTemplateRequest) (*longrunning.Operation, error)
	// Instantiates a template and begins execution.
	//
	// This method is equivalent to executing the sequence
	// [CreateWorkflowTemplate][google.cloud.dataproc.v1.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1.WorkflowTemplateService.InstantiateWorkflowTemplate],
	// [DeleteWorkflowTemplate][google.cloud.dataproc.v1.WorkflowTemplateService.DeleteWorkflowTemplate].
	//
	// The returned Operation can be used to track execution of
	// workflow by polling
	// [operations.get][google.longrunning.Operations.GetOperation].
	// The Operation will complete when entire workflow is finished.
	//
	// The running workflow can be aborted via
	// [operations.cancel][google.longrunning.Operations.CancelOperation].
	// This will cause any inflight jobs to be cancelled and workflow-owned
	// clusters to be deleted.
	//
	// The [Operation.metadata][google.longrunning.Operation.metadata] will be
	// [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
	// Also see [Using
	// WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
	//
	// On successful completion,
	// [Operation.response][google.longrunning.Operation.response] will be
	// [Empty][google.protobuf.Empty].
	InstantiateInlineWorkflowTemplate(context.Context, *InstantiateInlineWorkflowTemplateRequest) (*longrunning.Operation, error)
	// Updates (replaces) workflow template. The updated template
	// must contain version that matches the current server version.
	UpdateWorkflowTemplate(context.Context, *UpdateWorkflowTemplateRequest) (*WorkflowTemplate, error)
	// Lists workflows that match the specified filter in the request.
	ListWorkflowTemplates(context.Context, *ListWorkflowTemplatesRequest) (*ListWorkflowTemplatesResponse, error)
	// Deletes a workflow template. It does not cancel in-progress workflows.
	DeleteWorkflowTemplate(context.Context, *DeleteWorkflowTemplateRequest) (*emptypb.Empty, error)
}

WorkflowTemplateServiceServer is the server API for WorkflowTemplateService service.

type YarnApplication

type YarnApplication struct {

	// Required. The application name.
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// Required. The application state.
	State YarnApplication_State `protobuf:"varint,2,opt,name=state,proto3,enum=google.cloud.dataproc.v1.YarnApplication_State" json:"state,omitempty"`
	// Required. The numerical progress of the application, from 1 to 100.
	Progress float32 `protobuf:"fixed32,3,opt,name=progress,proto3" json:"progress,omitempty"`
	// Optional. The HTTP URL of the ApplicationMaster, HistoryServer, or
	// TimelineServer that provides application-specific information. The URL uses
	// the internal hostname, and requires a proxy server for resolution and,
	// possibly, access.
	TrackingUrl string `protobuf:"bytes,4,opt,name=tracking_url,json=trackingUrl,proto3" json:"tracking_url,omitempty"`
	// contains filtered or unexported fields
}

A YARN application created by a job. Application information is a subset of <code>org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto</code>.

**Beta Feature**: This report is available for testing purposes only. It may be changed before final release.

func (*YarnApplication) Descriptor deprecated

func (*YarnApplication) Descriptor() ([]byte, []int)

Deprecated: Use YarnApplication.ProtoReflect.Descriptor instead.

func (*YarnApplication) GetName

func (x *YarnApplication) GetName() string

func (*YarnApplication) GetProgress

func (x *YarnApplication) GetProgress() float32

func (*YarnApplication) GetState

func (x *YarnApplication) GetState() YarnApplication_State

func (*YarnApplication) GetTrackingUrl

func (x *YarnApplication) GetTrackingUrl() string

func (*YarnApplication) ProtoMessage

func (*YarnApplication) ProtoMessage()

func (*YarnApplication) ProtoReflect

func (x *YarnApplication) ProtoReflect() protoreflect.Message

func (*YarnApplication) Reset

func (x *YarnApplication) Reset()

func (*YarnApplication) String

func (x *YarnApplication) String() string

type YarnApplication_State

type YarnApplication_State int32

The application state, corresponding to <code>YarnProtos.YarnApplicationStateProto</code>.

const (
	// Status is unspecified.
	YarnApplication_STATE_UNSPECIFIED YarnApplication_State = 0
	// Status is NEW.
	YarnApplication_NEW YarnApplication_State = 1
	// Status is NEW_SAVING.
	YarnApplication_NEW_SAVING YarnApplication_State = 2
	// Status is SUBMITTED.
	YarnApplication_SUBMITTED YarnApplication_State = 3
	// Status is ACCEPTED.
	YarnApplication_ACCEPTED YarnApplication_State = 4
	// Status is RUNNING.
	YarnApplication_RUNNING YarnApplication_State = 5
	// Status is FINISHED.
	YarnApplication_FINISHED YarnApplication_State = 6
	// Status is FAILED.
	YarnApplication_FAILED YarnApplication_State = 7
	// Status is KILLED.
	YarnApplication_KILLED YarnApplication_State = 8
)

func (YarnApplication_State) Descriptor

func (YarnApplication_State) Enum

func (YarnApplication_State) EnumDescriptor deprecated

func (YarnApplication_State) EnumDescriptor() ([]byte, []int)

Deprecated: Use YarnApplication_State.Descriptor instead.

func (YarnApplication_State) Number

func (YarnApplication_State) String

func (x YarnApplication_State) String() string

func (YarnApplication_State) Type

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL