Documentation

Index

Constants

This section is empty.

Variables

View Source
var (
	AlertPolicy_ConditionCombinerType_name = map[int32]string{
		0: "COMBINE_UNSPECIFIED",
		1: "AND",
		2: "OR",
		3: "AND_WITH_MATCHING_RESOURCE",
	}
	AlertPolicy_ConditionCombinerType_value = map[string]int32{
		"COMBINE_UNSPECIFIED":        0,
		"AND":                        1,
		"OR":                         2,
		"AND_WITH_MATCHING_RESOURCE": 3,
	}
)

Enum value maps for AlertPolicy_ConditionCombinerType.

View Source
var (
	ComparisonType_name = map[int32]string{
		0: "COMPARISON_UNSPECIFIED",
		1: "COMPARISON_GT",
		2: "COMPARISON_GE",
		3: "COMPARISON_LT",
		4: "COMPARISON_LE",
		5: "COMPARISON_EQ",
		6: "COMPARISON_NE",
	}
	ComparisonType_value = map[string]int32{
		"COMPARISON_UNSPECIFIED": 0,
		"COMPARISON_GT":          1,
		"COMPARISON_GE":          2,
		"COMPARISON_LT":          3,
		"COMPARISON_LE":          4,
		"COMPARISON_EQ":          5,
		"COMPARISON_NE":          6,
	}
)

Enum value maps for ComparisonType.

View Source
var (
	ServiceTier_name = map[int32]string{
		0: "SERVICE_TIER_UNSPECIFIED",
		1: "SERVICE_TIER_BASIC",
		2: "SERVICE_TIER_PREMIUM",
	}
	ServiceTier_value = map[string]int32{
		"SERVICE_TIER_UNSPECIFIED": 0,
		"SERVICE_TIER_BASIC":       1,
		"SERVICE_TIER_PREMIUM":     2,
	}
)

Enum value maps for ServiceTier.

View Source
var (
	Aggregation_Aligner_name = map[int32]string{
		0:  "ALIGN_NONE",
		1:  "ALIGN_DELTA",
		2:  "ALIGN_RATE",
		3:  "ALIGN_INTERPOLATE",
		4:  "ALIGN_NEXT_OLDER",
		10: "ALIGN_MIN",
		11: "ALIGN_MAX",
		12: "ALIGN_MEAN",
		13: "ALIGN_COUNT",
		14: "ALIGN_SUM",
		15: "ALIGN_STDDEV",
		16: "ALIGN_COUNT_TRUE",
		24: "ALIGN_COUNT_FALSE",
		17: "ALIGN_FRACTION_TRUE",
		18: "ALIGN_PERCENTILE_99",
		19: "ALIGN_PERCENTILE_95",
		20: "ALIGN_PERCENTILE_50",
		21: "ALIGN_PERCENTILE_05",
		23: "ALIGN_PERCENT_CHANGE",
	}
	Aggregation_Aligner_value = map[string]int32{
		"ALIGN_NONE":           0,
		"ALIGN_DELTA":          1,
		"ALIGN_RATE":           2,
		"ALIGN_INTERPOLATE":    3,
		"ALIGN_NEXT_OLDER":     4,
		"ALIGN_MIN":            10,
		"ALIGN_MAX":            11,
		"ALIGN_MEAN":           12,
		"ALIGN_COUNT":          13,
		"ALIGN_SUM":            14,
		"ALIGN_STDDEV":         15,
		"ALIGN_COUNT_TRUE":     16,
		"ALIGN_COUNT_FALSE":    24,
		"ALIGN_FRACTION_TRUE":  17,
		"ALIGN_PERCENTILE_99":  18,
		"ALIGN_PERCENTILE_95":  19,
		"ALIGN_PERCENTILE_50":  20,
		"ALIGN_PERCENTILE_05":  21,
		"ALIGN_PERCENT_CHANGE": 23,
	}
)

Enum value maps for Aggregation_Aligner.

View Source
var (
	Aggregation_Reducer_name = map[int32]string{
		0:  "REDUCE_NONE",
		1:  "REDUCE_MEAN",
		2:  "REDUCE_MIN",
		3:  "REDUCE_MAX",
		4:  "REDUCE_SUM",
		5:  "REDUCE_STDDEV",
		6:  "REDUCE_COUNT",
		7:  "REDUCE_COUNT_TRUE",
		15: "REDUCE_COUNT_FALSE",
		8:  "REDUCE_FRACTION_TRUE",
		9:  "REDUCE_PERCENTILE_99",
		10: "REDUCE_PERCENTILE_95",
		11: "REDUCE_PERCENTILE_50",
		12: "REDUCE_PERCENTILE_05",
	}
	Aggregation_Reducer_value = map[string]int32{
		"REDUCE_NONE":          0,
		"REDUCE_MEAN":          1,
		"REDUCE_MIN":           2,
		"REDUCE_MAX":           3,
		"REDUCE_SUM":           4,
		"REDUCE_STDDEV":        5,
		"REDUCE_COUNT":         6,
		"REDUCE_COUNT_TRUE":    7,
		"REDUCE_COUNT_FALSE":   15,
		"REDUCE_FRACTION_TRUE": 8,
		"REDUCE_PERCENTILE_99": 9,
		"REDUCE_PERCENTILE_95": 10,
		"REDUCE_PERCENTILE_50": 11,
		"REDUCE_PERCENTILE_05": 12,
	}
)

Enum value maps for Aggregation_Reducer.

View Source
var (
	ListTimeSeriesRequest_TimeSeriesView_name = map[int32]string{
		0: "FULL",
		1: "HEADERS",
	}
	ListTimeSeriesRequest_TimeSeriesView_value = map[string]int32{
		"FULL":    0,
		"HEADERS": 1,
	}
)

Enum value maps for ListTimeSeriesRequest_TimeSeriesView.

View Source
var (
	NotificationChannel_VerificationStatus_name = map[int32]string{
		0: "VERIFICATION_STATUS_UNSPECIFIED",
		1: "UNVERIFIED",
		2: "VERIFIED",
	}
	NotificationChannel_VerificationStatus_value = map[string]int32{
		"VERIFICATION_STATUS_UNSPECIFIED": 0,
		"UNVERIFIED":                      1,
		"VERIFIED":                        2,
	}
)

Enum value maps for NotificationChannel_VerificationStatus.

View Source
var (
	ServiceLevelObjective_View_name = map[int32]string{
		0: "VIEW_UNSPECIFIED",
		2: "FULL",
		1: "EXPLICIT",
	}
	ServiceLevelObjective_View_value = map[string]int32{
		"VIEW_UNSPECIFIED": 0,
		"FULL":             2,
		"EXPLICIT":         1,
	}
)

Enum value maps for ServiceLevelObjective_View.

View Source
var (
	UptimeCheckRegion_name = map[int32]string{
		0: "REGION_UNSPECIFIED",
		1: "USA",
		2: "EUROPE",
		3: "SOUTH_AMERICA",
		4: "ASIA_PACIFIC",
	}
	UptimeCheckRegion_value = map[string]int32{
		"REGION_UNSPECIFIED": 0,
		"USA":                1,
		"EUROPE":             2,
		"SOUTH_AMERICA":      3,
		"ASIA_PACIFIC":       4,
	}
)

Enum value maps for UptimeCheckRegion.

View Source
var (
	GroupResourceType_name = map[int32]string{
		0: "RESOURCE_TYPE_UNSPECIFIED",
		1: "INSTANCE",
		2: "AWS_ELB_LOAD_BALANCER",
	}
	GroupResourceType_value = map[string]int32{
		"RESOURCE_TYPE_UNSPECIFIED": 0,
		"INSTANCE":                  1,
		"AWS_ELB_LOAD_BALANCER":     2,
	}
)

Enum value maps for GroupResourceType.

View Source
var (
	InternalChecker_State_name = map[int32]string{
		0: "UNSPECIFIED",
		1: "CREATING",
		2: "RUNNING",
	}
	InternalChecker_State_value = map[string]int32{
		"UNSPECIFIED": 0,
		"CREATING":    1,
		"RUNNING":     2,
	}
)

Enum value maps for InternalChecker_State.

View Source
var (
	UptimeCheckConfig_HttpCheck_RequestMethod_name = map[int32]string{
		0: "METHOD_UNSPECIFIED",
		1: "GET",
		2: "POST",
	}
	UptimeCheckConfig_HttpCheck_RequestMethod_value = map[string]int32{
		"METHOD_UNSPECIFIED": 0,
		"GET":                1,
		"POST":               2,
	}
)

Enum value maps for UptimeCheckConfig_HttpCheck_RequestMethod.

View Source
var (
	UptimeCheckConfig_HttpCheck_ContentType_name = map[int32]string{
		0: "TYPE_UNSPECIFIED",
		1: "URL_ENCODED",
	}
	UptimeCheckConfig_HttpCheck_ContentType_value = map[string]int32{
		"TYPE_UNSPECIFIED": 0,
		"URL_ENCODED":      1,
	}
)

Enum value maps for UptimeCheckConfig_HttpCheck_ContentType.

View Source
var (
	UptimeCheckConfig_ContentMatcher_ContentMatcherOption_name = map[int32]string{
		0: "CONTENT_MATCHER_OPTION_UNSPECIFIED",
		1: "CONTAINS_STRING",
		2: "NOT_CONTAINS_STRING",
		3: "MATCHES_REGEX",
		4: "NOT_MATCHES_REGEX",
	}
	UptimeCheckConfig_ContentMatcher_ContentMatcherOption_value = map[string]int32{
		"CONTENT_MATCHER_OPTION_UNSPECIFIED": 0,
		"CONTAINS_STRING":                    1,
		"NOT_CONTAINS_STRING":                2,
		"MATCHES_REGEX":                      3,
		"NOT_MATCHES_REGEX":                  4,
	}
)

Enum value maps for UptimeCheckConfig_ContentMatcher_ContentMatcherOption.

View Source
var File_google_monitoring_v3_alert_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_alert_service_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_common_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_dropped_labels_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_group_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_group_service_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_metric_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_metric_service_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_mutation_record_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_notification_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_notification_service_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_service_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_service_service_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_span_context_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_uptime_proto protoreflect.FileDescriptor
View Source
var File_google_monitoring_v3_uptime_service_proto protoreflect.FileDescriptor

Functions

func RegisterAlertPolicyServiceServer

func RegisterAlertPolicyServiceServer(s *grpc.Server, srv AlertPolicyServiceServer)

func RegisterGroupServiceServer

func RegisterGroupServiceServer(s *grpc.Server, srv GroupServiceServer)

func RegisterMetricServiceServer

func RegisterMetricServiceServer(s *grpc.Server, srv MetricServiceServer)

func RegisterNotificationChannelServiceServer

func RegisterNotificationChannelServiceServer(s *grpc.Server, srv NotificationChannelServiceServer)

func RegisterServiceMonitoringServiceServer

func RegisterServiceMonitoringServiceServer(s *grpc.Server, srv ServiceMonitoringServiceServer)

func RegisterUptimeCheckServiceServer

func RegisterUptimeCheckServiceServer(s *grpc.Server, srv UptimeCheckServiceServer)

Types

type Aggregation

type Aggregation struct {

	// The `alignment_period` specifies a time interval, in seconds, that is used
	// to divide the data in all the
	// [time series][google.monitoring.v3.TimeSeries] into consistent blocks of
	// time. This will be done before the per-series aligner can be applied to
	// the data.
	//
	// The value must be at least 60 seconds. If a per-series aligner other than
	// `ALIGN_NONE` is specified, this field is required or an error is returned.
	// If no per-series aligner is specified, or the aligner `ALIGN_NONE` is
	// specified, then this field is ignored.
	AlignmentPeriod *durationpb.Duration `protobuf:"bytes,1,opt,name=alignment_period,json=alignmentPeriod,proto3" json:"alignment_period,omitempty"`
	// An `Aligner` describes how to bring the data points in a single
	// time series into temporal alignment. Except for `ALIGN_NONE`, all
	// alignments cause all the data points in an `alignment_period` to be
	// mathematically grouped together, resulting in a single data point for
	// each `alignment_period` with end timestamp at the end of the period.
	//
	// Not all alignment operations may be applied to all time series. The valid
	// choices depend on the `metric_kind` and `value_type` of the original time
	// series. Alignment can change the `metric_kind` or the `value_type` of
	// the time series.
	//
	// Time series data must be aligned in order to perform cross-time
	// series reduction. If `cross_series_reducer` is specified, then
	// `per_series_aligner` must be specified and not equal to `ALIGN_NONE`
	// and `alignment_period` must be specified; otherwise, an error is
	// returned.
	PerSeriesAligner Aggregation_Aligner `` /* 158-byte string literal not displayed */
	// The reduction operation to be used to combine time series into a single
	// time series, where the value of each data point in the resulting series is
	// a function of all the already aligned values in the input time series.
	//
	// Not all reducer operations can be applied to all time series. The valid
	// choices depend on the `metric_kind` and the `value_type` of the original
	// time series. Reduction can yield a time series with a different
	// `metric_kind` or `value_type` than the input time series.
	//
	// Time series data must first be aligned (see `per_series_aligner`) in order
	// to perform cross-time series reduction. If `cross_series_reducer` is
	// specified, then `per_series_aligner` must be specified, and must not be
	// `ALIGN_NONE`. An `alignment_period` must also be specified; otherwise, an
	// error is returned.
	CrossSeriesReducer Aggregation_Reducer `` /* 164-byte string literal not displayed */
	// The set of fields to preserve when `cross_series_reducer` is
	// specified. The `group_by_fields` determine how the time series are
	// partitioned into subsets prior to applying the aggregation
	// operation. Each subset contains time series that have the same
	// value for each of the grouping fields. Each individual time
	// series is a member of exactly one subset. The
	// `cross_series_reducer` is applied to each subset of time series.
	// It is not possible to reduce across different resource types, so
	// this field implicitly contains `resource.type`.  Fields not
	// specified in `group_by_fields` are aggregated away.  If
	// `group_by_fields` is not specified and all the time series have
	// the same resource type, then the time series are aggregated into
	// a single output time series. If `cross_series_reducer` is not
	// defined, this field is ignored.
	GroupByFields []string `protobuf:"bytes,5,rep,name=group_by_fields,json=groupByFields,proto3" json:"group_by_fields,omitempty"`
	// contains filtered or unexported fields
}

Describes how to combine multiple time series to provide a different view of the data. Aggregation of time series is done in two steps. First, each time series in the set is _aligned_ to the same time interval boundaries, then the set of time series is optionally _reduced_ in number.

Alignment consists of applying the `per_series_aligner` operation to each time series after its data has been divided into regular `alignment_period` time intervals. This process takes _all_ of the data points in an alignment period, applies a mathematical transformation such as averaging, minimum, maximum, delta, etc., and converts them into a single data point per period.

Reduction is when the aligned and transformed time series can optionally be combined, reducing the number of time series through similar mathematical transformations. Reduction involves applying a `cross_series_reducer` to all the time series, optionally sorting the time series into subsets with `group_by_fields`, and applying the reducer to each subset.

The raw time series data can contain a huge amount of information from multiple sources. Alignment and reduction transforms this mass of data into a more manageable and representative collection of data, for example "the 95% latency across the average of all tasks in a cluster". This representative data can be more easily graphed and comprehended, and the individual time series data is still available for later drilldown. For more details, see [Filtering and aggregation](https://cloud.google.com/monitoring/api/v3/aggregation).

func (*Aggregation) Descriptor

func (*Aggregation) Descriptor() ([]byte, []int)

Deprecated: Use Aggregation.ProtoReflect.Descriptor instead.

func (*Aggregation) GetAlignmentPeriod

func (x *Aggregation) GetAlignmentPeriod() *durationpb.Duration

func (*Aggregation) GetCrossSeriesReducer

func (x *Aggregation) GetCrossSeriesReducer() Aggregation_Reducer

func (*Aggregation) GetGroupByFields

func (x *Aggregation) GetGroupByFields() []string

func (*Aggregation) GetPerSeriesAligner

func (x *Aggregation) GetPerSeriesAligner() Aggregation_Aligner

func (*Aggregation) ProtoMessage

func (*Aggregation) ProtoMessage()

func (*Aggregation) ProtoReflect

func (x *Aggregation) ProtoReflect() protoreflect.Message

func (*Aggregation) Reset

func (x *Aggregation) Reset()

func (*Aggregation) String

func (x *Aggregation) String() string

type Aggregation_Aligner

type Aggregation_Aligner int32

The `Aligner` specifies the operation that will be applied to the data points in each alignment period in a time series. Except for `ALIGN_NONE`, which specifies that no operation be applied, each alignment operation replaces the set of data values in each alignment period with a single value: the result of applying the operation to the data values. An aligned time series has a single data value at the end of each `alignment_period`.

An alignment operation can change the data type of the values, too. For example, if you apply a counting operation to boolean values, the data `value_type` in the original time series is `BOOLEAN`, but the `value_type` in the aligned result is `INT64`.

const (
	// No alignment. Raw data is returned. Not valid if cross-series reduction
	// is requested. The `value_type` of the result is the same as the
	// `value_type` of the input.
	Aggregation_ALIGN_NONE Aggregation_Aligner = 0
	// Align and convert to
	// [DELTA][google.api.MetricDescriptor.MetricKind.DELTA].
	// The output is `delta = y1 - y0`.
	//
	// This alignment is valid for
	// [CUMULATIVE][google.api.MetricDescriptor.MetricKind.CUMULATIVE] and
	// `DELTA` metrics. If the selected alignment period results in periods
	// with no data, then the aligned value for such a period is created by
	// interpolation. The `value_type`  of the aligned result is the same as
	// the `value_type` of the input.
	Aggregation_ALIGN_DELTA Aggregation_Aligner = 1
	// Align and convert to a rate. The result is computed as
	// `rate = (y1 - y0)/(t1 - t0)`, or "delta over time".
	// Think of this aligner as providing the slope of the line that passes
	// through the value at the start and at the end of the `alignment_period`.
	//
	// This aligner is valid for `CUMULATIVE`
	// and `DELTA` metrics with numeric values. If the selected alignment
	// period results in periods with no data, then the aligned value for
	// such a period is created by interpolation. The output is a `GAUGE`
	// metric with `value_type` `DOUBLE`.
	//
	// If, by "rate", you mean "percentage change", see the
	// `ALIGN_PERCENT_CHANGE` aligner instead.
	Aggregation_ALIGN_RATE Aggregation_Aligner = 2
	// Align by interpolating between adjacent points around the alignment
	// period boundary. This aligner is valid for `GAUGE` metrics with
	// numeric values. The `value_type` of the aligned result is the same as the
	// `value_type` of the input.
	Aggregation_ALIGN_INTERPOLATE Aggregation_Aligner = 3
	// Align by moving the most recent data point before the end of the
	// alignment period to the boundary at the end of the alignment
	// period. This aligner is valid for `GAUGE` metrics. The `value_type` of
	// the aligned result is the same as the `value_type` of the input.
	Aggregation_ALIGN_NEXT_OLDER Aggregation_Aligner = 4
	// Align the time series by returning the minimum value in each alignment
	// period. This aligner is valid for `GAUGE` and `DELTA` metrics with
	// numeric values. The `value_type` of the aligned result is the same as
	// the `value_type` of the input.
	Aggregation_ALIGN_MIN Aggregation_Aligner = 10
	// Align the time series by returning the maximum value in each alignment
	// period. This aligner is valid for `GAUGE` and `DELTA` metrics with
	// numeric values. The `value_type` of the aligned result is the same as
	// the `value_type` of the input.
	Aggregation_ALIGN_MAX Aggregation_Aligner = 11
	// Align the time series by returning the mean value in each alignment
	// period. This aligner is valid for `GAUGE` and `DELTA` metrics with
	// numeric values. The `value_type` of the aligned result is `DOUBLE`.
	Aggregation_ALIGN_MEAN Aggregation_Aligner = 12
	// Align the time series by returning the number of values in each alignment
	// period. This aligner is valid for `GAUGE` and `DELTA` metrics with
	// numeric or Boolean values. The `value_type` of the aligned result is
	// `INT64`.
	Aggregation_ALIGN_COUNT Aggregation_Aligner = 13
	// Align the time series by returning the sum of the values in each
	// alignment period. This aligner is valid for `GAUGE` and `DELTA`
	// metrics with numeric and distribution values. The `value_type` of the
	// aligned result is the same as the `value_type` of the input.
	Aggregation_ALIGN_SUM Aggregation_Aligner = 14
	// Align the time series by returning the standard deviation of the values
	// in each alignment period. This aligner is valid for `GAUGE` and
	// `DELTA` metrics with numeric values. The `value_type` of the output is
	// `DOUBLE`.
	Aggregation_ALIGN_STDDEV Aggregation_Aligner = 15
	// Align the time series by returning the number of `True` values in
	// each alignment period. This aligner is valid for `GAUGE` metrics with
	// Boolean values. The `value_type` of the output is `INT64`.
	Aggregation_ALIGN_COUNT_TRUE Aggregation_Aligner = 16
	// Align the time series by returning the number of `False` values in
	// each alignment period. This aligner is valid for `GAUGE` metrics with
	// Boolean values. The `value_type` of the output is `INT64`.
	Aggregation_ALIGN_COUNT_FALSE Aggregation_Aligner = 24
	// Align the time series by returning the ratio of the number of `True`
	// values to the total number of values in each alignment period. This
	// aligner is valid for `GAUGE` metrics with Boolean values. The output
	// value is in the range [0.0, 1.0] and has `value_type` `DOUBLE`.
	Aggregation_ALIGN_FRACTION_TRUE Aggregation_Aligner = 17
	// Align the time series by using [percentile
	// aggregation](https://en.wikipedia.org/wiki/Percentile). The resulting
	// data point in each alignment period is the 99th percentile of all data
	// points in the period. This aligner is valid for `GAUGE` and `DELTA`
	// metrics with distribution values. The output is a `GAUGE` metric with
	// `value_type` `DOUBLE`.
	Aggregation_ALIGN_PERCENTILE_99 Aggregation_Aligner = 18
	// Align the time series by using [percentile
	// aggregation](https://en.wikipedia.org/wiki/Percentile). The resulting
	// data point in each alignment period is the 95th percentile of all data
	// points in the period. This aligner is valid for `GAUGE` and `DELTA`
	// metrics with distribution values. The output is a `GAUGE` metric with
	// `value_type` `DOUBLE`.
	Aggregation_ALIGN_PERCENTILE_95 Aggregation_Aligner = 19
	// Align the time series by using [percentile
	// aggregation](https://en.wikipedia.org/wiki/Percentile). The resulting
	// data point in each alignment period is the 50th percentile of all data
	// points in the period. This aligner is valid for `GAUGE` and `DELTA`
	// metrics with distribution values. The output is a `GAUGE` metric with
	// `value_type` `DOUBLE`.
	Aggregation_ALIGN_PERCENTILE_50 Aggregation_Aligner = 20
	// Align the time series by using [percentile
	// aggregation](https://en.wikipedia.org/wiki/Percentile). The resulting
	// data point in each alignment period is the 5th percentile of all data
	// points in the period. This aligner is valid for `GAUGE` and `DELTA`
	// metrics with distribution values. The output is a `GAUGE` metric with
	// `value_type` `DOUBLE`.
	Aggregation_ALIGN_PERCENTILE_05 Aggregation_Aligner = 21
	// Align and convert to a percentage change. This aligner is valid for
	// `GAUGE` and `DELTA` metrics with numeric values. This alignment returns
	// `((current - previous)/previous) * 100`, where the value of `previous` is
	// determined based on the `alignment_period`.
	//
	// If the values of `current` and `previous` are both 0, then the returned
	// value is 0. If only `previous` is 0, the returned value is infinity.
	//
	// A 10-minute moving mean is computed at each point of the alignment period
	// prior to the above calculation to smooth the metric and prevent false
	// positives from very short-lived spikes. The moving mean is only
	// applicable for data whose values are `>= 0`. Any values `< 0` are
	// treated as a missing datapoint, and are ignored. While `DELTA`
	// metrics are accepted by this alignment, special care should be taken that
	// the values for the metric will always be positive. The output is a
	// `GAUGE` metric with `value_type` `DOUBLE`.
	Aggregation_ALIGN_PERCENT_CHANGE Aggregation_Aligner = 23
)

func (Aggregation_Aligner) Descriptor

func (Aggregation_Aligner) Enum

func (Aggregation_Aligner) EnumDescriptor

func (Aggregation_Aligner) EnumDescriptor() ([]byte, []int)

Deprecated: Use Aggregation_Aligner.Descriptor instead.

func (Aggregation_Aligner) Number

func (Aggregation_Aligner) String

func (x Aggregation_Aligner) String() string

func (Aggregation_Aligner) Type

type Aggregation_Reducer

type Aggregation_Reducer int32

A Reducer operation describes how to aggregate data points from multiple time series into a single time series, where the value of each data point in the resulting series is a function of all the already aligned values in the input time series.

const (
	// No cross-time series reduction. The output of the `Aligner` is
	// returned.
	Aggregation_REDUCE_NONE Aggregation_Reducer = 0
	// Reduce by computing the mean value across time series for each
	// alignment period. This reducer is valid for
	// [DELTA][google.api.MetricDescriptor.MetricKind.DELTA] and
	// [GAUGE][google.api.MetricDescriptor.MetricKind.GAUGE] metrics with
	// numeric or distribution values. The `value_type` of the output is
	// [DOUBLE][google.api.MetricDescriptor.ValueType.DOUBLE].
	Aggregation_REDUCE_MEAN Aggregation_Reducer = 1
	// Reduce by computing the minimum value across time series for each
	// alignment period. This reducer is valid for `DELTA` and `GAUGE` metrics
	// with numeric values. The `value_type` of the output is the same as the
	// `value_type` of the input.
	Aggregation_REDUCE_MIN Aggregation_Reducer = 2
	// Reduce by computing the maximum value across time series for each
	// alignment period. This reducer is valid for `DELTA` and `GAUGE` metrics
	// with numeric values. The `value_type` of the output is the same as the
	// `value_type` of the input.
	Aggregation_REDUCE_MAX Aggregation_Reducer = 3
	// Reduce by computing the sum across time series for each
	// alignment period. This reducer is valid for `DELTA` and `GAUGE` metrics
	// with numeric and distribution values. The `value_type` of the output is
	// the same as the `value_type` of the input.
	Aggregation_REDUCE_SUM Aggregation_Reducer = 4
	// Reduce by computing the standard deviation across time series
	// for each alignment period. This reducer is valid for `DELTA` and
	// `GAUGE` metrics with numeric or distribution values. The `value_type`
	// of the output is `DOUBLE`.
	Aggregation_REDUCE_STDDEV Aggregation_Reducer = 5
	// Reduce by computing the number of data points across time series
	// for each alignment period. This reducer is valid for `DELTA` and
	// `GAUGE` metrics of numeric, Boolean, distribution, and string
	// `value_type`. The `value_type` of the output is `INT64`.
	Aggregation_REDUCE_COUNT Aggregation_Reducer = 6
	// Reduce by computing the number of `True`-valued data points across time
	// series for each alignment period. This reducer is valid for `DELTA` and
	// `GAUGE` metrics of Boolean `value_type`. The `value_type` of the output
	// is `INT64`.
	Aggregation_REDUCE_COUNT_TRUE Aggregation_Reducer = 7
	// Reduce by computing the number of `False`-valued data points across time
	// series for each alignment period. This reducer is valid for `DELTA` and
	// `GAUGE` metrics of Boolean `value_type`. The `value_type` of the output
	// is `INT64`.
	Aggregation_REDUCE_COUNT_FALSE Aggregation_Reducer = 15
	// Reduce by computing the ratio of the number of `True`-valued data points
	// to the total number of data points for each alignment period. This
	// reducer is valid for `DELTA` and `GAUGE` metrics of Boolean `value_type`.
	// The output value is in the range [0.0, 1.0] and has `value_type`
	// `DOUBLE`.
	Aggregation_REDUCE_FRACTION_TRUE Aggregation_Reducer = 8
	// Reduce by computing the [99th
	// percentile](https://en.wikipedia.org/wiki/Percentile) of data points
	// across time series for each alignment period. This reducer is valid for
	// `GAUGE` and `DELTA` metrics of numeric and distribution type. The value
	// of the output is `DOUBLE`.
	Aggregation_REDUCE_PERCENTILE_99 Aggregation_Reducer = 9
	// Reduce by computing the [95th
	// percentile](https://en.wikipedia.org/wiki/Percentile) of data points
	// across time series for each alignment period. This reducer is valid for
	// `GAUGE` and `DELTA` metrics of numeric and distribution type. The value
	// of the output is `DOUBLE`.
	Aggregation_REDUCE_PERCENTILE_95 Aggregation_Reducer = 10
	// Reduce by computing the [50th
	// percentile](https://en.wikipedia.org/wiki/Percentile) of data points
	// across time series for each alignment period. This reducer is valid for
	// `GAUGE` and `DELTA` metrics of numeric and distribution type. The value
	// of the output is `DOUBLE`.
	Aggregation_REDUCE_PERCENTILE_50 Aggregation_Reducer = 11
	// Reduce by computing the [5th
	// percentile](https://en.wikipedia.org/wiki/Percentile) of data points
	// across time series for each alignment period. This reducer is valid for
	// `GAUGE` and `DELTA` metrics of numeric and distribution type. The value
	// of the output is `DOUBLE`.
	Aggregation_REDUCE_PERCENTILE_05 Aggregation_Reducer = 12
)

func (Aggregation_Reducer) Descriptor

func (Aggregation_Reducer) Enum

func (Aggregation_Reducer) EnumDescriptor

func (Aggregation_Reducer) EnumDescriptor() ([]byte, []int)

Deprecated: Use Aggregation_Reducer.Descriptor instead.

func (Aggregation_Reducer) Number

func (Aggregation_Reducer) String

func (x Aggregation_Reducer) String() string

func (Aggregation_Reducer) Type

type AlertPolicy

type AlertPolicy struct {

	// Required if the policy exists. The resource name for this policy. The
	// format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/alertPolicies/[ALERT_POLICY_ID]
	//
	// `[ALERT_POLICY_ID]` is assigned by Stackdriver Monitoring when the policy
	// is created.  When calling the
	// [alertPolicies.create][google.monitoring.v3.AlertPolicyService.CreateAlertPolicy]
	// method, do not include the `name` field in the alerting policy passed as
	// part of the request.
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// A short name or phrase used to identify the policy in dashboards,
	// notifications, and incidents. To avoid confusion, don't use the same
	// display name for multiple policies in the same project. The name is
	// limited to 512 Unicode characters.
	DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"`
	// Documentation that is included with notifications and incidents related to
	// this policy. Best practice is for the documentation to include information
	// to help responders understand, mitigate, escalate, and correct the
	// underlying problems detected by the alerting policy. Notification channels
	// that have limited capacity might not show this documentation.
	Documentation *AlertPolicy_Documentation `protobuf:"bytes,13,opt,name=documentation,proto3" json:"documentation,omitempty"`
	// User-supplied key/value data to be used for organizing and
	// identifying the `AlertPolicy` objects.
	//
	// The field can contain up to 64 entries. Each key and value is limited to
	// 63 Unicode characters or 128 bytes, whichever is smaller. Labels and
	// values can contain only lowercase letters, numerals, underscores, and
	// dashes. Keys must begin with a letter.
	UserLabels map[string]string `` /* 180-byte string literal not displayed */
	// A list of conditions for the policy. The conditions are combined by AND or
	// OR according to the `combiner` field. If the combined conditions evaluate
	// to true, then an incident is created. A policy can have from one to six
	// conditions.
	// If `condition_time_series_query_language` is present, it must be the only
	// `condition`.
	Conditions []*AlertPolicy_Condition `protobuf:"bytes,12,rep,name=conditions,proto3" json:"conditions,omitempty"`
	// How to combine the results of multiple conditions to determine if an
	// incident should be opened.
	// If `condition_time_series_query_language` is present, this must be
	// `COMBINE_UNSPECIFIED`.
	Combiner AlertPolicy_ConditionCombinerType `` /* 130-byte string literal not displayed */
	// Whether or not the policy is enabled. On write, the default interpretation
	// if unset is that the policy is enabled. On read, clients should not make
	// any assumption about the state if it has not been populated. The
	// field should always be populated on List and Get operations, unless
	// a field projection has been specified that strips it out.
	Enabled *wrapperspb.BoolValue `protobuf:"bytes,17,opt,name=enabled,proto3" json:"enabled,omitempty"`
	// Read-only description of how the alert policy is invalid. OK if the alert
	// policy is valid. If not OK, the alert policy will not generate incidents.
	Validity *status.Status `protobuf:"bytes,18,opt,name=validity,proto3" json:"validity,omitempty"`
	// Identifies the notification channels to which notifications should be sent
	// when incidents are opened or closed or when new violations occur on
	// an already opened incident. Each element of this array corresponds to
	// the `name` field in each of the
	// [`NotificationChannel`][google.monitoring.v3.NotificationChannel]
	// objects that are returned from the [`ListNotificationChannels`]
	// [google.monitoring.v3.NotificationChannelService.ListNotificationChannels]
	// method. The format of the entries in this field is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/notificationChannels/[CHANNEL_ID]
	NotificationChannels []string `protobuf:"bytes,14,rep,name=notification_channels,json=notificationChannels,proto3" json:"notification_channels,omitempty"`
	// A read-only record of the creation of the alerting policy. If provided
	// in a call to create or update, this field will be ignored.
	CreationRecord *MutationRecord `protobuf:"bytes,10,opt,name=creation_record,json=creationRecord,proto3" json:"creation_record,omitempty"`
	// A read-only record of the most recent change to the alerting policy. If
	// provided in a call to create or update, this field will be ignored.
	MutationRecord *MutationRecord `protobuf:"bytes,11,opt,name=mutation_record,json=mutationRecord,proto3" json:"mutation_record,omitempty"`
	// contains filtered or unexported fields
}

A description of the conditions under which some aspect of your system is considered to be "unhealthy" and the ways to notify people or services about this state. For an overview of alert policies, see [Introduction to Alerting](https://cloud.google.com/monitoring/alerts/).

func (*AlertPolicy) Descriptor

func (*AlertPolicy) Descriptor() ([]byte, []int)

Deprecated: Use AlertPolicy.ProtoReflect.Descriptor instead.

func (*AlertPolicy) GetCombiner

func (*AlertPolicy) GetConditions

func (x *AlertPolicy) GetConditions() []*AlertPolicy_Condition

func (*AlertPolicy) GetCreationRecord

func (x *AlertPolicy) GetCreationRecord() *MutationRecord

func (*AlertPolicy) GetDisplayName

func (x *AlertPolicy) GetDisplayName() string

func (*AlertPolicy) GetDocumentation

func (x *AlertPolicy) GetDocumentation() *AlertPolicy_Documentation

func (*AlertPolicy) GetEnabled

func (x *AlertPolicy) GetEnabled() *wrapperspb.BoolValue

func (*AlertPolicy) GetMutationRecord

func (x *AlertPolicy) GetMutationRecord() *MutationRecord

func (*AlertPolicy) GetName

func (x *AlertPolicy) GetName() string

func (*AlertPolicy) GetNotificationChannels

func (x *AlertPolicy) GetNotificationChannels() []string

func (*AlertPolicy) GetUserLabels

func (x *AlertPolicy) GetUserLabels() map[string]string

func (*AlertPolicy) GetValidity

func (x *AlertPolicy) GetValidity() *status.Status

func (*AlertPolicy) ProtoMessage

func (*AlertPolicy) ProtoMessage()

func (*AlertPolicy) ProtoReflect

func (x *AlertPolicy) ProtoReflect() protoreflect.Message

func (*AlertPolicy) Reset

func (x *AlertPolicy) Reset()

func (*AlertPolicy) String

func (x *AlertPolicy) String() string

type AlertPolicyServiceClient

type AlertPolicyServiceClient interface {
	// Lists the existing alerting policies for the project.
	ListAlertPolicies(ctx context.Context, in *ListAlertPoliciesRequest, opts ...grpc.CallOption) (*ListAlertPoliciesResponse, error)
	// Gets a single alerting policy.
	GetAlertPolicy(ctx context.Context, in *GetAlertPolicyRequest, opts ...grpc.CallOption) (*AlertPolicy, error)
	// Creates a new alerting policy.
	CreateAlertPolicy(ctx context.Context, in *CreateAlertPolicyRequest, opts ...grpc.CallOption) (*AlertPolicy, error)
	// Deletes an alerting policy.
	DeleteAlertPolicy(ctx context.Context, in *DeleteAlertPolicyRequest, opts ...grpc.CallOption) (*emptypb.Empty, error)
	// Updates an alerting policy. You can either replace the entire policy with
	// a new one or replace only certain fields in the current alerting policy by
	// specifying the fields to be updated via `updateMask`. Returns the
	// updated alerting policy.
	UpdateAlertPolicy(ctx context.Context, in *UpdateAlertPolicyRequest, opts ...grpc.CallOption) (*AlertPolicy, error)
}

AlertPolicyServiceClient is the client API for AlertPolicyService service.

For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.

type AlertPolicyServiceServer

type AlertPolicyServiceServer interface {
	// Lists the existing alerting policies for the project.
	ListAlertPolicies(context.Context, *ListAlertPoliciesRequest) (*ListAlertPoliciesResponse, error)
	// Gets a single alerting policy.
	GetAlertPolicy(context.Context, *GetAlertPolicyRequest) (*AlertPolicy, error)
	// Creates a new alerting policy.
	CreateAlertPolicy(context.Context, *CreateAlertPolicyRequest) (*AlertPolicy, error)
	// Deletes an alerting policy.
	DeleteAlertPolicy(context.Context, *DeleteAlertPolicyRequest) (*emptypb.Empty, error)
	// Updates an alerting policy. You can either replace the entire policy with
	// a new one or replace only certain fields in the current alerting policy by
	// specifying the fields to be updated via `updateMask`. Returns the
	// updated alerting policy.
	UpdateAlertPolicy(context.Context, *UpdateAlertPolicyRequest) (*AlertPolicy, error)
}

AlertPolicyServiceServer is the server API for AlertPolicyService service.

type AlertPolicy_Condition

type AlertPolicy_Condition struct {

	// Required if the condition exists. The unique resource name for this
	// condition. Its format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/alertPolicies/[POLICY_ID]/conditions/[CONDITION_ID]
	//
	// `[CONDITION_ID]` is assigned by Stackdriver Monitoring when the
	// condition is created as part of a new or updated alerting policy.
	//
	// When calling the
	// [alertPolicies.create][google.monitoring.v3.AlertPolicyService.CreateAlertPolicy]
	// method, do not include the `name` field in the conditions of the
	// requested alerting policy. Stackdriver Monitoring creates the
	// condition identifiers and includes them in the new policy.
	//
	// When calling the
	// [alertPolicies.update][google.monitoring.v3.AlertPolicyService.UpdateAlertPolicy]
	// method to update a policy, including a condition `name` causes the
	// existing condition to be updated. Conditions without names are added to
	// the updated policy. Existing conditions are deleted if they are not
	// updated.
	//
	// Best practice is to preserve `[CONDITION_ID]` if you make only small
	// changes, such as those to condition thresholds, durations, or trigger
	// values.  Otherwise, treat the change as a new condition and let the
	// existing condition be deleted.
	Name string `protobuf:"bytes,12,opt,name=name,proto3" json:"name,omitempty"`
	// A short name or phrase used to identify the condition in dashboards,
	// notifications, and incidents. To avoid confusion, don't use the same
	// display name for multiple conditions in the same policy.
	DisplayName string `protobuf:"bytes,6,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"`
	// Only one of the following condition types will be specified.
	//
	// Types that are assignable to Condition:
	//	*AlertPolicy_Condition_ConditionThreshold
	//	*AlertPolicy_Condition_ConditionAbsent
	Condition isAlertPolicy_Condition_Condition `protobuf_oneof:"condition"`
	// contains filtered or unexported fields
}

A condition is a true/false test that determines when an alerting policy should open an incident. If a condition evaluates to true, it signifies that something is wrong.

func (*AlertPolicy_Condition) Descriptor

func (*AlertPolicy_Condition) Descriptor() ([]byte, []int)

Deprecated: Use AlertPolicy_Condition.ProtoReflect.Descriptor instead.

func (*AlertPolicy_Condition) GetCondition

func (m *AlertPolicy_Condition) GetCondition() isAlertPolicy_Condition_Condition

func (*AlertPolicy_Condition) GetConditionAbsent

func (*AlertPolicy_Condition) GetConditionThreshold

func (*AlertPolicy_Condition) GetDisplayName

func (x *AlertPolicy_Condition) GetDisplayName() string

func (*AlertPolicy_Condition) GetName

func (x *AlertPolicy_Condition) GetName() string

func (*AlertPolicy_Condition) ProtoMessage

func (*AlertPolicy_Condition) ProtoMessage()

func (*AlertPolicy_Condition) ProtoReflect

func (x *AlertPolicy_Condition) ProtoReflect() protoreflect.Message

func (*AlertPolicy_Condition) Reset

func (x *AlertPolicy_Condition) Reset()

func (*AlertPolicy_Condition) String

func (x *AlertPolicy_Condition) String() string

type AlertPolicy_ConditionCombinerType

type AlertPolicy_ConditionCombinerType int32

Operators for combining conditions.

const (
	// An unspecified combiner.
	AlertPolicy_COMBINE_UNSPECIFIED AlertPolicy_ConditionCombinerType = 0
	// Combine conditions using the logical `AND` operator. An
	// incident is created only if all the conditions are met
	// simultaneously. This combiner is satisfied if all conditions are
	// met, even if they are met on completely different resources.
	AlertPolicy_AND AlertPolicy_ConditionCombinerType = 1
	// Combine conditions using the logical `OR` operator. An incident
	// is created if any of the listed conditions is met.
	AlertPolicy_OR AlertPolicy_ConditionCombinerType = 2
	// Combine conditions using logical `AND` operator, but unlike the regular
	// `AND` option, an incident is created only if all conditions are met
	// simultaneously on at least one resource.
	AlertPolicy_AND_WITH_MATCHING_RESOURCE AlertPolicy_ConditionCombinerType = 3
)

func (AlertPolicy_ConditionCombinerType) Descriptor

func (AlertPolicy_ConditionCombinerType) Enum

func (AlertPolicy_ConditionCombinerType) EnumDescriptor

func (AlertPolicy_ConditionCombinerType) EnumDescriptor() ([]byte, []int)

Deprecated: Use AlertPolicy_ConditionCombinerType.Descriptor instead.

func (AlertPolicy_ConditionCombinerType) Number

func (AlertPolicy_ConditionCombinerType) String

func (AlertPolicy_ConditionCombinerType) Type

type AlertPolicy_Condition_ConditionAbsent

type AlertPolicy_Condition_ConditionAbsent struct {
	// A condition that checks that a time series continues to
	// receive new data points.
	ConditionAbsent *AlertPolicy_Condition_MetricAbsence `protobuf:"bytes,2,opt,name=condition_absent,json=conditionAbsent,proto3,oneof"`
}

type AlertPolicy_Condition_ConditionThreshold

type AlertPolicy_Condition_ConditionThreshold struct {
	// A condition that compares a time series against a threshold.
	ConditionThreshold *AlertPolicy_Condition_MetricThreshold `protobuf:"bytes,1,opt,name=condition_threshold,json=conditionThreshold,proto3,oneof"`
}

type AlertPolicy_Condition_MetricAbsence

type AlertPolicy_Condition_MetricAbsence struct {

	// A [filter](https://cloud.google.com/monitoring/api/v3/filters) that
	// identifies which time series should be compared with the threshold.
	//
	// The filter is similar to the one that is specified in the
	// [`ListTimeSeries`
	// request](https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list)
	// (that call is useful to verify the time series that will be retrieved /
	// processed) and must specify the metric type and optionally may contain
	// restrictions on resource type, resource labels, and metric labels.
	// This field may not exceed 2048 Unicode characters in length.
	Filter string `protobuf:"bytes,1,opt,name=filter,proto3" json:"filter,omitempty"`
	// Specifies the alignment of data points in individual time series as
	// well as how to combine the retrieved time series together (such as
	// when aggregating multiple streams on each resource to a single
	// stream for each resource or when aggregating streams across all
	// members of a group of resrouces). Multiple aggregations
	// are applied in the order specified.
	//
	// This field is similar to the one in the [`ListTimeSeries`
	// request](https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list).
	// It is advisable to use the `ListTimeSeries` method when debugging this
	// field.
	Aggregations []*Aggregation `protobuf:"bytes,5,rep,name=aggregations,proto3" json:"aggregations,omitempty"`
	// The amount of time that a time series must fail to report new
	// data to be considered failing. Currently, only values that
	// are a multiple of a minute--e.g.  60, 120, or 300
	// seconds--are supported. If an invalid value is given, an
	// error will be returned. The `Duration.nanos` field is
	// ignored.
	Duration *durationpb.Duration `protobuf:"bytes,2,opt,name=duration,proto3" json:"duration,omitempty"`
	// The number/percent of time series for which the comparison must hold
	// in order for the condition to trigger. If unspecified, then the
	// condition will trigger if the comparison is true for any of the
	// time series that have been identified by `filter` and `aggregations`.
	Trigger *AlertPolicy_Condition_Trigger `protobuf:"bytes,3,opt,name=trigger,proto3" json:"trigger,omitempty"`
	// contains filtered or unexported fields
}

A condition type that checks that monitored resources are reporting data. The configuration defines a metric and a set of monitored resources. The predicate is considered in violation when a time series for the specified metric of a monitored resource does not include any data in the specified `duration`.

func (*AlertPolicy_Condition_MetricAbsence) Descriptor

func (*AlertPolicy_Condition_MetricAbsence) Descriptor() ([]byte, []int)

Deprecated: Use AlertPolicy_Condition_MetricAbsence.ProtoReflect.Descriptor instead.

func (*AlertPolicy_Condition_MetricAbsence) GetAggregations

func (x *AlertPolicy_Condition_MetricAbsence) GetAggregations() []*Aggregation

func (*AlertPolicy_Condition_MetricAbsence) GetDuration

func (*AlertPolicy_Condition_MetricAbsence) GetFilter

func (*AlertPolicy_Condition_MetricAbsence) GetTrigger

func (*AlertPolicy_Condition_MetricAbsence) ProtoMessage

func (*AlertPolicy_Condition_MetricAbsence) ProtoMessage()

func (*AlertPolicy_Condition_MetricAbsence) ProtoReflect

func (*AlertPolicy_Condition_MetricAbsence) Reset

func (*AlertPolicy_Condition_MetricAbsence) String

type AlertPolicy_Condition_MetricThreshold

type AlertPolicy_Condition_MetricThreshold struct {

	// A [filter](https://cloud.google.com/monitoring/api/v3/filters) that
	// identifies which time series should be compared with the threshold.
	//
	// The filter is similar to the one that is specified in the
	// [`ListTimeSeries`
	// request](https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list)
	// (that call is useful to verify the time series that will be retrieved /
	// processed) and must specify the metric type and optionally may contain
	// restrictions on resource type, resource labels, and metric labels.
	// This field may not exceed 2048 Unicode characters in length.
	Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"`
	// Specifies the alignment of data points in individual time series as
	// well as how to combine the retrieved time series together (such as
	// when aggregating multiple streams on each resource to a single
	// stream for each resource or when aggregating streams across all
	// members of a group of resrouces). Multiple aggregations
	// are applied in the order specified.
	//
	// This field is similar to the one in the [`ListTimeSeries`
	// request](https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list).
	// It is advisable to use the `ListTimeSeries` method when debugging this
	// field.
	Aggregations []*Aggregation `protobuf:"bytes,8,rep,name=aggregations,proto3" json:"aggregations,omitempty"`
	// A [filter](https://cloud.google.com/monitoring/api/v3/filters) that
	// identifies a time series that should be used as the denominator of a
	// ratio that will be compared with the threshold. If a
	// `denominator_filter` is specified, the time series specified by the
	// `filter` field will be used as the numerator.
	//
	// The filter must specify the metric type and optionally may contain
	// restrictions on resource type, resource labels, and metric labels.
	// This field may not exceed 2048 Unicode characters in length.
	DenominatorFilter string `protobuf:"bytes,9,opt,name=denominator_filter,json=denominatorFilter,proto3" json:"denominator_filter,omitempty"`
	// Specifies the alignment of data points in individual time series
	// selected by `denominatorFilter` as
	// well as how to combine the retrieved time series together (such as
	// when aggregating multiple streams on each resource to a single
	// stream for each resource or when aggregating streams across all
	// members of a group of resources).
	//
	// When computing ratios, the `aggregations` and
	// `denominator_aggregations` fields must use the same alignment period
	// and produce time series that have the same periodicity and labels.
	DenominatorAggregations []*Aggregation `` /* 131-byte string literal not displayed */
	// The comparison to apply between the time series (indicated by `filter`
	// and `aggregation`) and the threshold (indicated by `threshold_value`).
	// The comparison is applied on each time series, with the time series
	// on the left-hand side and the threshold on the right-hand side.
	//
	// Only `COMPARISON_LT` and `COMPARISON_GT` are supported currently.
	Comparison ComparisonType `protobuf:"varint,4,opt,name=comparison,proto3,enum=google.monitoring.v3.ComparisonType" json:"comparison,omitempty"`
	// A value against which to compare the time series.
	ThresholdValue float64 `protobuf:"fixed64,5,opt,name=threshold_value,json=thresholdValue,proto3" json:"threshold_value,omitempty"`
	// The amount of time that a time series must violate the
	// threshold to be considered failing. Currently, only values
	// that are a multiple of a minute--e.g., 0, 60, 120, or 300
	// seconds--are supported. If an invalid value is given, an
	// error will be returned. When choosing a duration, it is useful to
	// keep in mind the frequency of the underlying time series data
	// (which may also be affected by any alignments specified in the
	// `aggregations` field); a good duration is long enough so that a single
	// outlier does not generate spurious alerts, but short enough that
	// unhealthy states are detected and alerted on quickly.
	Duration *durationpb.Duration `protobuf:"bytes,6,opt,name=duration,proto3" json:"duration,omitempty"`
	// The number/percent of time series for which the comparison must hold
	// in order for the condition to trigger. If unspecified, then the
	// condition will trigger if the comparison is true for any of the
	// time series that have been identified by `filter` and `aggregations`,
	// or by the ratio, if `denominator_filter` and `denominator_aggregations`
	// are specified.
	Trigger *AlertPolicy_Condition_Trigger `protobuf:"bytes,7,opt,name=trigger,proto3" json:"trigger,omitempty"`
	// contains filtered or unexported fields
}

A condition type that compares a collection of time series against a threshold.

func (*AlertPolicy_Condition_MetricThreshold) Descriptor

func (*AlertPolicy_Condition_MetricThreshold) Descriptor() ([]byte, []int)

Deprecated: Use AlertPolicy_Condition_MetricThreshold.ProtoReflect.Descriptor instead.

func (*AlertPolicy_Condition_MetricThreshold) GetAggregations

func (x *AlertPolicy_Condition_MetricThreshold) GetAggregations() []*Aggregation

func (*AlertPolicy_Condition_MetricThreshold) GetComparison

func (*AlertPolicy_Condition_MetricThreshold) GetDenominatorAggregations

func (x *AlertPolicy_Condition_MetricThreshold) GetDenominatorAggregations() []*Aggregation

func (*AlertPolicy_Condition_MetricThreshold) GetDenominatorFilter

func (x *AlertPolicy_Condition_MetricThreshold) GetDenominatorFilter() string

func (*AlertPolicy_Condition_MetricThreshold) GetDuration

func (*AlertPolicy_Condition_MetricThreshold) GetFilter

func (*AlertPolicy_Condition_MetricThreshold) GetThresholdValue

func (x *AlertPolicy_Condition_MetricThreshold) GetThresholdValue() float64

func (*AlertPolicy_Condition_MetricThreshold) GetTrigger

func (*AlertPolicy_Condition_MetricThreshold) ProtoMessage

func (*AlertPolicy_Condition_MetricThreshold) ProtoMessage()

func (*AlertPolicy_Condition_MetricThreshold) ProtoReflect

func (*AlertPolicy_Condition_MetricThreshold) Reset

func (*AlertPolicy_Condition_MetricThreshold) String

type AlertPolicy_Condition_Trigger

type AlertPolicy_Condition_Trigger struct {

	// A type of trigger.
	//
	// Types that are assignable to Type:
	//	*AlertPolicy_Condition_Trigger_Count
	//	*AlertPolicy_Condition_Trigger_Percent
	Type isAlertPolicy_Condition_Trigger_Type `protobuf_oneof:"type"`
	// contains filtered or unexported fields
}

Specifies how many time series must fail a predicate to trigger a condition. If not specified, then a `{count: 1}` trigger is used.

func (*AlertPolicy_Condition_Trigger) Descriptor

func (*AlertPolicy_Condition_Trigger) Descriptor() ([]byte, []int)

Deprecated: Use AlertPolicy_Condition_Trigger.ProtoReflect.Descriptor instead.

func (*AlertPolicy_Condition_Trigger) GetCount

func (x *AlertPolicy_Condition_Trigger) GetCount() int32

func (*AlertPolicy_Condition_Trigger) GetPercent

func (x *AlertPolicy_Condition_Trigger) GetPercent() float64

func (*AlertPolicy_Condition_Trigger) GetType

func (m *AlertPolicy_Condition_Trigger) GetType() isAlertPolicy_Condition_Trigger_Type

func (*AlertPolicy_Condition_Trigger) ProtoMessage

func (*AlertPolicy_Condition_Trigger) ProtoMessage()

func (*AlertPolicy_Condition_Trigger) ProtoReflect

func (*AlertPolicy_Condition_Trigger) Reset

func (x *AlertPolicy_Condition_Trigger) Reset()

func (*AlertPolicy_Condition_Trigger) String

type AlertPolicy_Condition_Trigger_Count

type AlertPolicy_Condition_Trigger_Count struct {
	// The absolute number of time series that must fail
	// the predicate for the condition to be triggered.
	Count int32 `protobuf:"varint,1,opt,name=count,proto3,oneof"`
}

type AlertPolicy_Condition_Trigger_Percent

type AlertPolicy_Condition_Trigger_Percent struct {
	// The percentage of time series that must fail the
	// predicate for the condition to be triggered.
	Percent float64 `protobuf:"fixed64,2,opt,name=percent,proto3,oneof"`
}

type AlertPolicy_Documentation

type AlertPolicy_Documentation struct {

	// The text of the documentation, interpreted according to `mime_type`.
	// The content may not exceed 8,192 Unicode characters and may not exceed
	// more than 10,240 bytes when encoded in UTF-8 format, whichever is
	// smaller.
	Content string `protobuf:"bytes,1,opt,name=content,proto3" json:"content,omitempty"`
	// The format of the `content` field. Presently, only the value
	// `"text/markdown"` is supported. See
	// [Markdown](https://en.wikipedia.org/wiki/Markdown) for more information.
	MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"`
	// contains filtered or unexported fields
}

A content string and a MIME type that describes the content string's format.

func (*AlertPolicy_Documentation) Descriptor

func (*AlertPolicy_Documentation) Descriptor() ([]byte, []int)

Deprecated: Use AlertPolicy_Documentation.ProtoReflect.Descriptor instead.

func (*AlertPolicy_Documentation) GetContent

func (x *AlertPolicy_Documentation) GetContent() string

func (*AlertPolicy_Documentation) GetMimeType

func (x *AlertPolicy_Documentation) GetMimeType() string

func (*AlertPolicy_Documentation) ProtoMessage

func (*AlertPolicy_Documentation) ProtoMessage()

func (*AlertPolicy_Documentation) ProtoReflect

func (*AlertPolicy_Documentation) Reset

func (x *AlertPolicy_Documentation) Reset()

func (*AlertPolicy_Documentation) String

func (x *AlertPolicy_Documentation) String() string

type BasicSli

type BasicSli struct {

	// OPTIONAL: The set of RPCs to which this SLI is relevant. Telemetry from
	// other methods will not be used to calculate performance for this SLI. If
	// omitted, this SLI applies to all the Service's methods. For service types
	// that don't support breaking down by method, setting this field will result
	// in an error.
	Method []string `protobuf:"bytes,7,rep,name=method,proto3" json:"method,omitempty"`
	// OPTIONAL: The set of locations to which this SLI is relevant. Telemetry
	// from other locations will not be used to calculate performance for this
	// SLI. If omitted, this SLI applies to all locations in which the Service has
	// activity. For service types that don't support breaking down by location,
	// setting this field will result in an error.
	Location []string `protobuf:"bytes,8,rep,name=location,proto3" json:"location,omitempty"`
	// OPTIONAL: The set of API versions to which this SLI is relevant. Telemetry
	// from other API versions will not be used to calculate performance for this
	// SLI. If omitted, this SLI applies to all API versions. For service types
	// that don't support breaking down by version, setting this field will result
	// in an error.
	Version []string `protobuf:"bytes,9,rep,name=version,proto3" json:"version,omitempty"`
	// This SLI can be evaluated on the basis of availability or latency.
	//
	// Types that are assignable to SliCriteria:
	//	*BasicSli_Availability
	//	*BasicSli_Latency
	SliCriteria isBasicSli_SliCriteria `protobuf_oneof:"sli_criteria"`
	// contains filtered or unexported fields
}

An SLI measuring performance on a well-known service type. Performance will be computed on the basis of pre-defined metrics. The type of the `service_resource` determines the metrics to use and the `service_resource.labels` and `metric_labels` are used to construct a monitoring filter to filter that metric down to just the data relevant to this service.

func (*BasicSli) Descriptor

func (*BasicSli) Descriptor() ([]byte, []int)

Deprecated: Use BasicSli.ProtoReflect.Descriptor instead.

func (*BasicSli) GetAvailability

func (x *BasicSli) GetAvailability() *BasicSli_AvailabilityCriteria

func (*BasicSli) GetLatency

func (x *BasicSli) GetLatency() *BasicSli_LatencyCriteria

func (*BasicSli) GetLocation

func (x *BasicSli) GetLocation() []string

func (*BasicSli) GetMethod

func (x *BasicSli) GetMethod() []string

func (*BasicSli) GetSliCriteria

func (m *BasicSli) GetSliCriteria() isBasicSli_SliCriteria

func (*BasicSli) GetVersion

func (x *BasicSli) GetVersion() []string

func (*BasicSli) ProtoMessage

func (*BasicSli) ProtoMessage()

func (*BasicSli) ProtoReflect

func (x *BasicSli) ProtoReflect() protoreflect.Message

func (*BasicSli) Reset

func (x *BasicSli) Reset()

func (*BasicSli) String

func (x *BasicSli) String() string

type BasicSli_Availability

type BasicSli_Availability struct {
	// Good service is defined to be the count of requests made to this service
	// that return successfully.
	Availability *BasicSli_AvailabilityCriteria `protobuf:"bytes,2,opt,name=availability,proto3,oneof"`
}

type BasicSli_AvailabilityCriteria

type BasicSli_AvailabilityCriteria struct {
	// contains filtered or unexported fields
}

Future parameters for the availability SLI.

func (*BasicSli_AvailabilityCriteria) Descriptor

func (*BasicSli_AvailabilityCriteria) Descriptor() ([]byte, []int)

Deprecated: Use BasicSli_AvailabilityCriteria.ProtoReflect.Descriptor instead.

func (*BasicSli_AvailabilityCriteria) ProtoMessage

func (*BasicSli_AvailabilityCriteria) ProtoMessage()

func (*BasicSli_AvailabilityCriteria) ProtoReflect

func (*BasicSli_AvailabilityCriteria) Reset

func (x *BasicSli_AvailabilityCriteria) Reset()

func (*BasicSli_AvailabilityCriteria) String

type BasicSli_Latency

type BasicSli_Latency struct {
	// Good service is defined to be the count of requests made to this service
	// that are fast enough with respect to `latency.threshold`.
	Latency *BasicSli_LatencyCriteria `protobuf:"bytes,3,opt,name=latency,proto3,oneof"`
}

type BasicSli_LatencyCriteria

type BasicSli_LatencyCriteria struct {

	// Good service is defined to be the count of requests made to this service
	// that return in no more than `threshold`.
	Threshold *durationpb.Duration `protobuf:"bytes,3,opt,name=threshold,proto3" json:"threshold,omitempty"`
	// contains filtered or unexported fields
}

Parameters for a latency threshold SLI.

func (*BasicSli_LatencyCriteria) Descriptor

func (*BasicSli_LatencyCriteria) Descriptor() ([]byte, []int)

Deprecated: Use BasicSli_LatencyCriteria.ProtoReflect.Descriptor instead.

func (*BasicSli_LatencyCriteria) GetThreshold

func (x *BasicSli_LatencyCriteria) GetThreshold() *durationpb.Duration

func (*BasicSli_LatencyCriteria) ProtoMessage

func (*BasicSli_LatencyCriteria) ProtoMessage()

func (*BasicSli_LatencyCriteria) ProtoReflect

func (x *BasicSli_LatencyCriteria) ProtoReflect() protoreflect.Message

func (*BasicSli_LatencyCriteria) Reset

func (x *BasicSli_LatencyCriteria) Reset()

func (*BasicSli_LatencyCriteria) String

func (x *BasicSli_LatencyCriteria) String() string

type ComparisonType

type ComparisonType int32

Specifies an ordering relationship on two arguments, called `left` and `right`.

const (
	// No ordering relationship is specified.
	ComparisonType_COMPARISON_UNSPECIFIED ComparisonType = 0
	// True if the left argument is greater than the right argument.
	ComparisonType_COMPARISON_GT ComparisonType = 1
	// True if the left argument is greater than or equal to the right argument.
	ComparisonType_COMPARISON_GE ComparisonType = 2
	// True if the left argument is less than the right argument.
	ComparisonType_COMPARISON_LT ComparisonType = 3
	// True if the left argument is less than or equal to the right argument.
	ComparisonType_COMPARISON_LE ComparisonType = 4
	// True if the left argument is equal to the right argument.
	ComparisonType_COMPARISON_EQ ComparisonType = 5
	// True if the left argument is not equal to the right argument.
	ComparisonType_COMPARISON_NE ComparisonType = 6
)

func (ComparisonType) Descriptor

func (ComparisonType) Enum

func (x ComparisonType) Enum() *ComparisonType

func (ComparisonType) EnumDescriptor

func (ComparisonType) EnumDescriptor() ([]byte, []int)

Deprecated: Use ComparisonType.Descriptor instead.

func (ComparisonType) Number

func (ComparisonType) String

func (x ComparisonType) String() string

func (ComparisonType) Type

type CreateAlertPolicyRequest

type CreateAlertPolicyRequest struct {

	// Required. The project in which to create the alerting policy. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	//
	// Note that this field names the parent container in which the alerting
	// policy will be written, not the name of the created policy. The alerting
	// policy that is returned will have a name that contains a normalized
	// representation of this name as a prefix but adds a suffix of the form
	// `/alertPolicies/[ALERT_POLICY_ID]`, identifying the policy in the
	// container.
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// Required. The requested alerting policy. You should omit the `name` field in this
	// policy. The name will be returned in the new policy, including
	// a new `[ALERT_POLICY_ID]` value.
	AlertPolicy *AlertPolicy `protobuf:"bytes,2,opt,name=alert_policy,json=alertPolicy,proto3" json:"alert_policy,omitempty"`
	// contains filtered or unexported fields
}

The protocol for the `CreateAlertPolicy` request.

func (*CreateAlertPolicyRequest) Descriptor

func (*CreateAlertPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateAlertPolicyRequest.ProtoReflect.Descriptor instead.

func (*CreateAlertPolicyRequest) GetAlertPolicy

func (x *CreateAlertPolicyRequest) GetAlertPolicy() *AlertPolicy

func (*CreateAlertPolicyRequest) GetName

func (x *CreateAlertPolicyRequest) GetName() string

func (*CreateAlertPolicyRequest) ProtoMessage

func (*CreateAlertPolicyRequest) ProtoMessage()

func (*CreateAlertPolicyRequest) ProtoReflect

func (x *CreateAlertPolicyRequest) ProtoReflect() protoreflect.Message

func (*CreateAlertPolicyRequest) Reset

func (x *CreateAlertPolicyRequest) Reset()

func (*CreateAlertPolicyRequest) String

func (x *CreateAlertPolicyRequest) String() string

type CreateGroupRequest

type CreateGroupRequest struct {

	// Required. The project in which to create the group. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"`
	// Required. A group definition. It is an error to define the `name` field because
	// the system assigns the name.
	Group *Group `protobuf:"bytes,2,opt,name=group,proto3" json:"group,omitempty"`
	// If true, validate this request but do not create the group.
	ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly,proto3" json:"validate_only,omitempty"`
	// contains filtered or unexported fields
}

The `CreateGroup` request.

func (*CreateGroupRequest) Descriptor

func (*CreateGroupRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateGroupRequest.ProtoReflect.Descriptor instead.

func (*CreateGroupRequest) GetGroup

func (x *CreateGroupRequest) GetGroup() *Group

func (*CreateGroupRequest) GetName

func (x *CreateGroupRequest) GetName() string

func (*CreateGroupRequest) GetValidateOnly

func (x *CreateGroupRequest) GetValidateOnly() bool

func (*CreateGroupRequest) ProtoMessage

func (*CreateGroupRequest) ProtoMessage()

func (*CreateGroupRequest) ProtoReflect

func (x *CreateGroupRequest) ProtoReflect() protoreflect.Message

func (*CreateGroupRequest) Reset

func (x *CreateGroupRequest) Reset()

func (*CreateGroupRequest) String

func (x *CreateGroupRequest) String() string

type CreateMetricDescriptorRequest

type CreateMetricDescriptorRequest struct {

	// Required. The project on which to execute the request. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// Required. The new [custom metric](https://cloud.google.com/monitoring/custom-metrics)
	// descriptor.
	MetricDescriptor *metric.MetricDescriptor `protobuf:"bytes,2,opt,name=metric_descriptor,json=metricDescriptor,proto3" json:"metric_descriptor,omitempty"`
	// contains filtered or unexported fields
}

The `CreateMetricDescriptor` request.

func (*CreateMetricDescriptorRequest) Descriptor

func (*CreateMetricDescriptorRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateMetricDescriptorRequest.ProtoReflect.Descriptor instead.

func (*CreateMetricDescriptorRequest) GetMetricDescriptor

func (x *CreateMetricDescriptorRequest) GetMetricDescriptor() *metric.MetricDescriptor

func (*CreateMetricDescriptorRequest) GetName

func (*CreateMetricDescriptorRequest) ProtoMessage

func (*CreateMetricDescriptorRequest) ProtoMessage()

func (*CreateMetricDescriptorRequest) ProtoReflect

func (*CreateMetricDescriptorRequest) Reset

func (x *CreateMetricDescriptorRequest) Reset()

func (*CreateMetricDescriptorRequest) String

type CreateNotificationChannelRequest

type CreateNotificationChannelRequest struct {

	// Required. The project on which to execute the request. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	//
	// This names the container into which the channel will be
	// written, this does not name the newly created channel. The resulting
	// channel's name will have a normalized version of this field as a prefix,
	// but will add `/notificationChannels/[CHANNEL_ID]` to identify the channel.
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// Required. The definition of the `NotificationChannel` to create.
	NotificationChannel *NotificationChannel `protobuf:"bytes,2,opt,name=notification_channel,json=notificationChannel,proto3" json:"notification_channel,omitempty"`
	// contains filtered or unexported fields
}

The `CreateNotificationChannel` request.

func (*CreateNotificationChannelRequest) Descriptor

func (*CreateNotificationChannelRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateNotificationChannelRequest.ProtoReflect.Descriptor instead.

func (*CreateNotificationChannelRequest) GetName

func (*CreateNotificationChannelRequest) GetNotificationChannel

func (x *CreateNotificationChannelRequest) GetNotificationChannel() *NotificationChannel

func (*CreateNotificationChannelRequest) ProtoMessage

func (*CreateNotificationChannelRequest) ProtoMessage()

func (*CreateNotificationChannelRequest) ProtoReflect

func (*CreateNotificationChannelRequest) Reset

func (*CreateNotificationChannelRequest) String

type CreateServiceLevelObjectiveRequest

type CreateServiceLevelObjectiveRequest struct {

	// Required. Resource name of the parent `Service`. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/services/[SERVICE_ID]
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Optional. The ServiceLevelObjective id to use for this
	// ServiceLevelObjective. If omitted, an id will be generated instead. Must
	// match the pattern `[a-z0-9\-]+`
	ServiceLevelObjectiveId string `` /* 134-byte string literal not displayed */
	// Required. The `ServiceLevelObjective` to create.
	// The provided `name` will be respected if no `ServiceLevelObjective` exists
	// with this name.
	ServiceLevelObjective *ServiceLevelObjective `` /* 126-byte string literal not displayed */
	// contains filtered or unexported fields
}

The `CreateServiceLevelObjective` request.

func (*CreateServiceLevelObjectiveRequest) Descriptor

func (*CreateServiceLevelObjectiveRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateServiceLevelObjectiveRequest.ProtoReflect.Descriptor instead.

func (*CreateServiceLevelObjectiveRequest) GetParent

func (*CreateServiceLevelObjectiveRequest) GetServiceLevelObjective

func (x *CreateServiceLevelObjectiveRequest) GetServiceLevelObjective() *ServiceLevelObjective

func (*CreateServiceLevelObjectiveRequest) GetServiceLevelObjectiveId

func (x *CreateServiceLevelObjectiveRequest) GetServiceLevelObjectiveId() string

func (*CreateServiceLevelObjectiveRequest) ProtoMessage

func (*CreateServiceLevelObjectiveRequest) ProtoMessage()

func (*CreateServiceLevelObjectiveRequest) ProtoReflect

func (*CreateServiceLevelObjectiveRequest) Reset

func (*CreateServiceLevelObjectiveRequest) String

type CreateServiceRequest

type CreateServiceRequest struct {

	// Required. Resource name of the parent workspace. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Optional. The Service id to use for this Service. If omitted, an id will be
	// generated instead. Must match the pattern `[a-z0-9\-]+`
	ServiceId string `protobuf:"bytes,3,opt,name=service_id,json=serviceId,proto3" json:"service_id,omitempty"`
	// Required. The `Service` to create.
	Service *Service `protobuf:"bytes,2,opt,name=service,proto3" json:"service,omitempty"`
	// contains filtered or unexported fields
}

The `CreateService` request.

func (*CreateServiceRequest) Descriptor

func (*CreateServiceRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateServiceRequest.ProtoReflect.Descriptor instead.

func (*CreateServiceRequest) GetParent

func (x *CreateServiceRequest) GetParent() string

func (*CreateServiceRequest) GetService

func (x *CreateServiceRequest) GetService() *Service

func (*CreateServiceRequest) GetServiceId

func (x *CreateServiceRequest) GetServiceId() string

func (*CreateServiceRequest) ProtoMessage

func (*CreateServiceRequest) ProtoMessage()

func (*CreateServiceRequest) ProtoReflect

func (x *CreateServiceRequest) ProtoReflect() protoreflect.Message

func (*CreateServiceRequest) Reset

func (x *CreateServiceRequest) Reset()

func (*CreateServiceRequest) String

func (x *CreateServiceRequest) String() string

type CreateTimeSeriesError

type CreateTimeSeriesError struct {

	// DEPRECATED. Time series ID that resulted in the `status` error.
	//
	// Deprecated: Do not use.
	TimeSeries *TimeSeries `protobuf:"bytes,1,opt,name=time_series,json=timeSeries,proto3" json:"time_series,omitempty"`
	// DEPRECATED. The status of the requested write operation for `time_series`.
	//
	// Deprecated: Do not use.
	Status *status.Status `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"`
	// contains filtered or unexported fields
}

DEPRECATED. Used to hold per-time-series error status.

func (*CreateTimeSeriesError) Descriptor

func (*CreateTimeSeriesError) Descriptor() ([]byte, []int)

Deprecated: Use CreateTimeSeriesError.ProtoReflect.Descriptor instead.

func (*CreateTimeSeriesError) GetStatus

func (x *CreateTimeSeriesError) GetStatus() *status.Status

Deprecated: Do not use.

func (*CreateTimeSeriesError) GetTimeSeries

func (x *CreateTimeSeriesError) GetTimeSeries() *TimeSeries

Deprecated: Do not use.

func (*CreateTimeSeriesError) ProtoMessage

func (*CreateTimeSeriesError) ProtoMessage()

func (*CreateTimeSeriesError) ProtoReflect

func (x *CreateTimeSeriesError) ProtoReflect() protoreflect.Message

func (*CreateTimeSeriesError) Reset

func (x *CreateTimeSeriesError) Reset()

func (*CreateTimeSeriesError) String

func (x *CreateTimeSeriesError) String() string

type CreateTimeSeriesRequest

type CreateTimeSeriesRequest struct {

	// Required. The project on which to execute the request. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// Required. The new data to be added to a list of time series.
	// Adds at most one data point to each of several time series.  The new data
	// point must be more recent than any other point in its time series.  Each
	// `TimeSeries` value must fully specify a unique time series by supplying
	// all label values for the metric and the monitored resource.
	//
	// The maximum number of `TimeSeries` objects per `Create` request is 200.
	TimeSeries []*TimeSeries `protobuf:"bytes,2,rep,name=time_series,json=timeSeries,proto3" json:"time_series,omitempty"`
	// contains filtered or unexported fields
}

The `CreateTimeSeries` request.

func (*CreateTimeSeriesRequest) Descriptor

func (*CreateTimeSeriesRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateTimeSeriesRequest.ProtoReflect.Descriptor instead.

func (*CreateTimeSeriesRequest) GetName

func (x *CreateTimeSeriesRequest) GetName() string

func (*CreateTimeSeriesRequest) GetTimeSeries

func (x *CreateTimeSeriesRequest) GetTimeSeries() []*TimeSeries

func (*CreateTimeSeriesRequest) ProtoMessage

func (*CreateTimeSeriesRequest) ProtoMessage()

func (*CreateTimeSeriesRequest) ProtoReflect

func (x *CreateTimeSeriesRequest) ProtoReflect() protoreflect.Message

func (*CreateTimeSeriesRequest) Reset

func (x *CreateTimeSeriesRequest) Reset()

func (*CreateTimeSeriesRequest) String

func (x *CreateTimeSeriesRequest) String() string

type CreateTimeSeriesSummary

type CreateTimeSeriesSummary struct {

	// The number of points in the request.
	TotalPointCount int32 `protobuf:"varint,1,opt,name=total_point_count,json=totalPointCount,proto3" json:"total_point_count,omitempty"`
	// The number of points that were successfully written.
	SuccessPointCount int32 `protobuf:"varint,2,opt,name=success_point_count,json=successPointCount,proto3" json:"success_point_count,omitempty"`
	// The number of points that failed to be written. Order is not guaranteed.
	Errors []*CreateTimeSeriesSummary_Error `protobuf:"bytes,3,rep,name=errors,proto3" json:"errors,omitempty"`
	// contains filtered or unexported fields
}

Summary of the result of a failed request to write data to a time series.

func (*CreateTimeSeriesSummary) Descriptor

func (*CreateTimeSeriesSummary) Descriptor() ([]byte, []int)

Deprecated: Use CreateTimeSeriesSummary.ProtoReflect.Descriptor instead.

func (*CreateTimeSeriesSummary) GetErrors

func (*CreateTimeSeriesSummary) GetSuccessPointCount

func (x *CreateTimeSeriesSummary) GetSuccessPointCount() int32

func (*CreateTimeSeriesSummary) GetTotalPointCount

func (x *CreateTimeSeriesSummary) GetTotalPointCount() int32

func (*CreateTimeSeriesSummary) ProtoMessage

func (*CreateTimeSeriesSummary) ProtoMessage()

func (*CreateTimeSeriesSummary) ProtoReflect

func (x *CreateTimeSeriesSummary) ProtoReflect() protoreflect.Message

func (*CreateTimeSeriesSummary) Reset

func (x *CreateTimeSeriesSummary) Reset()

func (*CreateTimeSeriesSummary) String

func (x *CreateTimeSeriesSummary) String() string

type CreateTimeSeriesSummary_Error

type CreateTimeSeriesSummary_Error struct {

	// The status of the requested write operation.
	Status *status.Status `protobuf:"bytes,1,opt,name=status,proto3" json:"status,omitempty"`
	// The number of points that couldn't be written because of `status`.
	PointCount int32 `protobuf:"varint,2,opt,name=point_count,json=pointCount,proto3" json:"point_count,omitempty"`
	// contains filtered or unexported fields
}

Detailed information about an error category.

func (*CreateTimeSeriesSummary_Error) Descriptor

func (*CreateTimeSeriesSummary_Error) Descriptor() ([]byte, []int)

Deprecated: Use CreateTimeSeriesSummary_Error.ProtoReflect.Descriptor instead.

func (*CreateTimeSeriesSummary_Error) GetPointCount

func (x *CreateTimeSeriesSummary_Error) GetPointCount() int32

func (*CreateTimeSeriesSummary_Error) GetStatus

func (*CreateTimeSeriesSummary_Error) ProtoMessage

func (*CreateTimeSeriesSummary_Error) ProtoMessage()

func (*CreateTimeSeriesSummary_Error) ProtoReflect

func (*CreateTimeSeriesSummary_Error) Reset

func (x *CreateTimeSeriesSummary_Error) Reset()

func (*CreateTimeSeriesSummary_Error) String

type CreateUptimeCheckConfigRequest

type CreateUptimeCheckConfigRequest struct {

	// Required. The project in which to create the Uptime check. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]
	Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"`
	// Required. The new Uptime check configuration.
	UptimeCheckConfig *UptimeCheckConfig `protobuf:"bytes,2,opt,name=uptime_check_config,json=uptimeCheckConfig,proto3" json:"uptime_check_config,omitempty"`
	// contains filtered or unexported fields
}

The protocol for the `CreateUptimeCheckConfig` request.

func (*CreateUptimeCheckConfigRequest) Descriptor

func (*CreateUptimeCheckConfigRequest) Descriptor() ([]byte, []int)

Deprecated: Use CreateUptimeCheckConfigRequest.ProtoReflect.Descriptor instead.

func (*CreateUptimeCheckConfigRequest) GetParent

func (x *CreateUptimeCheckConfigRequest) GetParent() string

func (*CreateUptimeCheckConfigRequest) GetUptimeCheckConfig

func (x *CreateUptimeCheckConfigRequest) GetUptimeCheckConfig() *UptimeCheckConfig

func (*CreateUptimeCheckConfigRequest) ProtoMessage

func (*CreateUptimeCheckConfigRequest) ProtoMessage()

func (*CreateUptimeCheckConfigRequest) ProtoReflect

func (*CreateUptimeCheckConfigRequest) Reset

func (x *CreateUptimeCheckConfigRequest) Reset()

func (*CreateUptimeCheckConfigRequest) String

type DeleteAlertPolicyRequest

type DeleteAlertPolicyRequest struct {

	// Required. The alerting policy to delete. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/alertPolicies/[ALERT_POLICY_ID]
	//
	// For more information, see [AlertPolicy][google.monitoring.v3.AlertPolicy].
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The protocol for the `DeleteAlertPolicy` request.

func (*DeleteAlertPolicyRequest) Descriptor

func (*DeleteAlertPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteAlertPolicyRequest.ProtoReflect.Descriptor instead.

func (*DeleteAlertPolicyRequest) GetName

func (x *DeleteAlertPolicyRequest) GetName() string

func (*DeleteAlertPolicyRequest) ProtoMessage

func (*DeleteAlertPolicyRequest) ProtoMessage()

func (*DeleteAlertPolicyRequest) ProtoReflect

func (x *DeleteAlertPolicyRequest) ProtoReflect() protoreflect.Message

func (*DeleteAlertPolicyRequest) Reset

func (x *DeleteAlertPolicyRequest) Reset()

func (*DeleteAlertPolicyRequest) String

func (x *DeleteAlertPolicyRequest) String() string

type DeleteGroupRequest

type DeleteGroupRequest struct {

	// Required. The group to delete. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/groups/[GROUP_ID]
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// If this field is true, then the request means to delete a group with all
	// its descendants. Otherwise, the request means to delete a group only when
	// it has no descendants. The default value is false.
	Recursive bool `protobuf:"varint,4,opt,name=recursive,proto3" json:"recursive,omitempty"`
	// contains filtered or unexported fields
}

The `DeleteGroup` request. The default behavior is to be able to delete a single group without any descendants.

func (*DeleteGroupRequest) Descriptor

func (*DeleteGroupRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteGroupRequest.ProtoReflect.Descriptor instead.

func (*DeleteGroupRequest) GetName

func (x *DeleteGroupRequest) GetName() string

func (*DeleteGroupRequest) GetRecursive

func (x *DeleteGroupRequest) GetRecursive() bool

func (*DeleteGroupRequest) ProtoMessage

func (*DeleteGroupRequest) ProtoMessage()

func (*DeleteGroupRequest) ProtoReflect

func (x *DeleteGroupRequest) ProtoReflect() protoreflect.Message

func (*DeleteGroupRequest) Reset

func (x *DeleteGroupRequest) Reset()

func (*DeleteGroupRequest) String

func (x *DeleteGroupRequest) String() string

type DeleteMetricDescriptorRequest

type DeleteMetricDescriptorRequest struct {

	// Required. The metric descriptor on which to execute the request. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/metricDescriptors/[METRIC_ID]
	//
	// An example of `[METRIC_ID]` is:
	// `"custom.googleapis.com/my_test_metric"`.
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The `DeleteMetricDescriptor` request.

func (*DeleteMetricDescriptorRequest) Descriptor

func (*DeleteMetricDescriptorRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteMetricDescriptorRequest.ProtoReflect.Descriptor instead.

func (*DeleteMetricDescriptorRequest) GetName

func (*DeleteMetricDescriptorRequest) ProtoMessage

func (*DeleteMetricDescriptorRequest) ProtoMessage()

func (*DeleteMetricDescriptorRequest) ProtoReflect

func (*DeleteMetricDescriptorRequest) Reset

func (x *DeleteMetricDescriptorRequest) Reset()

func (*DeleteMetricDescriptorRequest) String

type DeleteNotificationChannelRequest

type DeleteNotificationChannelRequest struct {

	// Required. The channel for which to execute the request. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/notificationChannels/[CHANNEL_ID]
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// If true, the notification channel will be deleted regardless of its
	// use in alert policies (the policies will be updated to remove the
	// channel). If false, channels that are still referenced by an existing
	// alerting policy will fail to be deleted in a delete operation.
	Force bool `protobuf:"varint,5,opt,name=force,proto3" json:"force,omitempty"`
	// contains filtered or unexported fields
}

The `DeleteNotificationChannel` request.

func (*DeleteNotificationChannelRequest) Descriptor

func (*DeleteNotificationChannelRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteNotificationChannelRequest.ProtoReflect.Descriptor instead.

func (*DeleteNotificationChannelRequest) GetForce

func (x *DeleteNotificationChannelRequest) GetForce() bool

func (*DeleteNotificationChannelRequest) GetName

func (*DeleteNotificationChannelRequest) ProtoMessage

func (*DeleteNotificationChannelRequest) ProtoMessage()

func (*DeleteNotificationChannelRequest) ProtoReflect

func (*DeleteNotificationChannelRequest) Reset

func (*DeleteNotificationChannelRequest) String

type DeleteServiceLevelObjectiveRequest

type DeleteServiceLevelObjectiveRequest struct {

	// Required. Resource name of the `ServiceLevelObjective` to delete. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/services/[SERVICE_ID]/serviceLevelObjectives/[SLO_NAME]
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The `DeleteServiceLevelObjective` request.

func (*DeleteServiceLevelObjectiveRequest) Descriptor

func (*DeleteServiceLevelObjectiveRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteServiceLevelObjectiveRequest.ProtoReflect.Descriptor instead.

func (*DeleteServiceLevelObjectiveRequest) GetName

func (*DeleteServiceLevelObjectiveRequest) ProtoMessage

func (*DeleteServiceLevelObjectiveRequest) ProtoMessage()

func (*DeleteServiceLevelObjectiveRequest) ProtoReflect

func (*DeleteServiceLevelObjectiveRequest) Reset

func (*DeleteServiceLevelObjectiveRequest) String

type DeleteServiceRequest

type DeleteServiceRequest struct {

	// Required. Resource name of the `Service` to delete. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/services/[SERVICE_ID]
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The `DeleteService` request.

func (*DeleteServiceRequest) Descriptor

func (*DeleteServiceRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteServiceRequest.ProtoReflect.Descriptor instead.

func (*DeleteServiceRequest) GetName

func (x *DeleteServiceRequest) GetName() string

func (*DeleteServiceRequest) ProtoMessage

func (*DeleteServiceRequest) ProtoMessage()

func (*DeleteServiceRequest) ProtoReflect

func (x *DeleteServiceRequest) ProtoReflect() protoreflect.Message

func (*DeleteServiceRequest) Reset

func (x *DeleteServiceRequest) Reset()

func (*DeleteServiceRequest) String

func (x *DeleteServiceRequest) String() string

type DeleteUptimeCheckConfigRequest

type DeleteUptimeCheckConfigRequest struct {

	// Required. The Uptime check configuration to delete. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/uptimeCheckConfigs/[UPTIME_CHECK_ID]
	Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The protocol for the `DeleteUptimeCheckConfig` request.

func (*DeleteUptimeCheckConfigRequest) Descriptor

func (*DeleteUptimeCheckConfigRequest) Descriptor() ([]byte, []int)

Deprecated: Use DeleteUptimeCheckConfigRequest.ProtoReflect.Descriptor instead.

func (*DeleteUptimeCheckConfigRequest) GetName

func (*DeleteUptimeCheckConfigRequest) ProtoMessage

func (*DeleteUptimeCheckConfigRequest) ProtoMessage()

func (*DeleteUptimeCheckConfigRequest) ProtoReflect

func (*DeleteUptimeCheckConfigRequest) Reset

func (x *DeleteUptimeCheckConfigRequest) Reset()

func (*DeleteUptimeCheckConfigRequest) String

type DistributionCut

type DistributionCut struct {

	// A [monitoring filter](https://cloud.google.com/monitoring/api/v3/filters)
	// specifying a `TimeSeries` aggregating values. Must have `ValueType =
	// DISTRIBUTION` and `MetricKind = DELTA` or `MetricKind = CUMULATIVE`.
	DistributionFilter string `protobuf:"bytes,4,opt,name=distribution_filter,json=distributionFilter,proto3" json:"distribution_filter,omitempty"`
	// Range of values considered "good." For a one-sided range, set one bound to
	// an infinite value.
	Range *Range `protobuf:"bytes,5,opt,name=range,proto3" json:"range,omitempty"`
	// contains filtered or unexported fields
}

A `DistributionCut` defines a `TimeSeries` and thresholds used for measuring good service and total service. The `TimeSeries` must have `ValueType = DISTRIBUTION` and `MetricKind = DELTA` or `MetricKind = CUMULATIVE`. The computed `good_service` will be the count of values x in the `Distribution` such that `range.min <= x < range.max`.

func (*DistributionCut) Descriptor

func (*DistributionCut) Descriptor() ([]byte, []int)

Deprecated: Use DistributionCut.ProtoReflect.Descriptor instead.

func (*DistributionCut) GetDistributionFilter

func (x *DistributionCut) GetDistributionFilter() string

func (*DistributionCut) GetRange

func (x *DistributionCut) GetRange() *Range

func (*DistributionCut) ProtoMessage

func (*DistributionCut) ProtoMessage()

func (*DistributionCut) ProtoReflect

func (x *DistributionCut) ProtoReflect() protoreflect.Message

func (*DistributionCut) Reset

func (x *DistributionCut) Reset()

func (*DistributionCut) String

func (x *DistributionCut) String() string

type DroppedLabels

type DroppedLabels struct {

	// Map from label to its value, for all labels dropped in any aggregation.
	Label map[string]string `` /* 151-byte string literal not displayed */
	// contains filtered or unexported fields
}

A set of (label, value) pairs which were dropped during aggregation, attached to google.api.Distribution.Exemplars in google.api.Distribution values during aggregation.

These values are used in combination with the label values that remain on the aggregated Distribution timeseries to construct the full label set for the exemplar values. The resulting full label set may be used to identify the specific task/job/instance (for example) which may be contributing to a long-tail, while allowing the storage savings of only storing aggregated distribution values for a large group.

Note that there are no guarantees on ordering of the labels from exemplar-to-exemplar and from distribution-to-distribution in the same stream, and there may be duplicates. It is up to clients to resolve any ambiguities.

func (*DroppedLabels) Descriptor

func (*DroppedLabels) Descriptor() ([]byte, []int)

Deprecated: Use DroppedLabels.ProtoReflect.Descriptor instead.

func (*DroppedLabels) GetLabel

func (x *DroppedLabels) GetLabel() map[string]string

func (*DroppedLabels) ProtoMessage

func (*DroppedLabels) ProtoMessage()

func (*DroppedLabels) ProtoReflect

func (x *DroppedLabels) ProtoReflect() protoreflect.Message

func (*DroppedLabels) Reset

func (x *DroppedLabels) Reset()

func (*DroppedLabels) String

func (x *DroppedLabels) String() string

type GetAlertPolicyRequest

type GetAlertPolicyRequest struct {

	// Required. The alerting policy to retrieve. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/alertPolicies/[ALERT_POLICY_ID]
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The protocol for the `GetAlertPolicy` request.

func (*GetAlertPolicyRequest) Descriptor

func (*GetAlertPolicyRequest) Descriptor() ([]byte, []int)

Deprecated: Use GetAlertPolicyRequest.ProtoReflect.Descriptor instead.

func (*GetAlertPolicyRequest) GetName

func (x *GetAlertPolicyRequest) GetName() string

func (*GetAlertPolicyRequest) ProtoMessage

func (*GetAlertPolicyRequest) ProtoMessage()

func (*GetAlertPolicyRequest) ProtoReflect

func (x *GetAlertPolicyRequest) ProtoReflect() protoreflect.Message

func (*GetAlertPolicyRequest) Reset

func (x *GetAlertPolicyRequest) Reset()

func (*GetAlertPolicyRequest) String

func (x *GetAlertPolicyRequest) String() string

type GetGroupRequest

type GetGroupRequest struct {

	// Required. The group to retrieve. The format is:
	//
	//     projects/[PROJECT_ID_OR_NUMBER]/groups/[GROUP_ID]
	Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"`
	// contains filtered or unexported fields
}

The `GetGroup` request.