Documentation
¶
Index ¶
- func BigqueryJob_GenerateConfigForImport(scope constructs.Construct, importToId *string, importFromId *string, ...) cdktf.ImportableResource
- func BigqueryJob_IsConstruct(x interface{}) *bool
- func BigqueryJob_IsTerraformElement(x interface{}) *bool
- func BigqueryJob_IsTerraformResource(x interface{}) *bool
- func BigqueryJob_TfResourceType() *string
- func NewBigqueryJobCopyDestinationEncryptionConfigurationOutputReference_Override(b BigqueryJobCopyDestinationEncryptionConfigurationOutputReference, ...)
- func NewBigqueryJobCopyDestinationTableOutputReference_Override(b BigqueryJobCopyDestinationTableOutputReference, ...)
- func NewBigqueryJobCopyOutputReference_Override(b BigqueryJobCopyOutputReference, terraformResource cdktf.IInterpolatingParent, ...)
- func NewBigqueryJobCopySourceTablesList_Override(b BigqueryJobCopySourceTablesList, ...)
- func NewBigqueryJobCopySourceTablesOutputReference_Override(b BigqueryJobCopySourceTablesOutputReference, ...)
- func NewBigqueryJobExtractOutputReference_Override(b BigqueryJobExtractOutputReference, ...)
- func NewBigqueryJobExtractSourceModelOutputReference_Override(b BigqueryJobExtractSourceModelOutputReference, ...)
- func NewBigqueryJobExtractSourceTableOutputReference_Override(b BigqueryJobExtractSourceTableOutputReference, ...)
- func NewBigqueryJobLoadDestinationEncryptionConfigurationOutputReference_Override(b BigqueryJobLoadDestinationEncryptionConfigurationOutputReference, ...)
- func NewBigqueryJobLoadDestinationTableOutputReference_Override(b BigqueryJobLoadDestinationTableOutputReference, ...)
- func NewBigqueryJobLoadOutputReference_Override(b BigqueryJobLoadOutputReference, terraformResource cdktf.IInterpolatingParent, ...)
- func NewBigqueryJobLoadParquetOptionsOutputReference_Override(b BigqueryJobLoadParquetOptionsOutputReference, ...)
- func NewBigqueryJobLoadTimePartitioningOutputReference_Override(b BigqueryJobLoadTimePartitioningOutputReference, ...)
- func NewBigqueryJobQueryDefaultDatasetOutputReference_Override(b BigqueryJobQueryDefaultDatasetOutputReference, ...)
- func NewBigqueryJobQueryDestinationEncryptionConfigurationOutputReference_Override(b BigqueryJobQueryDestinationEncryptionConfigurationOutputReference, ...)
- func NewBigqueryJobQueryDestinationTableOutputReference_Override(b BigqueryJobQueryDestinationTableOutputReference, ...)
- func NewBigqueryJobQueryOutputReference_Override(b BigqueryJobQueryOutputReference, ...)
- func NewBigqueryJobQueryScriptOptionsOutputReference_Override(b BigqueryJobQueryScriptOptionsOutputReference, ...)
- func NewBigqueryJobQueryUserDefinedFunctionResourcesList_Override(b BigqueryJobQueryUserDefinedFunctionResourcesList, ...)
- func NewBigqueryJobQueryUserDefinedFunctionResourcesOutputReference_Override(b BigqueryJobQueryUserDefinedFunctionResourcesOutputReference, ...)
- func NewBigqueryJobStatusErrorResultList_Override(b BigqueryJobStatusErrorResultList, ...)
- func NewBigqueryJobStatusErrorResultOutputReference_Override(b BigqueryJobStatusErrorResultOutputReference, ...)
- func NewBigqueryJobStatusErrorsList_Override(b BigqueryJobStatusErrorsList, terraformResource cdktf.IInterpolatingParent, ...)
- func NewBigqueryJobStatusErrorsOutputReference_Override(b BigqueryJobStatusErrorsOutputReference, ...)
- func NewBigqueryJobStatusList_Override(b BigqueryJobStatusList, terraformResource cdktf.IInterpolatingParent, ...)
- func NewBigqueryJobStatusOutputReference_Override(b BigqueryJobStatusOutputReference, ...)
- func NewBigqueryJobTimeoutsOutputReference_Override(b BigqueryJobTimeoutsOutputReference, ...)
- func NewBigqueryJob_Override(b BigqueryJob, scope constructs.Construct, id *string, ...)
- type BigqueryJob
- type BigqueryJobConfig
- type BigqueryJobCopy
- type BigqueryJobCopyDestinationEncryptionConfiguration
- type BigqueryJobCopyDestinationEncryptionConfigurationOutputReference
- type BigqueryJobCopyDestinationTable
- type BigqueryJobCopyDestinationTableOutputReference
- type BigqueryJobCopyOutputReference
- type BigqueryJobCopySourceTables
- type BigqueryJobCopySourceTablesList
- type BigqueryJobCopySourceTablesOutputReference
- type BigqueryJobExtract
- type BigqueryJobExtractOutputReference
- type BigqueryJobExtractSourceModel
- type BigqueryJobExtractSourceModelOutputReference
- type BigqueryJobExtractSourceTable
- type BigqueryJobExtractSourceTableOutputReference
- type BigqueryJobLoad
- type BigqueryJobLoadDestinationEncryptionConfiguration
- type BigqueryJobLoadDestinationEncryptionConfigurationOutputReference
- type BigqueryJobLoadDestinationTable
- type BigqueryJobLoadDestinationTableOutputReference
- type BigqueryJobLoadOutputReference
- type BigqueryJobLoadParquetOptions
- type BigqueryJobLoadParquetOptionsOutputReference
- type BigqueryJobLoadTimePartitioning
- type BigqueryJobLoadTimePartitioningOutputReference
- type BigqueryJobQuery
- type BigqueryJobQueryDefaultDataset
- type BigqueryJobQueryDefaultDatasetOutputReference
- type BigqueryJobQueryDestinationEncryptionConfiguration
- type BigqueryJobQueryDestinationEncryptionConfigurationOutputReference
- type BigqueryJobQueryDestinationTable
- type BigqueryJobQueryDestinationTableOutputReference
- type BigqueryJobQueryOutputReference
- type BigqueryJobQueryScriptOptions
- type BigqueryJobQueryScriptOptionsOutputReference
- type BigqueryJobQueryUserDefinedFunctionResources
- type BigqueryJobQueryUserDefinedFunctionResourcesList
- type BigqueryJobQueryUserDefinedFunctionResourcesOutputReference
- type BigqueryJobStatus
- type BigqueryJobStatusErrorResult
- type BigqueryJobStatusErrorResultList
- type BigqueryJobStatusErrorResultOutputReference
- type BigqueryJobStatusErrors
- type BigqueryJobStatusErrorsList
- type BigqueryJobStatusErrorsOutputReference
- type BigqueryJobStatusList
- type BigqueryJobStatusOutputReference
- type BigqueryJobTimeouts
- type BigqueryJobTimeoutsOutputReference
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func BigqueryJob_GenerateConfigForImport ¶
func BigqueryJob_GenerateConfigForImport(scope constructs.Construct, importToId *string, importFromId *string, provider cdktf.TerraformProvider) cdktf.ImportableResource
Generates CDKTF code for importing a BigqueryJob resource upon running "cdktf plan <stack-name>".
func BigqueryJob_IsConstruct ¶
func BigqueryJob_IsConstruct(x interface{}) *bool
Checks if `x` is a construct.
Use this method instead of `instanceof` to properly detect `Construct` instances, even when the construct library is symlinked.
Explanation: in JavaScript, multiple copies of the `constructs` library on disk are seen as independent, completely different libraries. As a consequence, the class `Construct` in each copy of the `constructs` library is seen as a different class, and an instance of one class will not test as `instanceof` the other class. `npm install` will not create installations like this, but users may manually symlink construct libraries together or use a monorepo tool: in those cases, multiple copies of the `constructs` library can be accidentally installed, and `instanceof` will behave unpredictably. It is safest to avoid using `instanceof`, and using this type-testing method instead.
Returns: true if `x` is an object created from a class which extends `Construct`.
func BigqueryJob_IsTerraformElement ¶
func BigqueryJob_IsTerraformElement(x interface{}) *bool
Experimental.
func BigqueryJob_IsTerraformResource ¶
func BigqueryJob_IsTerraformResource(x interface{}) *bool
Experimental.
func BigqueryJob_TfResourceType ¶
func BigqueryJob_TfResourceType() *string
func NewBigqueryJobCopyDestinationEncryptionConfigurationOutputReference_Override ¶
func NewBigqueryJobCopyDestinationEncryptionConfigurationOutputReference_Override(b BigqueryJobCopyDestinationEncryptionConfigurationOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobCopyDestinationTableOutputReference_Override ¶
func NewBigqueryJobCopyDestinationTableOutputReference_Override(b BigqueryJobCopyDestinationTableOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobCopyOutputReference_Override ¶
func NewBigqueryJobCopyOutputReference_Override(b BigqueryJobCopyOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobCopySourceTablesList_Override ¶
func NewBigqueryJobCopySourceTablesList_Override(b BigqueryJobCopySourceTablesList, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool)
func NewBigqueryJobCopySourceTablesOutputReference_Override ¶
func NewBigqueryJobCopySourceTablesOutputReference_Override(b BigqueryJobCopySourceTablesOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool)
func NewBigqueryJobExtractOutputReference_Override ¶
func NewBigqueryJobExtractOutputReference_Override(b BigqueryJobExtractOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobExtractSourceModelOutputReference_Override ¶
func NewBigqueryJobExtractSourceModelOutputReference_Override(b BigqueryJobExtractSourceModelOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobExtractSourceTableOutputReference_Override ¶
func NewBigqueryJobExtractSourceTableOutputReference_Override(b BigqueryJobExtractSourceTableOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobLoadDestinationEncryptionConfigurationOutputReference_Override ¶
func NewBigqueryJobLoadDestinationEncryptionConfigurationOutputReference_Override(b BigqueryJobLoadDestinationEncryptionConfigurationOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobLoadDestinationTableOutputReference_Override ¶
func NewBigqueryJobLoadDestinationTableOutputReference_Override(b BigqueryJobLoadDestinationTableOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobLoadOutputReference_Override ¶
func NewBigqueryJobLoadOutputReference_Override(b BigqueryJobLoadOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobLoadParquetOptionsOutputReference_Override ¶
func NewBigqueryJobLoadParquetOptionsOutputReference_Override(b BigqueryJobLoadParquetOptionsOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobLoadTimePartitioningOutputReference_Override ¶
func NewBigqueryJobLoadTimePartitioningOutputReference_Override(b BigqueryJobLoadTimePartitioningOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobQueryDefaultDatasetOutputReference_Override ¶
func NewBigqueryJobQueryDefaultDatasetOutputReference_Override(b BigqueryJobQueryDefaultDatasetOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobQueryDestinationEncryptionConfigurationOutputReference_Override ¶
func NewBigqueryJobQueryDestinationEncryptionConfigurationOutputReference_Override(b BigqueryJobQueryDestinationEncryptionConfigurationOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobQueryDestinationTableOutputReference_Override ¶
func NewBigqueryJobQueryDestinationTableOutputReference_Override(b BigqueryJobQueryDestinationTableOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobQueryOutputReference_Override ¶
func NewBigqueryJobQueryOutputReference_Override(b BigqueryJobQueryOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobQueryScriptOptionsOutputReference_Override ¶
func NewBigqueryJobQueryScriptOptionsOutputReference_Override(b BigqueryJobQueryScriptOptionsOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJobQueryUserDefinedFunctionResourcesList_Override ¶
func NewBigqueryJobQueryUserDefinedFunctionResourcesList_Override(b BigqueryJobQueryUserDefinedFunctionResourcesList, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool)
func NewBigqueryJobQueryUserDefinedFunctionResourcesOutputReference_Override ¶
func NewBigqueryJobQueryUserDefinedFunctionResourcesOutputReference_Override(b BigqueryJobQueryUserDefinedFunctionResourcesOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool)
func NewBigqueryJobStatusErrorResultList_Override ¶
func NewBigqueryJobStatusErrorResultList_Override(b BigqueryJobStatusErrorResultList, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool)
func NewBigqueryJobStatusErrorResultOutputReference_Override ¶
func NewBigqueryJobStatusErrorResultOutputReference_Override(b BigqueryJobStatusErrorResultOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool)
func NewBigqueryJobStatusErrorsList_Override ¶
func NewBigqueryJobStatusErrorsList_Override(b BigqueryJobStatusErrorsList, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool)
func NewBigqueryJobStatusErrorsOutputReference_Override ¶
func NewBigqueryJobStatusErrorsOutputReference_Override(b BigqueryJobStatusErrorsOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool)
func NewBigqueryJobStatusList_Override ¶
func NewBigqueryJobStatusList_Override(b BigqueryJobStatusList, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool)
func NewBigqueryJobStatusOutputReference_Override ¶
func NewBigqueryJobStatusOutputReference_Override(b BigqueryJobStatusOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool)
func NewBigqueryJobTimeoutsOutputReference_Override ¶
func NewBigqueryJobTimeoutsOutputReference_Override(b BigqueryJobTimeoutsOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewBigqueryJob_Override ¶
func NewBigqueryJob_Override(b BigqueryJob, scope constructs.Construct, id *string, config *BigqueryJobConfig)
Create a new {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job google_bigquery_job} Resource.
Types ¶
type BigqueryJob ¶
type BigqueryJob interface {
cdktf.TerraformResource
// Experimental.
CdktfStack() cdktf.TerraformStack
// Experimental.
Connection() interface{}
// Experimental.
SetConnection(val interface{})
// Experimental.
ConstructNodeMetadata() *map[string]interface{}
Copy() BigqueryJobCopyOutputReference
CopyInput() *BigqueryJobCopy
// Experimental.
Count() interface{}
// Experimental.
SetCount(val interface{})
// Experimental.
DependsOn() *[]*string
// Experimental.
SetDependsOn(val *[]*string)
EffectiveLabels() cdktf.StringMap
Extract() BigqueryJobExtractOutputReference
ExtractInput() *BigqueryJobExtract
// Experimental.
ForEach() cdktf.ITerraformIterator
// Experimental.
SetForEach(val cdktf.ITerraformIterator)
// Experimental.
Fqn() *string
// Experimental.
FriendlyUniqueId() *string
Id() *string
SetId(val *string)
IdInput() *string
JobId() *string
SetJobId(val *string)
JobIdInput() *string
JobTimeoutMs() *string
SetJobTimeoutMs(val *string)
JobTimeoutMsInput() *string
JobType() *string
Labels() *map[string]*string
SetLabels(val *map[string]*string)
LabelsInput() *map[string]*string
// Experimental.
Lifecycle() *cdktf.TerraformResourceLifecycle
// Experimental.
SetLifecycle(val *cdktf.TerraformResourceLifecycle)
Load() BigqueryJobLoadOutputReference
LoadInput() *BigqueryJobLoad
Location() *string
SetLocation(val *string)
LocationInput() *string
// The tree node.
Node() constructs.Node
Project() *string
SetProject(val *string)
ProjectInput() *string
// Experimental.
Provider() cdktf.TerraformProvider
// Experimental.
SetProvider(val cdktf.TerraformProvider)
// Experimental.
Provisioners() *[]interface{}
// Experimental.
SetProvisioners(val *[]interface{})
Query() BigqueryJobQueryOutputReference
QueryInput() *BigqueryJobQuery
// Experimental.
RawOverrides() interface{}
Status() BigqueryJobStatusList
// Experimental.
TerraformGeneratorMetadata() *cdktf.TerraformProviderGeneratorMetadata
TerraformLabels() cdktf.StringMap
// Experimental.
TerraformMetaArguments() *map[string]interface{}
// Experimental.
TerraformResourceType() *string
Timeouts() BigqueryJobTimeoutsOutputReference
TimeoutsInput() interface{}
UserEmail() *string
// Adds a user defined moveTarget string to this resource to be later used in .moveTo(moveTarget) to resolve the location of the move.
// Experimental.
AddMoveTarget(moveTarget *string)
// Experimental.
AddOverride(path *string, value interface{})
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
ImportFrom(id *string, provider cdktf.TerraformProvider)
// Experimental.
InterpolationForAttribute(terraformAttribute *string) cdktf.IResolvable
// Moves this resource to the target resource given by moveTarget.
// Experimental.
MoveTo(moveTarget *string, index interface{})
// Overrides the auto-generated logical ID with a specific ID.
// Experimental.
OverrideLogicalId(newLogicalId *string)
PutCopy(value *BigqueryJobCopy)
PutExtract(value *BigqueryJobExtract)
PutLoad(value *BigqueryJobLoad)
PutQuery(value *BigqueryJobQuery)
PutTimeouts(value *BigqueryJobTimeouts)
ResetCopy()
ResetExtract()
ResetId()
ResetJobTimeoutMs()
ResetLabels()
ResetLoad()
ResetLocation()
// Resets a previously passed logical Id to use the auto-generated logical id again.
// Experimental.
ResetOverrideLogicalId()
ResetProject()
ResetQuery()
ResetTimeouts()
SynthesizeAttributes() *map[string]interface{}
// Experimental.
ToMetadata() interface{}
// Returns a string representation of this construct.
ToString() *string
// Adds this resource to the terraform JSON output.
// Experimental.
ToTerraform() interface{}
}
Represents a {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job google_bigquery_job}.
func NewBigqueryJob ¶
func NewBigqueryJob(scope constructs.Construct, id *string, config *BigqueryJobConfig) BigqueryJob
Create a new {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job google_bigquery_job} Resource.
type BigqueryJobConfig ¶
type BigqueryJobConfig struct {
// Experimental.
Connection interface{} `field:"optional" json:"connection" yaml:"connection"`
// Experimental.
Count interface{} `field:"optional" json:"count" yaml:"count"`
// Experimental.
DependsOn *[]cdktf.ITerraformDependable `field:"optional" json:"dependsOn" yaml:"dependsOn"`
// Experimental.
ForEach cdktf.ITerraformIterator `field:"optional" json:"forEach" yaml:"forEach"`
// Experimental.
Lifecycle *cdktf.TerraformResourceLifecycle `field:"optional" json:"lifecycle" yaml:"lifecycle"`
// Experimental.
Provider cdktf.TerraformProvider `field:"optional" json:"provider" yaml:"provider"`
// Experimental.
Provisioners *[]interface{} `field:"optional" json:"provisioners" yaml:"provisioners"`
// The ID of the job.
//
// The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#job_id BigqueryJob#job_id}
JobId *string `field:"required" json:"jobId" yaml:"jobId"`
// copy block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#copy BigqueryJob#copy}
Copy *BigqueryJobCopy `field:"optional" json:"copy" yaml:"copy"`
// extract block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#extract BigqueryJob#extract}
Extract *BigqueryJobExtract `field:"optional" json:"extract" yaml:"extract"`
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#id BigqueryJob#id}.
//
// Please be aware that the id field is automatically added to all resources in Terraform providers using a Terraform provider SDK version below 2.
// If you experience problems setting this value it might not be settable. Please take a look at the provider documentation to ensure it should be settable.
Id *string `field:"optional" json:"id" yaml:"id"`
// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#job_timeout_ms BigqueryJob#job_timeout_ms}
JobTimeoutMs *string `field:"optional" json:"jobTimeoutMs" yaml:"jobTimeoutMs"`
// The labels associated with this job. You can use these to organize and group your jobs.
//
// **Note**: This field is non-authoritative, and will only manage the labels present in your configuration.
// Please refer to the field 'effective_labels' for all of the labels present on the resource.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#labels BigqueryJob#labels}
Labels *map[string]*string `field:"optional" json:"labels" yaml:"labels"`
// load block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#load BigqueryJob#load}
Load *BigqueryJobLoad `field:"optional" json:"load" yaml:"load"`
// The geographic location of the job. The default value is US.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#location BigqueryJob#location}
Location *string `field:"optional" json:"location" yaml:"location"`
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project BigqueryJob#project}.
Project *string `field:"optional" json:"project" yaml:"project"`
// query block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#query BigqueryJob#query}
Query *BigqueryJobQuery `field:"optional" json:"query" yaml:"query"`
// timeouts block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#timeouts BigqueryJob#timeouts}
Timeouts *BigqueryJobTimeouts `field:"optional" json:"timeouts" yaml:"timeouts"`
}
type BigqueryJobCopy ¶
type BigqueryJobCopy struct {
// source_tables block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#source_tables BigqueryJob#source_tables}
SourceTables interface{} `field:"required" json:"sourceTables" yaml:"sourceTables"`
// Specifies whether the job is allowed to create new tables.
//
// The following values are supported:
// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
// Creation, truncation and append actions occur as one atomic update upon job completion Default value: "CREATE_IF_NEEDED" Possible values: ["CREATE_IF_NEEDED", "CREATE_NEVER"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#create_disposition BigqueryJob#create_disposition}
CreateDisposition *string `field:"optional" json:"createDisposition" yaml:"createDisposition"`
// destination_encryption_configuration block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_encryption_configuration BigqueryJob#destination_encryption_configuration}
DestinationEncryptionConfiguration *BigqueryJobCopyDestinationEncryptionConfiguration `field:"optional" json:"destinationEncryptionConfiguration" yaml:"destinationEncryptionConfiguration"`
// destination_table block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_table BigqueryJob#destination_table}
DestinationTable *BigqueryJobCopyDestinationTable `field:"optional" json:"destinationTable" yaml:"destinationTable"`
// Specifies the action that occurs if the destination table already exists.
//
// The following values are supported:
// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
// Creation, truncation and append actions occur as one atomic update upon job completion. Default value: "WRITE_EMPTY" Possible values: ["WRITE_TRUNCATE", "WRITE_APPEND", "WRITE_EMPTY"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#write_disposition BigqueryJob#write_disposition}
WriteDisposition *string `field:"optional" json:"writeDisposition" yaml:"writeDisposition"`
}
type BigqueryJobCopyDestinationEncryptionConfiguration ¶
type BigqueryJobCopyDestinationEncryptionConfiguration struct {
// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
//
// The BigQuery Service Account associated with your project requires access to this encryption key.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#kms_key_name BigqueryJob#kms_key_name}
KmsKeyName *string `field:"required" json:"kmsKeyName" yaml:"kmsKeyName"`
}
type BigqueryJobCopyDestinationEncryptionConfigurationOutputReference ¶
type BigqueryJobCopyDestinationEncryptionConfigurationOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobCopyDestinationEncryptionConfiguration
SetInternalValue(val *BigqueryJobCopyDestinationEncryptionConfiguration)
KmsKeyName() *string
SetKmsKeyName(val *string)
KmsKeyNameInput() *string
KmsKeyVersion() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobCopyDestinationEncryptionConfigurationOutputReference ¶
func NewBigqueryJobCopyDestinationEncryptionConfigurationOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobCopyDestinationEncryptionConfigurationOutputReference
type BigqueryJobCopyDestinationTable ¶
type BigqueryJobCopyDestinationTable struct {
// The table. Can be specified '{{table_id}}' if 'project_id' and 'dataset_id' are also set, or of the form 'projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}' if not.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#table_id BigqueryJob#table_id}
TableId *string `field:"required" json:"tableId" yaml:"tableId"`
// The ID of the dataset containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"optional" json:"datasetId" yaml:"datasetId"`
// The ID of the project containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"optional" json:"projectId" yaml:"projectId"`
}
type BigqueryJobCopyDestinationTableOutputReference ¶
type BigqueryJobCopyDestinationTableOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobCopyDestinationTable
SetInternalValue(val *BigqueryJobCopyDestinationTable)
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
TableId() *string
SetTableId(val *string)
TableIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetDatasetId()
ResetProjectId()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobCopyDestinationTableOutputReference ¶
func NewBigqueryJobCopyDestinationTableOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobCopyDestinationTableOutputReference
type BigqueryJobCopyOutputReference ¶
type BigqueryJobCopyOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
CreateDisposition() *string
SetCreateDisposition(val *string)
CreateDispositionInput() *string
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DestinationEncryptionConfiguration() BigqueryJobCopyDestinationEncryptionConfigurationOutputReference
DestinationEncryptionConfigurationInput() *BigqueryJobCopyDestinationEncryptionConfiguration
DestinationTable() BigqueryJobCopyDestinationTableOutputReference
DestinationTableInput() *BigqueryJobCopyDestinationTable
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobCopy
SetInternalValue(val *BigqueryJobCopy)
SourceTables() BigqueryJobCopySourceTablesList
SourceTablesInput() interface{}
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
WriteDisposition() *string
SetWriteDisposition(val *string)
WriteDispositionInput() *string
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
PutDestinationEncryptionConfiguration(value *BigqueryJobCopyDestinationEncryptionConfiguration)
PutDestinationTable(value *BigqueryJobCopyDestinationTable)
PutSourceTables(value interface{})
ResetCreateDisposition()
ResetDestinationEncryptionConfiguration()
ResetDestinationTable()
ResetWriteDisposition()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobCopyOutputReference ¶
func NewBigqueryJobCopyOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobCopyOutputReference
type BigqueryJobCopySourceTables ¶
type BigqueryJobCopySourceTables struct {
// The table. Can be specified '{{table_id}}' if 'project_id' and 'dataset_id' are also set, or of the form 'projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}' if not.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#table_id BigqueryJob#table_id}
TableId *string `field:"required" json:"tableId" yaml:"tableId"`
// The ID of the dataset containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"optional" json:"datasetId" yaml:"datasetId"`
// The ID of the project containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"optional" json:"projectId" yaml:"projectId"`
}
type BigqueryJobCopySourceTablesList ¶
type BigqueryJobCopySourceTablesList interface {
cdktf.ComplexList
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() interface{}
SetInternalValue(val interface{})
// The attribute on the parent resource this class is referencing.
TerraformAttribute() *string
SetTerraformAttribute(val *string)
// The parent resource.
TerraformResource() cdktf.IInterpolatingParent
SetTerraformResource(val cdktf.IInterpolatingParent)
// whether the list is wrapping a set (will add tolist() to be able to access an item via an index).
WrapsSet() *bool
SetWrapsSet(val *bool)
// Experimental.
ComputeFqn() *string
Get(index *float64) BigqueryJobCopySourceTablesOutputReference
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobCopySourceTablesList ¶
func NewBigqueryJobCopySourceTablesList(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool) BigqueryJobCopySourceTablesList
type BigqueryJobCopySourceTablesOutputReference ¶
type BigqueryJobCopySourceTablesOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() interface{}
SetInternalValue(val interface{})
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
TableId() *string
SetTableId(val *string)
TableIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetDatasetId()
ResetProjectId()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobCopySourceTablesOutputReference ¶
func NewBigqueryJobCopySourceTablesOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool) BigqueryJobCopySourceTablesOutputReference
type BigqueryJobExtract ¶
type BigqueryJobExtract struct {
// A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_uris BigqueryJob#destination_uris}
DestinationUris *[]*string `field:"required" json:"destinationUris" yaml:"destinationUris"`
// The compression type to use for exported files.
//
// Possible values include GZIP, DEFLATE, SNAPPY, and NONE.
// The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#compression BigqueryJob#compression}
Compression *string `field:"optional" json:"compression" yaml:"compression"`
// The exported file format.
//
// Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models.
// The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV.
// The default value for models is SAVED_MODEL.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_format BigqueryJob#destination_format}
DestinationFormat *string `field:"optional" json:"destinationFormat" yaml:"destinationFormat"`
// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
//
// Default is ','
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#field_delimiter BigqueryJob#field_delimiter}
FieldDelimiter *string `field:"optional" json:"fieldDelimiter" yaml:"fieldDelimiter"`
// Whether to print out a header row in the results. Default is true.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#print_header BigqueryJob#print_header}
PrintHeader interface{} `field:"optional" json:"printHeader" yaml:"printHeader"`
// source_model block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#source_model BigqueryJob#source_model}
SourceModel *BigqueryJobExtractSourceModel `field:"optional" json:"sourceModel" yaml:"sourceModel"`
// source_table block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#source_table BigqueryJob#source_table}
SourceTable *BigqueryJobExtractSourceTable `field:"optional" json:"sourceTable" yaml:"sourceTable"`
// Whether to use logical types when extracting to AVRO format.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#use_avro_logical_types BigqueryJob#use_avro_logical_types}
UseAvroLogicalTypes interface{} `field:"optional" json:"useAvroLogicalTypes" yaml:"useAvroLogicalTypes"`
}
type BigqueryJobExtractOutputReference ¶
type BigqueryJobExtractOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
Compression() *string
SetCompression(val *string)
CompressionInput() *string
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DestinationFormat() *string
SetDestinationFormat(val *string)
DestinationFormatInput() *string
DestinationUris() *[]*string
SetDestinationUris(val *[]*string)
DestinationUrisInput() *[]*string
FieldDelimiter() *string
SetFieldDelimiter(val *string)
FieldDelimiterInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobExtract
SetInternalValue(val *BigqueryJobExtract)
PrintHeader() interface{}
SetPrintHeader(val interface{})
PrintHeaderInput() interface{}
SourceModel() BigqueryJobExtractSourceModelOutputReference
SourceModelInput() *BigqueryJobExtractSourceModel
SourceTable() BigqueryJobExtractSourceTableOutputReference
SourceTableInput() *BigqueryJobExtractSourceTable
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
UseAvroLogicalTypes() interface{}
SetUseAvroLogicalTypes(val interface{})
UseAvroLogicalTypesInput() interface{}
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
PutSourceModel(value *BigqueryJobExtractSourceModel)
PutSourceTable(value *BigqueryJobExtractSourceTable)
ResetCompression()
ResetDestinationFormat()
ResetFieldDelimiter()
ResetPrintHeader()
ResetSourceModel()
ResetSourceTable()
ResetUseAvroLogicalTypes()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobExtractOutputReference ¶
func NewBigqueryJobExtractOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobExtractOutputReference
type BigqueryJobExtractSourceModel ¶
type BigqueryJobExtractSourceModel struct {
// The ID of the dataset containing this model.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"required" json:"datasetId" yaml:"datasetId"`
// The ID of the model.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#model_id BigqueryJob#model_id}
ModelId *string `field:"required" json:"modelId" yaml:"modelId"`
// The ID of the project containing this model.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"required" json:"projectId" yaml:"projectId"`
}
type BigqueryJobExtractSourceModelOutputReference ¶
type BigqueryJobExtractSourceModelOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobExtractSourceModel
SetInternalValue(val *BigqueryJobExtractSourceModel)
ModelId() *string
SetModelId(val *string)
ModelIdInput() *string
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobExtractSourceModelOutputReference ¶
func NewBigqueryJobExtractSourceModelOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobExtractSourceModelOutputReference
type BigqueryJobExtractSourceTable ¶
type BigqueryJobExtractSourceTable struct {
// The table. Can be specified '{{table_id}}' if 'project_id' and 'dataset_id' are also set, or of the form 'projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}' if not.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#table_id BigqueryJob#table_id}
TableId *string `field:"required" json:"tableId" yaml:"tableId"`
// The ID of the dataset containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"optional" json:"datasetId" yaml:"datasetId"`
// The ID of the project containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"optional" json:"projectId" yaml:"projectId"`
}
type BigqueryJobExtractSourceTableOutputReference ¶
type BigqueryJobExtractSourceTableOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobExtractSourceTable
SetInternalValue(val *BigqueryJobExtractSourceTable)
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
TableId() *string
SetTableId(val *string)
TableIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetDatasetId()
ResetProjectId()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobExtractSourceTableOutputReference ¶
func NewBigqueryJobExtractSourceTableOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobExtractSourceTableOutputReference
type BigqueryJobLoad ¶
type BigqueryJobLoad struct {
// destination_table block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_table BigqueryJob#destination_table}
DestinationTable *BigqueryJobLoadDestinationTable `field:"required" json:"destinationTable" yaml:"destinationTable"`
// The fully-qualified URIs that point to your data in Google Cloud.
//
// For Google Cloud Storage URIs: Each URI can contain one '\*' wildcard character
// and it must come after the 'bucket' name. Size limits related to load jobs apply
// to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be
// specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table.
// For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '\*' wildcard character is not allowed.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#source_uris BigqueryJob#source_uris}
SourceUris *[]*string `field:"required" json:"sourceUris" yaml:"sourceUris"`
// Accept rows that are missing trailing optional columns.
//
// The missing values are treated as nulls.
// If false, records with missing trailing columns are treated as bad records, and if there are too many bad records,
// an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#allow_jagged_rows BigqueryJob#allow_jagged_rows}
AllowJaggedRows interface{} `field:"optional" json:"allowJaggedRows" yaml:"allowJaggedRows"`
// Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file.
//
// The default value is false.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#allow_quoted_newlines BigqueryJob#allow_quoted_newlines}
AllowQuotedNewlines interface{} `field:"optional" json:"allowQuotedNewlines" yaml:"allowQuotedNewlines"`
// Indicates if we should automatically infer the options and schema for CSV and JSON sources.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#autodetect BigqueryJob#autodetect}
Autodetect interface{} `field:"optional" json:"autodetect" yaml:"autodetect"`
// Specifies whether the job is allowed to create new tables.
//
// The following values are supported:
// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
// Creation, truncation and append actions occur as one atomic update upon job completion Default value: "CREATE_IF_NEEDED" Possible values: ["CREATE_IF_NEEDED", "CREATE_NEVER"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#create_disposition BigqueryJob#create_disposition}
CreateDisposition *string `field:"optional" json:"createDisposition" yaml:"createDisposition"`
// destination_encryption_configuration block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_encryption_configuration BigqueryJob#destination_encryption_configuration}
DestinationEncryptionConfiguration *BigqueryJobLoadDestinationEncryptionConfiguration `field:"optional" json:"destinationEncryptionConfiguration" yaml:"destinationEncryptionConfiguration"`
// The character encoding of the data.
//
// The supported values are UTF-8 or ISO-8859-1.
// The default value is UTF-8. BigQuery decodes the data after the raw, binary data
// has been split using the values of the quote and fieldDelimiter properties.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#encoding BigqueryJob#encoding}
Encoding *string `field:"optional" json:"encoding" yaml:"encoding"`
// The separator for fields in a CSV file.
//
// The separator can be any ISO-8859-1 single-byte character.
// To use a character in the range 128-255, you must encode the character as UTF8. BigQuery converts
// the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the
// data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator.
// The default value is a comma (',').
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#field_delimiter BigqueryJob#field_delimiter}
FieldDelimiter *string `field:"optional" json:"fieldDelimiter" yaml:"fieldDelimiter"`
// Indicates if BigQuery should allow extra values that are not represented in the table schema.
//
// If true, the extra values are ignored. If false, records with extra columns are treated as bad records,
// and if there are too many bad records, an invalid error is returned in the job result.
// The default value is false. The sourceFormat property determines what BigQuery treats as an extra value:
// CSV: Trailing columns
// JSON: Named values that don't match any column names
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#ignore_unknown_values BigqueryJob#ignore_unknown_values}
IgnoreUnknownValues interface{} `field:"optional" json:"ignoreUnknownValues" yaml:"ignoreUnknownValues"`
// If sourceFormat is set to newline-delimited JSON, indicates whether it should be processed as a JSON variant such as GeoJSON.
//
// For a sourceFormat other than JSON, omit this field. If the sourceFormat is newline-delimited JSON: - for newline-delimited
// GeoJSON: set to GEOJSON.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#json_extension BigqueryJob#json_extension}
JsonExtension *string `field:"optional" json:"jsonExtension" yaml:"jsonExtension"`
// The maximum number of bad records that BigQuery can ignore when running the job.
//
// If the number of bad records exceeds this value,
// an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#max_bad_records BigqueryJob#max_bad_records}
MaxBadRecords *float64 `field:"optional" json:"maxBadRecords" yaml:"maxBadRecords"`
// Specifies a string that represents a null value in a CSV file.
//
// For example, if you specify "\N", BigQuery interprets "\N" as a null value
// when loading a CSV file. The default value is the empty string. If you set this property to a custom value, BigQuery throws an error if an
// empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as
// an empty value.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#null_marker BigqueryJob#null_marker}
NullMarker *string `field:"optional" json:"nullMarker" yaml:"nullMarker"`
// parquet_options block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#parquet_options BigqueryJob#parquet_options}
ParquetOptions *BigqueryJobLoadParquetOptions `field:"optional" json:"parquetOptions" yaml:"parquetOptions"`
// If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup.
//
// Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties.
// If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#projection_fields BigqueryJob#projection_fields}
ProjectionFields *[]*string `field:"optional" json:"projectionFields" yaml:"projectionFields"`
// The value that is used to quote data sections in a CSV file.
//
// BigQuery converts the string to ISO-8859-1 encoding,
// and then uses the first byte of the encoded string to split the data in its raw, binary state.
// The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string.
// If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#quote BigqueryJob#quote}
Quote *string `field:"optional" json:"quote" yaml:"quote"`
// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration.
//
// Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#schema_update_options BigqueryJob#schema_update_options}
SchemaUpdateOptions *[]*string `field:"optional" json:"schemaUpdateOptions" yaml:"schemaUpdateOptions"`
// The number of rows at the top of a CSV file that BigQuery will skip when loading the data.
//
// The default value is 0. This property is useful if you have header rows in the file that should be skipped.
// When autodetect is on, the behavior is the following:
// skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected,
// the row is read as data. Otherwise data is read starting from the second row.
// skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row.
// skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected,
// row N is just skipped. Otherwise row N is used to extract column names for the detected schema.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#skip_leading_rows BigqueryJob#skip_leading_rows}
SkipLeadingRows *float64 `field:"optional" json:"skipLeadingRows" yaml:"skipLeadingRows"`
// The format of the data files.
//
// For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP".
// For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET".
// For orc, specify "ORC". [Beta] For Bigtable, specify "BIGTABLE".
// The default value is CSV.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#source_format BigqueryJob#source_format}
SourceFormat *string `field:"optional" json:"sourceFormat" yaml:"sourceFormat"`
// time_partitioning block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#time_partitioning BigqueryJob#time_partitioning}
TimePartitioning *BigqueryJobLoadTimePartitioning `field:"optional" json:"timePartitioning" yaml:"timePartitioning"`
// Specifies the action that occurs if the destination table already exists.
//
// The following values are supported:
// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
// Creation, truncation and append actions occur as one atomic update upon job completion. Default value: "WRITE_EMPTY" Possible values: ["WRITE_TRUNCATE", "WRITE_APPEND", "WRITE_EMPTY"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#write_disposition BigqueryJob#write_disposition}
WriteDisposition *string `field:"optional" json:"writeDisposition" yaml:"writeDisposition"`
}
type BigqueryJobLoadDestinationEncryptionConfiguration ¶
type BigqueryJobLoadDestinationEncryptionConfiguration struct {
// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
//
// The BigQuery Service Account associated with your project requires access to this encryption key.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#kms_key_name BigqueryJob#kms_key_name}
KmsKeyName *string `field:"required" json:"kmsKeyName" yaml:"kmsKeyName"`
}
type BigqueryJobLoadDestinationEncryptionConfigurationOutputReference ¶
type BigqueryJobLoadDestinationEncryptionConfigurationOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobLoadDestinationEncryptionConfiguration
SetInternalValue(val *BigqueryJobLoadDestinationEncryptionConfiguration)
KmsKeyName() *string
SetKmsKeyName(val *string)
KmsKeyNameInput() *string
KmsKeyVersion() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobLoadDestinationEncryptionConfigurationOutputReference ¶
func NewBigqueryJobLoadDestinationEncryptionConfigurationOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobLoadDestinationEncryptionConfigurationOutputReference
type BigqueryJobLoadDestinationTable ¶
type BigqueryJobLoadDestinationTable struct {
// The table. Can be specified '{{table_id}}' if 'project_id' and 'dataset_id' are also set, or of the form 'projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}' if not.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#table_id BigqueryJob#table_id}
TableId *string `field:"required" json:"tableId" yaml:"tableId"`
// The ID of the dataset containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"optional" json:"datasetId" yaml:"datasetId"`
// The ID of the project containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"optional" json:"projectId" yaml:"projectId"`
}
type BigqueryJobLoadDestinationTableOutputReference ¶
type BigqueryJobLoadDestinationTableOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobLoadDestinationTable
SetInternalValue(val *BigqueryJobLoadDestinationTable)
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
TableId() *string
SetTableId(val *string)
TableIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetDatasetId()
ResetProjectId()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobLoadDestinationTableOutputReference ¶
func NewBigqueryJobLoadDestinationTableOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobLoadDestinationTableOutputReference
type BigqueryJobLoadOutputReference ¶
type BigqueryJobLoadOutputReference interface {
cdktf.ComplexObject
AllowJaggedRows() interface{}
SetAllowJaggedRows(val interface{})
AllowJaggedRowsInput() interface{}
AllowQuotedNewlines() interface{}
SetAllowQuotedNewlines(val interface{})
AllowQuotedNewlinesInput() interface{}
Autodetect() interface{}
SetAutodetect(val interface{})
AutodetectInput() interface{}
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
CreateDisposition() *string
SetCreateDisposition(val *string)
CreateDispositionInput() *string
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DestinationEncryptionConfiguration() BigqueryJobLoadDestinationEncryptionConfigurationOutputReference
DestinationEncryptionConfigurationInput() *BigqueryJobLoadDestinationEncryptionConfiguration
DestinationTable() BigqueryJobLoadDestinationTableOutputReference
DestinationTableInput() *BigqueryJobLoadDestinationTable
Encoding() *string
SetEncoding(val *string)
EncodingInput() *string
FieldDelimiter() *string
SetFieldDelimiter(val *string)
FieldDelimiterInput() *string
// Experimental.
Fqn() *string
IgnoreUnknownValues() interface{}
SetIgnoreUnknownValues(val interface{})
IgnoreUnknownValuesInput() interface{}
InternalValue() *BigqueryJobLoad
SetInternalValue(val *BigqueryJobLoad)
JsonExtension() *string
SetJsonExtension(val *string)
JsonExtensionInput() *string
MaxBadRecords() *float64
SetMaxBadRecords(val *float64)
MaxBadRecordsInput() *float64
NullMarker() *string
SetNullMarker(val *string)
NullMarkerInput() *string
ParquetOptions() BigqueryJobLoadParquetOptionsOutputReference
ParquetOptionsInput() *BigqueryJobLoadParquetOptions
ProjectionFields() *[]*string
SetProjectionFields(val *[]*string)
ProjectionFieldsInput() *[]*string
Quote() *string
SetQuote(val *string)
QuoteInput() *string
SchemaUpdateOptions() *[]*string
SetSchemaUpdateOptions(val *[]*string)
SchemaUpdateOptionsInput() *[]*string
SkipLeadingRows() *float64
SetSkipLeadingRows(val *float64)
SkipLeadingRowsInput() *float64
SourceFormat() *string
SetSourceFormat(val *string)
SourceFormatInput() *string
SourceUris() *[]*string
SetSourceUris(val *[]*string)
SourceUrisInput() *[]*string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
TimePartitioning() BigqueryJobLoadTimePartitioningOutputReference
TimePartitioningInput() *BigqueryJobLoadTimePartitioning
WriteDisposition() *string
SetWriteDisposition(val *string)
WriteDispositionInput() *string
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
PutDestinationEncryptionConfiguration(value *BigqueryJobLoadDestinationEncryptionConfiguration)
PutDestinationTable(value *BigqueryJobLoadDestinationTable)
PutParquetOptions(value *BigqueryJobLoadParquetOptions)
PutTimePartitioning(value *BigqueryJobLoadTimePartitioning)
ResetAllowJaggedRows()
ResetAllowQuotedNewlines()
ResetAutodetect()
ResetCreateDisposition()
ResetDestinationEncryptionConfiguration()
ResetEncoding()
ResetFieldDelimiter()
ResetIgnoreUnknownValues()
ResetJsonExtension()
ResetMaxBadRecords()
ResetNullMarker()
ResetParquetOptions()
ResetProjectionFields()
ResetQuote()
ResetSchemaUpdateOptions()
ResetSkipLeadingRows()
ResetSourceFormat()
ResetTimePartitioning()
ResetWriteDisposition()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobLoadOutputReference ¶
func NewBigqueryJobLoadOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobLoadOutputReference
type BigqueryJobLoadParquetOptions ¶
type BigqueryJobLoadParquetOptions struct {
// If sourceFormat is set to PARQUET, indicates whether to use schema inference specifically for Parquet LIST logical type.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#enable_list_inference BigqueryJob#enable_list_inference}
EnableListInference interface{} `field:"optional" json:"enableListInference" yaml:"enableListInference"`
// If sourceFormat is set to PARQUET, indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#enum_as_string BigqueryJob#enum_as_string}
EnumAsString interface{} `field:"optional" json:"enumAsString" yaml:"enumAsString"`
}
type BigqueryJobLoadParquetOptionsOutputReference ¶
type BigqueryJobLoadParquetOptionsOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
EnableListInference() interface{}
SetEnableListInference(val interface{})
EnableListInferenceInput() interface{}
EnumAsString() interface{}
SetEnumAsString(val interface{})
EnumAsStringInput() interface{}
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobLoadParquetOptions
SetInternalValue(val *BigqueryJobLoadParquetOptions)
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetEnableListInference()
ResetEnumAsString()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobLoadParquetOptionsOutputReference ¶
func NewBigqueryJobLoadParquetOptionsOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobLoadParquetOptionsOutputReference
type BigqueryJobLoadTimePartitioning ¶
type BigqueryJobLoadTimePartitioning struct {
// The only type supported is DAY, which will generate one partition per day.
//
// Providing an empty string used to cause an error,
// but in OnePlatform the field will be treated as unset.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#type BigqueryJob#type}
Type *string `field:"required" json:"type" yaml:"type"`
// Number of milliseconds for which to keep the storage for a partition.
//
// A wrapper is used here because 0 is an invalid value.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#expiration_ms BigqueryJob#expiration_ms}
ExpirationMs *string `field:"optional" json:"expirationMs" yaml:"expirationMs"`
// If not set, the table is partitioned by pseudo column '_PARTITIONTIME';
//
// if set, the table is partitioned by this field.
// The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED.
// A wrapper is used here because an empty string is an invalid value.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#field BigqueryJob#field}
Field *string `field:"optional" json:"field" yaml:"field"`
}
type BigqueryJobLoadTimePartitioningOutputReference ¶
type BigqueryJobLoadTimePartitioningOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
ExpirationMs() *string
SetExpirationMs(val *string)
ExpirationMsInput() *string
Field() *string
SetField(val *string)
FieldInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobLoadTimePartitioning
SetInternalValue(val *BigqueryJobLoadTimePartitioning)
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
Type() *string
SetType(val *string)
TypeInput() *string
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetExpirationMs()
ResetField()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobLoadTimePartitioningOutputReference ¶
func NewBigqueryJobLoadTimePartitioningOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobLoadTimePartitioningOutputReference
type BigqueryJobQuery ¶
type BigqueryJobQuery struct {
// SQL query text to execute.
//
// The useLegacySql field can be used to indicate whether the query uses legacy SQL or standard SQL.
// *NOTE*: queries containing [DML language](https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language)
// ('DELETE', 'UPDATE', 'MERGE', 'INSERT') must specify 'create_disposition = ""' and 'write_disposition = ""'.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#query BigqueryJob#query}
Query *string `field:"required" json:"query" yaml:"query"`
// If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance.
//
// Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed.
// However, you must still set destinationTable when result size exceeds the allowed maximum response size.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#allow_large_results BigqueryJob#allow_large_results}
AllowLargeResults interface{} `field:"optional" json:"allowLargeResults" yaml:"allowLargeResults"`
// Specifies whether the job is allowed to create new tables.
//
// The following values are supported:
// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
// Creation, truncation and append actions occur as one atomic update upon job completion Default value: "CREATE_IF_NEEDED" Possible values: ["CREATE_IF_NEEDED", "CREATE_NEVER"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#create_disposition BigqueryJob#create_disposition}
CreateDisposition *string `field:"optional" json:"createDisposition" yaml:"createDisposition"`
// default_dataset block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#default_dataset BigqueryJob#default_dataset}
DefaultDataset *BigqueryJobQueryDefaultDataset `field:"optional" json:"defaultDataset" yaml:"defaultDataset"`
// destination_encryption_configuration block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_encryption_configuration BigqueryJob#destination_encryption_configuration}
DestinationEncryptionConfiguration *BigqueryJobQueryDestinationEncryptionConfiguration `field:"optional" json:"destinationEncryptionConfiguration" yaml:"destinationEncryptionConfiguration"`
// destination_table block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#destination_table BigqueryJob#destination_table}
DestinationTable *BigqueryJobQueryDestinationTable `field:"optional" json:"destinationTable" yaml:"destinationTable"`
// If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results.
//
// allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#flatten_results BigqueryJob#flatten_results}
FlattenResults interface{} `field:"optional" json:"flattenResults" yaml:"flattenResults"`
// Limits the billing tier for this job.
//
// Queries that have resource usage beyond this tier will fail (without incurring a charge).
// If unspecified, this will be set to your project default.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#maximum_billing_tier BigqueryJob#maximum_billing_tier}
MaximumBillingTier *float64 `field:"optional" json:"maximumBillingTier" yaml:"maximumBillingTier"`
// Limits the bytes billed for this job.
//
// Queries that will have bytes billed beyond this limit will fail (without incurring a charge).
// If unspecified, this will be set to your project default.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#maximum_bytes_billed BigqueryJob#maximum_bytes_billed}
MaximumBytesBilled *string `field:"optional" json:"maximumBytesBilled" yaml:"maximumBytesBilled"`
// Standard SQL only.
//
// Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#parameter_mode BigqueryJob#parameter_mode}
ParameterMode *string `field:"optional" json:"parameterMode" yaml:"parameterMode"`
// Specifies a priority for the query. Default value: "INTERACTIVE" Possible values: ["INTERACTIVE", "BATCH"].
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#priority BigqueryJob#priority}
Priority *string `field:"optional" json:"priority" yaml:"priority"`
// Allows the schema of the destination table to be updated as a side effect of the query job.
//
// Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table,
// specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema.
// One or more of the following values are specified:
// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#schema_update_options BigqueryJob#schema_update_options}
SchemaUpdateOptions *[]*string `field:"optional" json:"schemaUpdateOptions" yaml:"schemaUpdateOptions"`
// script_options block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#script_options BigqueryJob#script_options}
ScriptOptions *BigqueryJobQueryScriptOptions `field:"optional" json:"scriptOptions" yaml:"scriptOptions"`
// Specifies whether to use BigQuery's legacy SQL dialect for this query.
//
// The default value is true.
// If set to false, the query will use BigQuery's standard SQL.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#use_legacy_sql BigqueryJob#use_legacy_sql}
UseLegacySql interface{} `field:"optional" json:"useLegacySql" yaml:"useLegacySql"`
// Whether to look for the result in the query cache.
//
// The query cache is a best-effort cache that will be flushed whenever
// tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
// The default value is true.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#use_query_cache BigqueryJob#use_query_cache}
UseQueryCache interface{} `field:"optional" json:"useQueryCache" yaml:"useQueryCache"`
// user_defined_function_resources block.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#user_defined_function_resources BigqueryJob#user_defined_function_resources}
UserDefinedFunctionResources interface{} `field:"optional" json:"userDefinedFunctionResources" yaml:"userDefinedFunctionResources"`
// Specifies the action that occurs if the destination table already exists.
//
// The following values are supported:
// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
// Creation, truncation and append actions occur as one atomic update upon job completion. Default value: "WRITE_EMPTY" Possible values: ["WRITE_TRUNCATE", "WRITE_APPEND", "WRITE_EMPTY"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#write_disposition BigqueryJob#write_disposition}
WriteDisposition *string `field:"optional" json:"writeDisposition" yaml:"writeDisposition"`
}
type BigqueryJobQueryDefaultDataset ¶
type BigqueryJobQueryDefaultDataset struct {
// The dataset. Can be specified '{{dataset_id}}' if 'project_id' is also set, or of the form 'projects/{{project}}/datasets/{{dataset_id}}' if not.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"required" json:"datasetId" yaml:"datasetId"`
// The ID of the project containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"optional" json:"projectId" yaml:"projectId"`
}
type BigqueryJobQueryDefaultDatasetOutputReference ¶
type BigqueryJobQueryDefaultDatasetOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobQueryDefaultDataset
SetInternalValue(val *BigqueryJobQueryDefaultDataset)
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetProjectId()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryDefaultDatasetOutputReference ¶
func NewBigqueryJobQueryDefaultDatasetOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobQueryDefaultDatasetOutputReference
type BigqueryJobQueryDestinationEncryptionConfiguration ¶
type BigqueryJobQueryDestinationEncryptionConfiguration struct {
// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
//
// The BigQuery Service Account associated with your project requires access to this encryption key.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#kms_key_name BigqueryJob#kms_key_name}
KmsKeyName *string `field:"required" json:"kmsKeyName" yaml:"kmsKeyName"`
}
type BigqueryJobQueryDestinationEncryptionConfigurationOutputReference ¶
type BigqueryJobQueryDestinationEncryptionConfigurationOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobQueryDestinationEncryptionConfiguration
SetInternalValue(val *BigqueryJobQueryDestinationEncryptionConfiguration)
KmsKeyName() *string
SetKmsKeyName(val *string)
KmsKeyNameInput() *string
KmsKeyVersion() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryDestinationEncryptionConfigurationOutputReference ¶
func NewBigqueryJobQueryDestinationEncryptionConfigurationOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobQueryDestinationEncryptionConfigurationOutputReference
type BigqueryJobQueryDestinationTable ¶
type BigqueryJobQueryDestinationTable struct {
// The table. Can be specified '{{table_id}}' if 'project_id' and 'dataset_id' are also set, or of the form 'projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}' if not.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#table_id BigqueryJob#table_id}
TableId *string `field:"required" json:"tableId" yaml:"tableId"`
// The ID of the dataset containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#dataset_id BigqueryJob#dataset_id}
DatasetId *string `field:"optional" json:"datasetId" yaml:"datasetId"`
// The ID of the project containing this table.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#project_id BigqueryJob#project_id}
ProjectId *string `field:"optional" json:"projectId" yaml:"projectId"`
}
type BigqueryJobQueryDestinationTableOutputReference ¶
type BigqueryJobQueryDestinationTableOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DatasetId() *string
SetDatasetId(val *string)
DatasetIdInput() *string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobQueryDestinationTable
SetInternalValue(val *BigqueryJobQueryDestinationTable)
ProjectId() *string
SetProjectId(val *string)
ProjectIdInput() *string
TableId() *string
SetTableId(val *string)
TableIdInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetDatasetId()
ResetProjectId()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryDestinationTableOutputReference ¶
func NewBigqueryJobQueryDestinationTableOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobQueryDestinationTableOutputReference
type BigqueryJobQueryOutputReference ¶
type BigqueryJobQueryOutputReference interface {
cdktf.ComplexObject
AllowLargeResults() interface{}
SetAllowLargeResults(val interface{})
AllowLargeResultsInput() interface{}
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
CreateDisposition() *string
SetCreateDisposition(val *string)
CreateDispositionInput() *string
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
DefaultDataset() BigqueryJobQueryDefaultDatasetOutputReference
DefaultDatasetInput() *BigqueryJobQueryDefaultDataset
DestinationEncryptionConfiguration() BigqueryJobQueryDestinationEncryptionConfigurationOutputReference
DestinationEncryptionConfigurationInput() *BigqueryJobQueryDestinationEncryptionConfiguration
DestinationTable() BigqueryJobQueryDestinationTableOutputReference
DestinationTableInput() *BigqueryJobQueryDestinationTable
FlattenResults() interface{}
SetFlattenResults(val interface{})
FlattenResultsInput() interface{}
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobQuery
SetInternalValue(val *BigqueryJobQuery)
MaximumBillingTier() *float64
SetMaximumBillingTier(val *float64)
MaximumBillingTierInput() *float64
MaximumBytesBilled() *string
SetMaximumBytesBilled(val *string)
MaximumBytesBilledInput() *string
ParameterMode() *string
SetParameterMode(val *string)
ParameterModeInput() *string
Priority() *string
SetPriority(val *string)
PriorityInput() *string
Query() *string
SetQuery(val *string)
QueryInput() *string
SchemaUpdateOptions() *[]*string
SetSchemaUpdateOptions(val *[]*string)
SchemaUpdateOptionsInput() *[]*string
ScriptOptions() BigqueryJobQueryScriptOptionsOutputReference
ScriptOptionsInput() *BigqueryJobQueryScriptOptions
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
UseLegacySql() interface{}
SetUseLegacySql(val interface{})
UseLegacySqlInput() interface{}
UseQueryCache() interface{}
SetUseQueryCache(val interface{})
UseQueryCacheInput() interface{}
UserDefinedFunctionResources() BigqueryJobQueryUserDefinedFunctionResourcesList
UserDefinedFunctionResourcesInput() interface{}
WriteDisposition() *string
SetWriteDisposition(val *string)
WriteDispositionInput() *string
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
PutDefaultDataset(value *BigqueryJobQueryDefaultDataset)
PutDestinationEncryptionConfiguration(value *BigqueryJobQueryDestinationEncryptionConfiguration)
PutDestinationTable(value *BigqueryJobQueryDestinationTable)
PutScriptOptions(value *BigqueryJobQueryScriptOptions)
PutUserDefinedFunctionResources(value interface{})
ResetAllowLargeResults()
ResetCreateDisposition()
ResetDefaultDataset()
ResetDestinationEncryptionConfiguration()
ResetDestinationTable()
ResetFlattenResults()
ResetMaximumBillingTier()
ResetMaximumBytesBilled()
ResetParameterMode()
ResetPriority()
ResetSchemaUpdateOptions()
ResetScriptOptions()
ResetUseLegacySql()
ResetUseQueryCache()
ResetUserDefinedFunctionResources()
ResetWriteDisposition()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryOutputReference ¶
func NewBigqueryJobQueryOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobQueryOutputReference
type BigqueryJobQueryScriptOptions ¶
type BigqueryJobQueryScriptOptions struct {
// Determines which statement in the script represents the "key result", used to populate the schema and query results of the script job.
//
// Possible values: ["LAST", "FIRST_SELECT"]
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#key_result_statement BigqueryJob#key_result_statement}
KeyResultStatement *string `field:"optional" json:"keyResultStatement" yaml:"keyResultStatement"`
// Limit on the number of bytes billed per statement. Exceeding this budget results in an error.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#statement_byte_budget BigqueryJob#statement_byte_budget}
StatementByteBudget *string `field:"optional" json:"statementByteBudget" yaml:"statementByteBudget"`
// Timeout period for each statement in a script.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#statement_timeout_ms BigqueryJob#statement_timeout_ms}
StatementTimeoutMs *string `field:"optional" json:"statementTimeoutMs" yaml:"statementTimeoutMs"`
}
type BigqueryJobQueryScriptOptionsOutputReference ¶
type BigqueryJobQueryScriptOptionsOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobQueryScriptOptions
SetInternalValue(val *BigqueryJobQueryScriptOptions)
KeyResultStatement() *string
SetKeyResultStatement(val *string)
KeyResultStatementInput() *string
StatementByteBudget() *string
SetStatementByteBudget(val *string)
StatementByteBudgetInput() *string
StatementTimeoutMs() *string
SetStatementTimeoutMs(val *string)
StatementTimeoutMsInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetKeyResultStatement()
ResetStatementByteBudget()
ResetStatementTimeoutMs()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryScriptOptionsOutputReference ¶
func NewBigqueryJobQueryScriptOptionsOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobQueryScriptOptionsOutputReference
type BigqueryJobQueryUserDefinedFunctionResources ¶
type BigqueryJobQueryUserDefinedFunctionResources struct {
// An inline resource that contains code for a user-defined function (UDF).
//
// Providing a inline code resource is equivalent to providing a URI for a file containing the same code.
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#inline_code BigqueryJob#inline_code}
InlineCode *string `field:"optional" json:"inlineCode" yaml:"inlineCode"`
// A code resource to load from a Google Cloud Storage URI (gs://bucket/path).
//
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#resource_uri BigqueryJob#resource_uri}
ResourceUri *string `field:"optional" json:"resourceUri" yaml:"resourceUri"`
}
type BigqueryJobQueryUserDefinedFunctionResourcesList ¶
type BigqueryJobQueryUserDefinedFunctionResourcesList interface {
cdktf.ComplexList
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() interface{}
SetInternalValue(val interface{})
// The attribute on the parent resource this class is referencing.
TerraformAttribute() *string
SetTerraformAttribute(val *string)
// The parent resource.
TerraformResource() cdktf.IInterpolatingParent
SetTerraformResource(val cdktf.IInterpolatingParent)
// whether the list is wrapping a set (will add tolist() to be able to access an item via an index).
WrapsSet() *bool
SetWrapsSet(val *bool)
// Experimental.
ComputeFqn() *string
Get(index *float64) BigqueryJobQueryUserDefinedFunctionResourcesOutputReference
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryUserDefinedFunctionResourcesList ¶
func NewBigqueryJobQueryUserDefinedFunctionResourcesList(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool) BigqueryJobQueryUserDefinedFunctionResourcesList
type BigqueryJobQueryUserDefinedFunctionResourcesOutputReference ¶
type BigqueryJobQueryUserDefinedFunctionResourcesOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InlineCode() *string
SetInlineCode(val *string)
InlineCodeInput() *string
InternalValue() interface{}
SetInternalValue(val interface{})
ResourceUri() *string
SetResourceUri(val *string)
ResourceUriInput() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetInlineCode()
ResetResourceUri()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobQueryUserDefinedFunctionResourcesOutputReference ¶
func NewBigqueryJobQueryUserDefinedFunctionResourcesOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool) BigqueryJobQueryUserDefinedFunctionResourcesOutputReference
type BigqueryJobStatus ¶
type BigqueryJobStatus struct {
}
type BigqueryJobStatusErrorResult ¶
type BigqueryJobStatusErrorResult struct {
}
type BigqueryJobStatusErrorResultList ¶
type BigqueryJobStatusErrorResultList interface {
cdktf.ComplexList
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
// The attribute on the parent resource this class is referencing.
TerraformAttribute() *string
SetTerraformAttribute(val *string)
// The parent resource.
TerraformResource() cdktf.IInterpolatingParent
SetTerraformResource(val cdktf.IInterpolatingParent)
// whether the list is wrapping a set (will add tolist() to be able to access an item via an index).
WrapsSet() *bool
SetWrapsSet(val *bool)
// Experimental.
ComputeFqn() *string
Get(index *float64) BigqueryJobStatusErrorResultOutputReference
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobStatusErrorResultList ¶
func NewBigqueryJobStatusErrorResultList(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool) BigqueryJobStatusErrorResultList
type BigqueryJobStatusErrorResultOutputReference ¶
type BigqueryJobStatusErrorResultOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobStatusErrorResult
SetInternalValue(val *BigqueryJobStatusErrorResult)
Location() *string
Message() *string
Reason() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobStatusErrorResultOutputReference ¶
func NewBigqueryJobStatusErrorResultOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool) BigqueryJobStatusErrorResultOutputReference
type BigqueryJobStatusErrors ¶
type BigqueryJobStatusErrors struct {
}
type BigqueryJobStatusErrorsList ¶
type BigqueryJobStatusErrorsList interface {
cdktf.ComplexList
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
// The attribute on the parent resource this class is referencing.
TerraformAttribute() *string
SetTerraformAttribute(val *string)
// The parent resource.
TerraformResource() cdktf.IInterpolatingParent
SetTerraformResource(val cdktf.IInterpolatingParent)
// whether the list is wrapping a set (will add tolist() to be able to access an item via an index).
WrapsSet() *bool
SetWrapsSet(val *bool)
// Experimental.
ComputeFqn() *string
Get(index *float64) BigqueryJobStatusErrorsOutputReference
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobStatusErrorsList ¶
func NewBigqueryJobStatusErrorsList(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool) BigqueryJobStatusErrorsList
type BigqueryJobStatusErrorsOutputReference ¶
type BigqueryJobStatusErrorsOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobStatusErrors
SetInternalValue(val *BigqueryJobStatusErrors)
Location() *string
Message() *string
Reason() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobStatusErrorsOutputReference ¶
func NewBigqueryJobStatusErrorsOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool) BigqueryJobStatusErrorsOutputReference
type BigqueryJobStatusList ¶
type BigqueryJobStatusList interface {
cdktf.ComplexList
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
// Experimental.
Fqn() *string
// The attribute on the parent resource this class is referencing.
TerraformAttribute() *string
SetTerraformAttribute(val *string)
// The parent resource.
TerraformResource() cdktf.IInterpolatingParent
SetTerraformResource(val cdktf.IInterpolatingParent)
// whether the list is wrapping a set (will add tolist() to be able to access an item via an index).
WrapsSet() *bool
SetWrapsSet(val *bool)
// Experimental.
ComputeFqn() *string
Get(index *float64) BigqueryJobStatusOutputReference
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobStatusList ¶
func NewBigqueryJobStatusList(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, wrapsSet *bool) BigqueryJobStatusList
type BigqueryJobStatusOutputReference ¶
type BigqueryJobStatusOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
ErrorResult() BigqueryJobStatusErrorResultList
Errors() BigqueryJobStatusErrorsList
// Experimental.
Fqn() *string
InternalValue() *BigqueryJobStatus
SetInternalValue(val *BigqueryJobStatus)
State() *string
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobStatusOutputReference ¶
func NewBigqueryJobStatusOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string, complexObjectIndex *float64, complexObjectIsFromSet *bool) BigqueryJobStatusOutputReference
type BigqueryJobTimeouts ¶
type BigqueryJobTimeouts struct {
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#create BigqueryJob#create}.
Create *string `field:"optional" json:"create" yaml:"create"`
// Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/5.3.0/docs/resources/bigquery_job#delete BigqueryJob#delete}.
Delete *string `field:"optional" json:"delete" yaml:"delete"`
}
type BigqueryJobTimeoutsOutputReference ¶
type BigqueryJobTimeoutsOutputReference interface {
cdktf.ComplexObject
// the index of the complex object in a list.
// Experimental.
ComplexObjectIndex() interface{}
// Experimental.
SetComplexObjectIndex(val interface{})
// set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items.
// Experimental.
ComplexObjectIsFromSet() *bool
// Experimental.
SetComplexObjectIsFromSet(val *bool)
Create() *string
SetCreate(val *string)
CreateInput() *string
// The creation stack of this resolvable which will be appended to errors thrown during resolution.
//
// If this returns an empty array the stack will not be attached.
// Experimental.
CreationStack() *[]*string
Delete() *string
SetDelete(val *string)
DeleteInput() *string
// Experimental.
Fqn() *string
InternalValue() interface{}
SetInternalValue(val interface{})
// Experimental.
TerraformAttribute() *string
// Experimental.
SetTerraformAttribute(val *string)
// Experimental.
TerraformResource() cdktf.IInterpolatingParent
// Experimental.
SetTerraformResource(val cdktf.IInterpolatingParent)
// Experimental.
ComputeFqn() *string
// Experimental.
GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{}
// Experimental.
GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable
// Experimental.
GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool
// Experimental.
GetListAttribute(terraformAttribute *string) *[]*string
// Experimental.
GetNumberAttribute(terraformAttribute *string) *float64
// Experimental.
GetNumberListAttribute(terraformAttribute *string) *[]*float64
// Experimental.
GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64
// Experimental.
GetStringAttribute(terraformAttribute *string) *string
// Experimental.
GetStringMapAttribute(terraformAttribute *string) *map[string]*string
// Experimental.
InterpolationAsList() cdktf.IResolvable
// Experimental.
InterpolationForAttribute(property *string) cdktf.IResolvable
ResetCreate()
ResetDelete()
// Produce the Token's value at resolution time.
// Experimental.
Resolve(_context cdktf.IResolveContext) interface{}
// Return a string representation of this resolvable object.
//
// Returns a reversible string representation.
// Experimental.
ToString() *string
}
func NewBigqueryJobTimeoutsOutputReference ¶
func NewBigqueryJobTimeoutsOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) BigqueryJobTimeoutsOutputReference
Source Files
¶
- BigqueryJob.go
- BigqueryJobConfig.go
- BigqueryJobCopy.go
- BigqueryJobCopyDestinationEncryptionConfiguration.go
- BigqueryJobCopyDestinationEncryptionConfigurationOutputReference.go
- BigqueryJobCopyDestinationEncryptionConfigurationOutputReference__checks.go
- BigqueryJobCopyDestinationTable.go
- BigqueryJobCopyDestinationTableOutputReference.go
- BigqueryJobCopyDestinationTableOutputReference__checks.go
- BigqueryJobCopyOutputReference.go
- BigqueryJobCopyOutputReference__checks.go
- BigqueryJobCopySourceTables.go
- BigqueryJobCopySourceTablesList.go
- BigqueryJobCopySourceTablesList__checks.go
- BigqueryJobCopySourceTablesOutputReference.go
- BigqueryJobCopySourceTablesOutputReference__checks.go
- BigqueryJobExtract.go
- BigqueryJobExtractOutputReference.go
- BigqueryJobExtractOutputReference__checks.go
- BigqueryJobExtractSourceModel.go
- BigqueryJobExtractSourceModelOutputReference.go
- BigqueryJobExtractSourceModelOutputReference__checks.go
- BigqueryJobExtractSourceTable.go
- BigqueryJobExtractSourceTableOutputReference.go
- BigqueryJobExtractSourceTableOutputReference__checks.go
- BigqueryJobLoad.go
- BigqueryJobLoadDestinationEncryptionConfiguration.go
- BigqueryJobLoadDestinationEncryptionConfigurationOutputReference.go
- BigqueryJobLoadDestinationEncryptionConfigurationOutputReference__checks.go
- BigqueryJobLoadDestinationTable.go
- BigqueryJobLoadDestinationTableOutputReference.go
- BigqueryJobLoadDestinationTableOutputReference__checks.go
- BigqueryJobLoadOutputReference.go
- BigqueryJobLoadOutputReference__checks.go
- BigqueryJobLoadParquetOptions.go
- BigqueryJobLoadParquetOptionsOutputReference.go
- BigqueryJobLoadParquetOptionsOutputReference__checks.go
- BigqueryJobLoadTimePartitioning.go
- BigqueryJobLoadTimePartitioningOutputReference.go
- BigqueryJobLoadTimePartitioningOutputReference__checks.go
- BigqueryJobQuery.go
- BigqueryJobQueryDefaultDataset.go
- BigqueryJobQueryDefaultDatasetOutputReference.go
- BigqueryJobQueryDefaultDatasetOutputReference__checks.go
- BigqueryJobQueryDestinationEncryptionConfiguration.go
- BigqueryJobQueryDestinationEncryptionConfigurationOutputReference.go
- BigqueryJobQueryDestinationEncryptionConfigurationOutputReference__checks.go
- BigqueryJobQueryDestinationTable.go
- BigqueryJobQueryDestinationTableOutputReference.go
- BigqueryJobQueryDestinationTableOutputReference__checks.go
- BigqueryJobQueryOutputReference.go
- BigqueryJobQueryOutputReference__checks.go
- BigqueryJobQueryScriptOptions.go
- BigqueryJobQueryScriptOptionsOutputReference.go
- BigqueryJobQueryScriptOptionsOutputReference__checks.go
- BigqueryJobQueryUserDefinedFunctionResources.go
- BigqueryJobQueryUserDefinedFunctionResourcesList.go
- BigqueryJobQueryUserDefinedFunctionResourcesList__checks.go
- BigqueryJobQueryUserDefinedFunctionResourcesOutputReference.go
- BigqueryJobQueryUserDefinedFunctionResourcesOutputReference__checks.go
- BigqueryJobStatus.go
- BigqueryJobStatusErrorResult.go
- BigqueryJobStatusErrorResultList.go
- BigqueryJobStatusErrorResultList__checks.go
- BigqueryJobStatusErrorResultOutputReference.go
- BigqueryJobStatusErrorResultOutputReference__checks.go
- BigqueryJobStatusErrors.go
- BigqueryJobStatusErrorsList.go
- BigqueryJobStatusErrorsList__checks.go
- BigqueryJobStatusErrorsOutputReference.go
- BigqueryJobStatusErrorsOutputReference__checks.go
- BigqueryJobStatusList.go
- BigqueryJobStatusList__checks.go
- BigqueryJobStatusOutputReference.go
- BigqueryJobStatusOutputReference__checks.go
- BigqueryJobTimeouts.go
- BigqueryJobTimeoutsOutputReference.go
- BigqueryJobTimeoutsOutputReference__checks.go
- BigqueryJob__checks.go
- main.go