From dcc47bea32125bf5e1f2c08f12638c5af0f500fb Mon Sep 17 00:00:00 2001 From: xiaoweim Date: Wed, 23 Oct 2024 22:05:58 +0000 Subject: [PATCH] feat: add direct controller for BigQueryDataset --- ...tasets.bigquery.cnrm.cloud.google.com.yaml | 16 +- .../bigquery/v1beta1/bigquerydataset_types.go | 42 ---- .../bigquery/v1beta1/zz_generated.deepcopy.go | 46 +--- .../direct/bigquery/v2/mapper.generated.go | 213 +----------------- .../v1beta1/bigqueryjob/dependencies.yaml | 5 - .../dependencies.yaml | 6 - .../dependencies.yaml | 2 - .../resource-docs/bigquery/bigquerydataset.md | 24 +- 8 files changed, 32 insertions(+), 322 deletions(-) diff --git a/config/crds/resources/apiextensions.k8s.io_v1_customresourcedefinition_bigquerydatasets.bigquery.cnrm.cloud.google.com.yaml b/config/crds/resources/apiextensions.k8s.io_v1_customresourcedefinition_bigquerydatasets.bigquery.cnrm.cloud.google.com.yaml index 8a8374041dc..e9d1e444b49 100644 --- a/config/crds/resources/apiextensions.k8s.io_v1_customresourcedefinition_bigquerydatasets.bigquery.cnrm.cloud.google.com.yaml +++ b/config/crds/resources/apiextensions.k8s.io_v1_customresourcedefinition_bigquerydatasets.bigquery.cnrm.cloud.google.com.yaml @@ -74,7 +74,7 @@ spec: its access needs to be granted again via an update operation.' properties: dataset: - description: The dataset this entry applies to + description: The dataset this entry applies to. properties: datasetId: description: A unique Id for this dataset, without the @@ -85,6 +85,9 @@ spec: projectId: description: The ID of the project containing this dataset. type: string + required: + - datasetId + - projectId type: object targetTypes: description: Which resources in the dataset this entry applies @@ -93,6 +96,9 @@ spec: items: type: string type: array + required: + - dataset + - targetTypes type: object domain: description: '[Pick one] A domain to grant access to. Any users @@ -143,6 +149,10 @@ spec: only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 256 characters. type: string + required: + - datasetId + - projectId + - routineId type: object specialGroup: description: |- @@ -183,6 +193,10 @@ spec: allow suffixing of the table Id with a partition decorator, such as `sample_table$20190123`. type: string + required: + - datasetId + - projectId + - tableId type: object type: object type: array diff --git a/pkg/clients/generated/apis/bigquery/v1beta1/bigquerydataset_types.go b/pkg/clients/generated/apis/bigquery/v1beta1/bigquerydataset_types.go index 222bb6cf695..3e13dba3647 100644 --- a/pkg/clients/generated/apis/bigquery/v1beta1/bigquerydataset_types.go +++ b/pkg/clients/generated/apis/bigquery/v1beta1/bigquerydataset_types.go @@ -91,21 +91,11 @@ type DatasetAccess struct { } type DatasetDataset struct { -<<<<<<< HEAD /* A unique Id for this dataset, without the project name. The Id must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters. */ DatasetId string `json:"datasetId"` /* The ID of the project containing this dataset. */ ProjectId string `json:"projectId"` -======= - /* Required. A unique Id for this dataset, without the project name. The Id must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters. */ - // +optional - DatasetId *string `json:"datasetId,omitempty"` - - /* Required. The Id of the project containing this dataset. */ - // +optional - ProjectId *string `json:"projectId,omitempty"` ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) } type DatasetDefaultEncryptionConfiguration struct { @@ -115,7 +105,6 @@ type DatasetDefaultEncryptionConfiguration struct { } type DatasetRoutine struct { -<<<<<<< HEAD /* The ID of the dataset containing this routine. */ DatasetId string `json:"datasetId"` @@ -135,33 +124,6 @@ type DatasetView struct { /* The Id of the table. The Id can contain Unicode characters in category L (letter), M (mark), N (number), Pc (connector, including underscore), Pd (dash), and Zs (space). For more information, see [General Category](https://wikipedia.org/wiki/Unicode_character_property#General_Category). The maximum length is 1,024 characters. Certain operations allow suffixing of the table Id with a partition decorator, such as `sample_table$20190123`. */ TableId string `json:"tableId"` -======= - /* Required. The Id of the dataset containing this routine. */ - // +optional - DatasetId *string `json:"datasetId,omitempty"` - - /* Required. The Id of the project containing this routine. */ - // +optional - ProjectId *string `json:"projectId,omitempty"` - - /* Required. The Id of the routine. The Id must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 256 characters. */ - // +optional - RoutineId *string `json:"routineId,omitempty"` -} - -type DatasetView struct { - /* Required. The Id of the dataset containing this table. */ - // +optional - DatasetId *string `json:"datasetId,omitempty"` - - /* Required. The Id of the project containing this table. */ - // +optional - ProjectId *string `json:"projectId,omitempty"` - - /* Required. The Id of the table. The Id can contain Unicode characters in category L (letter), M (mark), N (number), Pc (connector, including underscore), Pd (dash), and Zs (space). For more information, see [General Category](https://wikipedia.org/wiki/Unicode_character_property#General_Category). The maximum length is 1,024 characters. Certain operations allow suffixing of the table Id with a partition decorator, such as `sample_table$20190123`. */ - // +optional - TableId *string `json:"tableId,omitempty"` ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) } type BigQueryDatasetSpec struct { @@ -223,11 +185,7 @@ type BigQueryDatasetSpec struct { // +optional MaxTimeTravelHours *string `json:"maxTimeTravelHours,omitempty"` -<<<<<<< HEAD /* Optional. The project that this resource belongs to. */ -======= - /* The project that this resource belongs to. required */ ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) // +optional ProjectRef *v1alpha1.ResourceRef `json:"projectRef,omitempty"` diff --git a/pkg/clients/generated/apis/bigquery/v1beta1/zz_generated.deepcopy.go b/pkg/clients/generated/apis/bigquery/v1beta1/zz_generated.deepcopy.go index 63ef06146be..1bc18b90280 100644 --- a/pkg/clients/generated/apis/bigquery/v1beta1/zz_generated.deepcopy.go +++ b/pkg/clients/generated/apis/bigquery/v1beta1/zz_generated.deepcopy.go @@ -761,7 +761,7 @@ func (in *DatasetAccess) DeepCopyInto(out *DatasetAccess) { if in.Dataset != nil { in, out := &in.Dataset, &out.Dataset *out = new(DatasetDataset) - (*in).DeepCopyInto(*out) + **out = **in } if in.Domain != nil { in, out := &in.Domain, &out.Domain @@ -786,7 +786,7 @@ func (in *DatasetAccess) DeepCopyInto(out *DatasetAccess) { if in.Routine != nil { in, out := &in.Routine, &out.Routine *out = new(DatasetRoutine) - (*in).DeepCopyInto(*out) + **out = **in } if in.SpecialGroup != nil { in, out := &in.SpecialGroup, &out.SpecialGroup @@ -801,7 +801,7 @@ func (in *DatasetAccess) DeepCopyInto(out *DatasetAccess) { if in.View != nil { in, out := &in.View, &out.View *out = new(DatasetView) - (*in).DeepCopyInto(*out) + **out = **in } return } @@ -819,16 +819,6 @@ func (in *DatasetAccess) DeepCopy() *DatasetAccess { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DatasetDataset) DeepCopyInto(out *DatasetDataset) { *out = *in - if in.DatasetId != nil { - in, out := &in.DatasetId, &out.DatasetId - *out = new(string) - **out = **in - } - if in.ProjectId != nil { - in, out := &in.ProjectId, &out.ProjectId - *out = new(string) - **out = **in - } return } @@ -866,21 +856,6 @@ func (in *DatasetDefaultEncryptionConfiguration) DeepCopy() *DatasetDefaultEncry // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DatasetRoutine) DeepCopyInto(out *DatasetRoutine) { *out = *in - if in.DatasetId != nil { - in, out := &in.DatasetId, &out.DatasetId - *out = new(string) - **out = **in - } - if in.ProjectId != nil { - in, out := &in.ProjectId, &out.ProjectId - *out = new(string) - **out = **in - } - if in.RoutineId != nil { - in, out := &in.RoutineId, &out.RoutineId - *out = new(string) - **out = **in - } return } @@ -897,21 +872,6 @@ func (in *DatasetRoutine) DeepCopy() *DatasetRoutine { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DatasetView) DeepCopyInto(out *DatasetView) { *out = *in - if in.DatasetId != nil { - in, out := &in.DatasetId, &out.DatasetId - *out = new(string) - **out = **in - } - if in.ProjectId != nil { - in, out := &in.ProjectId, &out.ProjectId - *out = new(string) - **out = **in - } - if in.TableId != nil { - in, out := &in.TableId, &out.TableId - *out = new(string) - **out = **in - } return } diff --git a/pkg/controller/direct/bigquery/v2/mapper.generated.go b/pkg/controller/direct/bigquery/v2/mapper.generated.go index de64a9b0b92..f58fc9b0698 100644 --- a/pkg/controller/direct/bigquery/v2/mapper.generated.go +++ b/pkg/controller/direct/bigquery/v2/mapper.generated.go @@ -52,81 +52,7 @@ func Access_ToProto(mapCtx *direct.MapContext, in *krm.Access) *pb.DatasetAccess out.Dataset = DatasetAccessEntry_ToProto(mapCtx, in.Dataset) return out } -<<<<<<< HEAD -======= -func Dataset_FromProto(mapCtx *direct.MapContext, in *pb.Dataset) *krm.Dataset { - if in == nil { - return nil - } - out := &krm.Dataset{} - out.Kind = direct.LazyPtr(in.GetKind()) - out.Etag = direct.LazyPtr(in.GetEtag()) - out.ID = direct.LazyPtr(in.GetId()) - out.SelfLink = direct.LazyPtr(in.GetSelfLink()) - out.DatasetReference = DatasetReference_FromProto(mapCtx, in.GetDatasetReference()) - out.FriendlyName = direct.LazyPtr(in.GetFriendlyName()) - out.Description = direct.LazyPtr(in.GetDescription()) - out.DefaultTableExpirationMs = direct.LazyPtr(in.GetDefaultTableExpirationMs()) - out.DefaultPartitionExpirationMs = direct.LazyPtr(in.GetDefaultPartitionExpirationMs()) - out.Labels = in.Labels - out.Access = direct.Slice_FromProto(mapCtx, in.Access, Access_FromProto) - out.CreationTime = direct.LazyPtr(in.GetCreationTime()) - out.LastModifiedTime = direct.LazyPtr(in.GetLastModifiedTime()) - out.Location = direct.LazyPtr(in.GetLocation()) - out.DefaultEncryptionConfiguration = EncryptionConfiguration_FromProto(mapCtx, in.GetDefaultEncryptionConfiguration()) - out.SatisfiesPzs.Value = direct.LazyPtr(in.GetSatisfiesPzs()) - out.SatisfiesPzi.Value = direct.LazyPtr(in.GetSatisfiesPzi()) - out.Type = direct.LazyPtr(in.GetType()) - out.LinkedDatasetSource = LinkedDatasetSource_FromProto(mapCtx, in.GetLinkedDatasetSource()) - out.LinkedDatasetMetadata = LinkedDatasetMetadata_FromProto(mapCtx, in.GetLinkedDatasetMetadata()) - out.ExternalDatasetReference = ExternalDatasetReference_FromProto(mapCtx, in.GetExternalDatasetReference()) - out.ExternalCatalogDatasetOptions = ExternalCatalogDatasetOptions_FromProto(mapCtx, in.GetExternalCatalogDatasetOptions()) - out.IsCaseInsensitive.Value = direct.LazyPtr(in.GetIsCaseInsensitive()) - out.DefaultCollation = direct.LazyPtr(in.GetDefaultCollation()) - out.DefaultRoundingMode = direct.LazyPtr(in.GetDefaultRoundingMode()) - out.MaxTimeTravelHours = direct.LazyPtr(in.GetMaxTimeTravelHours()) - out.Tags = direct.Slice_FromProto(mapCtx, in.Tags, GcpTag_FromProto) - out.StorageBillingModel = direct.LazyPtr(in.GetStorageBillingModel()) - out.Restrictions = RestrictionConfig_FromProto(mapCtx, in.GetRestrictions()) - return out -} -func Dataset_ToProto(mapCtx *direct.MapContext, in *krm.Dataset) *pb.Dataset { - if in == nil { - return nil - } - out := &pb.Dataset{} - out.Kind = in.Kind - out.Etag = in.Etag - out.Id = in.ID - out.SelfLink = in.SelfLink - out.DatasetReference = DatasetReference_ToProto(mapCtx, in.DatasetReference) - out.FriendlyName = in.FriendlyName - out.Description = in.Description - out.DefaultTableExpirationMs = in.DefaultTableExpirationMs - out.DefaultPartitionExpirationMs = in.DefaultPartitionExpirationMs - out.Labels = in.Labels - out.Access = direct.Slice_ToProto(mapCtx, in.Access, Access_ToProto) - out.CreationTime = in.CreationTime - out.LastModifiedTime = in.LastModifiedTime - out.Location = in.Location - out.DefaultEncryptionConfiguration = EncryptionConfiguration_ToProto(mapCtx, in.DefaultEncryptionConfiguration) - out.SatisfiesPzs = in.SatisfiesPzs.Value - out.SatisfiesPzi = in.SatisfiesPzi.Value - out.Type = in.Type - out.LinkedDatasetSource = LinkedDatasetSource_ToProto(mapCtx, in.LinkedDatasetSource) - out.LinkedDatasetMetadata = LinkedDatasetMetadata_ToProto(mapCtx, in.LinkedDatasetMetadata) - out.ExternalDatasetReference = ExternalDatasetReference_ToProto(mapCtx, in.ExternalDatasetReference) - out.ExternalCatalogDatasetOptions = ExternalCatalogDatasetOptions_ToProto(mapCtx, in.ExternalCatalogDatasetOptions) - out.IsCaseInsensitive = in.IsCaseInsensitive.Value - out.DefaultCollation = in.DefaultCollation - out.DefaultRoundingMode = in.DefaultRoundingMode - out.MaxTimeTravelHours = in.MaxTimeTravelHours - out.Tags = direct.Slice_ToProto(mapCtx, in.Tags, GcpTag_ToProto) - out.StorageBillingModel = in.StorageBillingModel - out.Restrictions = RestrictionConfig_ToProto(mapCtx, in.Restrictions) - return out -} ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) + func DatasetAccessEntry_FromProto(mapCtx *direct.MapContext, in *pb.DatasetAccessEntry) *krm.DatasetAccessEntry { if in == nil { return nil @@ -179,142 +105,7 @@ func EncryptionConfiguration_ToProto(mapCtx *direct.MapContext, in *krm.Encrypti out.KmsKeyName = &in.KmsKeyRef.Name return out } -<<<<<<< HEAD -======= -func ErrorProto_FromProto(mapCtx *direct.MapContext, in *pb.ErrorProto) *krm.ErrorProto { - if in == nil { - return nil - } - out := &krm.ErrorProto{} - out.Reason = direct.LazyPtr(in.GetReason()) - out.Location = direct.LazyPtr(in.GetLocation()) - out.DebugInfo = direct.LazyPtr(in.GetDebugInfo()) - out.Message = direct.LazyPtr(in.GetMessage()) - return out -} -func ErrorProto_ToProto(mapCtx *direct.MapContext, in *krm.ErrorProto) *pb.ErrorProto { - if in == nil { - return nil - } - out := &pb.ErrorProto{} - out.Reason = in.Reason - out.Location = in.Location - out.DebugInfo = in.DebugInfo - out.Message = in.Message - return out -} -func ExplainQueryStage_FromProto(mapCtx *direct.MapContext, in *pb.ExplainQueryStage) *krm.ExplainQueryStage { - if in == nil { - return nil - } - out := &krm.ExplainQueryStage{} - out.Name = direct.LazyPtr(in.GetName()) - out.ID = direct.LazyPtr(in.GetId()) - out.StartMs = direct.LazyPtr(in.GetStartMs()) - out.EndMs = direct.LazyPtr(in.GetEndMs()) - out.InputStages = in.InputStages - out.WaitRatioAvg = direct.LazyPtr(in.GetWaitRatioAvg()) - out.WaitMsAvg = direct.LazyPtr(in.GetWaitMsAvg()) - out.WaitRatioMax = direct.LazyPtr(in.GetWaitRatioMax()) - out.WaitMsMax = direct.LazyPtr(in.GetWaitMsMax()) - out.ReadRatioAvg = direct.LazyPtr(in.GetReadRatioAvg()) - out.ReadMsAvg = direct.LazyPtr(in.GetReadMsAvg()) - out.ReadRatioMax = direct.LazyPtr(in.GetReadRatioMax()) - out.ReadMsMax = direct.LazyPtr(in.GetReadMsMax()) - out.ComputeRatioAvg = direct.LazyPtr(in.GetComputeRatioAvg()) - out.ComputeMsAvg = direct.LazyPtr(in.GetComputeMsAvg()) - out.ComputeRatioMax = direct.LazyPtr(in.GetComputeRatioMax()) - out.ComputeMsMax = direct.LazyPtr(in.GetComputeMsMax()) - out.WriteRatioAvg = direct.LazyPtr(in.GetWriteRatioAvg()) - out.WriteMsAvg = direct.LazyPtr(in.GetWriteMsAvg()) - out.WriteRatioMax = direct.LazyPtr(in.GetWriteRatioMax()) - out.WriteMsMax = direct.LazyPtr(in.GetWriteMsMax()) - out.ShuffleOutputBytes = direct.LazyPtr(in.GetShuffleOutputBytes()) - out.ShuffleOutputBytesSpilled = direct.LazyPtr(in.GetShuffleOutputBytesSpilled()) - out.RecordsRead = direct.LazyPtr(in.GetRecordsRead()) - out.RecordsWritten = direct.LazyPtr(in.GetRecordsWritten()) - out.ParallelInputs = direct.LazyPtr(in.GetParallelInputs()) - out.CompletedParallelInputs = direct.LazyPtr(in.GetCompletedParallelInputs()) - out.Status = direct.LazyPtr(in.GetStatus()) - out.Steps = direct.Slice_FromProto(mapCtx, in.Steps, ExplainQueryStep_FromProto) - out.SlotMs = direct.LazyPtr(in.GetSlotMs()) - out.ComputeMode = direct.LazyPtr(in.GetComputeMode()) - return out -} -func ExplainQueryStage_ToProto(mapCtx *direct.MapContext, in *krm.ExplainQueryStage) *pb.ExplainQueryStage { - if in == nil { - return nil - } - out := &pb.ExplainQueryStage{} - out.Name = in.Name - out.Id = in.ID - out.StartMs = in.StartMs - out.EndMs = in.EndMs - out.InputStages = in.InputStages - out.WaitRatioAvg = in.WaitRatioAvg - out.WaitMsAvg = in.WaitMsAvg - out.WaitRatioMax = in.WaitRatioMax - out.WaitMsMax = in.WaitMsMax - out.ReadRatioAvg = in.ReadRatioAvg - out.ReadMsAvg = in.ReadMsAvg - out.ReadRatioMax = in.ReadRatioMax - out.ReadMsMax = in.ReadMsMax - out.ComputeMsAvg = in.ComputeMsAvg - out.ComputeRatioMax = in.ComputeRatioMax - out.ComputeMsMax = in.ComputeMsMax - out.WriteRatioAvg = in.WriteRatioAvg - out.WriteMsAvg = in.WriteMsAvg - out.WriteRatioMax = in.WriteRatioMax - out.WriteMsMax = in.WriteMsMax - out.ShuffleOutputBytes = in.ShuffleOutputBytes - out.ShuffleOutputBytesSpilled = in.ShuffleOutputBytesSpilled - out.RecordsRead = in.RecordsRead - out.RecordsWritten = in.RecordsWritten - out.ParallelInputs = in.ParallelInputs - out.CompletedParallelInputs = in.CompletedParallelInputs - out.Status = in.Status - out.Steps = direct.Slice_ToProto(mapCtx, in.Steps, ExplainQueryStep_ToProto) - out.SlotMs = in.SlotMs - out.ComputeMode = in.ComputeMode - return out -} -func ExplainQueryStep_FromProto(mapCtx *direct.MapContext, in *pb.ExplainQueryStep) *krm.ExplainQueryStep { - if in == nil { - return nil - } - out := &krm.ExplainQueryStep{} - out.Kind = direct.LazyPtr(in.GetKind()) - out.Substeps = in.Substeps - return out -} -func ExplainQueryStep_ToProto(mapCtx *direct.MapContext, in *krm.ExplainQueryStep) *pb.ExplainQueryStep { - if in == nil { - return nil - } - out := &pb.ExplainQueryStep{} - out.Kind = in.Kind - out.Substeps = in.Substeps - return out -} -func ExportDataStatistics_FromProto(mapCtx *direct.MapContext, in *pb.ExportDataStatistics) *krm.ExportDataStatistics { - if in == nil { - return nil - } - out := &krm.ExportDataStatistics{} - out.FileCount = direct.LazyPtr(in.GetFileCount()) - out.RowCount = direct.LazyPtr(in.GetRowCount()) - return out -} -func ExportDataStatistics_ToProto(mapCtx *direct.MapContext, in *krm.ExportDataStatistics) *pb.ExportDataStatistics { - if in == nil { - return nil - } - out := &pb.ExportDataStatistics{} - out.FileCount = in.FileCount - out.RowCount = in.RowCount - return out -} ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) + func ExternalCatalogDatasetOptions_FromProto(mapCtx *direct.MapContext, in *pb.ExternalCatalogDatasetOptions) *krm.ExternalCatalogDatasetOptions { if in == nil { return nil diff --git a/pkg/test/resourcefixture/testdata/basic/bigquery/v1beta1/bigqueryjob/dependencies.yaml b/pkg/test/resourcefixture/testdata/basic/bigquery/v1beta1/bigqueryjob/dependencies.yaml index 0e0d535972e..072a5fee490 100644 --- a/pkg/test/resourcefixture/testdata/basic/bigquery/v1beta1/bigqueryjob/dependencies.yaml +++ b/pkg/test/resourcefixture/testdata/basic/bigquery/v1beta1/bigqueryjob/dependencies.yaml @@ -25,11 +25,6 @@ metadata: name: bigquerydataset2${uniqueId} spec: location: us-central1 -<<<<<<< HEAD -======= - projectRef: - external: ${projectId} ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) --- apiVersion: bigquery.cnrm.cloud.google.com/v1beta1 kind: BigQueryTable diff --git a/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-salesforce/dependencies.yaml b/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-salesforce/dependencies.yaml index f909288155a..f6108223d9f 100644 --- a/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-salesforce/dependencies.yaml +++ b/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-salesforce/dependencies.yaml @@ -17,10 +17,4 @@ kind: BigQueryDataset metadata: name: bigquerydataset${uniqueId} # "-" is not supported in dataset ID spec: -<<<<<<< HEAD location: us-central1 -======= - location: us-central1 - projectRef: - external: ${projectId} ->>>>>>> 86defe244 (feat: add direct controller for BigQueryDataset) diff --git a/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-scheduledquery/dependencies.yaml b/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-scheduledquery/dependencies.yaml index 0dd2492f034..8d886a5807a 100644 --- a/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-scheduledquery/dependencies.yaml +++ b/pkg/test/resourcefixture/testdata/basic/bigquerydatatransfer/v1beta1/bigquerydatatransferconfig/bigquerydatatransferconfig-scheduledquery/dependencies.yaml @@ -18,8 +18,6 @@ metadata: name: bigquerydataset${uniqueId} # "-" is not supported in dataset ID spec: location: us-central1 - projectRef: - external: ${projectId} --- apiVersion: iam.cnrm.cloud.google.com/v1beta1 kind: IAMServiceAccount diff --git a/scripts/generate-google3-docs/resource-reference/generated/resource-docs/bigquery/bigquerydataset.md b/scripts/generate-google3-docs/resource-reference/generated/resource-docs/bigquery/bigquerydataset.md index ae62050086e..9236b54a29f 100644 --- a/scripts/generate-google3-docs/resource-reference/generated/resource-docs/bigquery/bigquerydataset.md +++ b/scripts/generate-google3-docs/resource-reference/generated/resource-docs/bigquery/bigquerydataset.md @@ -175,17 +175,17 @@ storageBillingModel: string

access[].dataset.dataset

-

Optional

+

Required*

object

-

{% verbatim %}The dataset this entry applies to{% endverbatim %}

+

{% verbatim %}The dataset this entry applies to.{% endverbatim %}

access[].dataset.dataset.datasetId

-

Optional

+

Required*

string

@@ -195,7 +195,7 @@ storageBillingModel: string

access[].dataset.dataset.projectId

-

Optional

+

Required*

string

@@ -205,7 +205,7 @@ storageBillingModel: string

access[].dataset.targetTypes

-

Optional

+

Required*

list (string)

@@ -215,7 +215,7 @@ storageBillingModel: string

access[].dataset.targetTypes[]

-

Optional

+

Required*

string

@@ -285,7 +285,7 @@ storageBillingModel: string

access[].routine.datasetId

-

Optional

+

Required*

string

@@ -295,7 +295,7 @@ storageBillingModel: string

access[].routine.projectId

-

Optional

+

Required*

string

@@ -305,7 +305,7 @@ storageBillingModel: string

access[].routine.routineId

-

Optional

+

Required*

string

@@ -352,7 +352,7 @@ storageBillingModel: string

access[].view.datasetId

-

Optional

+

Required*

string

@@ -362,7 +362,7 @@ storageBillingModel: string

access[].view.projectId

-

Optional

+

Required*

string

@@ -372,7 +372,7 @@ storageBillingModel: string

access[].view.tableId

-

Optional

+

Required*

string