diff --git a/apis/amplify/v1beta1/zz_generated.deepcopy.go b/apis/amplify/v1beta1/zz_generated.deepcopy.go index 759b418c57..766d1aa8e5 100644 --- a/apis/amplify/v1beta1/zz_generated.deepcopy.go +++ b/apis/amplify/v1beta1/zz_generated.deepcopy.go @@ -76,6 +76,13 @@ func (in *AppInitParameters) DeepCopyInto(out *AppInitParameters) { *out = new(string) **out = **in } + if in.CacheConfig != nil { + in, out := &in.CacheConfig, &out.CacheConfig + *out = make([]CacheConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.CustomHeaders != nil { in, out := &in.CustomHeaders, &out.CustomHeaders *out = new(string) @@ -255,6 +262,13 @@ func (in *AppObservation) DeepCopyInto(out *AppObservation) { *out = new(string) **out = **in } + if in.CacheConfig != nil { + in, out := &in.CacheConfig, &out.CacheConfig + *out = make([]CacheConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.CustomHeaders != nil { in, out := &in.CustomHeaders, &out.CustomHeaders *out = new(string) @@ -425,6 +439,13 @@ func (in *AppParameters) DeepCopyInto(out *AppParameters) { *out = new(string) **out = **in } + if in.CacheConfig != nil { + in, out := &in.CacheConfig, &out.CacheConfig + *out = make([]CacheConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.CustomHeaders != nil { in, out := &in.CustomHeaders, &out.CustomHeaders *out = new(string) @@ -1516,6 +1537,66 @@ func (in *BranchStatus) DeepCopy() *BranchStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CacheConfigInitParameters) DeepCopyInto(out *CacheConfigInitParameters) { + *out = *in + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CacheConfigInitParameters. +func (in *CacheConfigInitParameters) DeepCopy() *CacheConfigInitParameters { + if in == nil { + return nil + } + out := new(CacheConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CacheConfigObservation) DeepCopyInto(out *CacheConfigObservation) { + *out = *in + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CacheConfigObservation. +func (in *CacheConfigObservation) DeepCopy() *CacheConfigObservation { + if in == nil { + return nil + } + out := new(CacheConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CacheConfigParameters) DeepCopyInto(out *CacheConfigParameters) { + *out = *in + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CacheConfigParameters. +func (in *CacheConfigParameters) DeepCopy() *CacheConfigParameters { + if in == nil { + return nil + } + out := new(CacheConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *CustomRuleInitParameters) DeepCopyInto(out *CustomRuleInitParameters) { *out = *in diff --git a/apis/appflow/v1beta1/zz_generated.deepcopy.go b/apis/appflow/v1beta1/zz_generated.deepcopy.go index 626aa19150..ed56f40182 100644 --- a/apis/appflow/v1beta1/zz_generated.deepcopy.go +++ b/apis/appflow/v1beta1/zz_generated.deepcopy.go @@ -1536,6 +1536,13 @@ func (in *FlowInitParameters) DeepCopyInto(out *FlowInitParameters) { *out = new(string) **out = **in } + if in.MetadataCatalogConfig != nil { + in, out := &in.MetadataCatalogConfig, &out.MetadataCatalogConfig + *out = make([]MetadataCatalogConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.SourceFlowConfig != nil { in, out := &in.SourceFlowConfig, &out.SourceFlowConfig *out = make([]SourceFlowConfigInitParameters, len(*in)) @@ -1652,6 +1659,13 @@ func (in *FlowObservation) DeepCopyInto(out *FlowObservation) { *out = new(string) **out = **in } + if in.MetadataCatalogConfig != nil { + in, out := &in.MetadataCatalogConfig, &out.MetadataCatalogConfig + *out = make([]MetadataCatalogConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.SourceFlowConfig != nil { in, out := &in.SourceFlowConfig, &out.SourceFlowConfig *out = make([]SourceFlowConfigObservation, len(*in)) @@ -1737,6 +1751,13 @@ func (in *FlowParameters) DeepCopyInto(out *FlowParameters) { *out = new(string) **out = **in } + if in.MetadataCatalogConfig != nil { + in, out := &in.MetadataCatalogConfig, &out.MetadataCatalogConfig + *out = make([]MetadataCatalogConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.Region != nil { in, out := &in.Region, &out.Region *out = new(string) @@ -1826,6 +1847,96 @@ func (in *FlowStatus) DeepCopy() *FlowStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *GlueDataCatalogInitParameters) DeepCopyInto(out *GlueDataCatalogInitParameters) { + *out = *in + if in.DatabaseName != nil { + in, out := &in.DatabaseName, &out.DatabaseName + *out = new(string) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.TablePrefix != nil { + in, out := &in.TablePrefix, &out.TablePrefix + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GlueDataCatalogInitParameters. +func (in *GlueDataCatalogInitParameters) DeepCopy() *GlueDataCatalogInitParameters { + if in == nil { + return nil + } + out := new(GlueDataCatalogInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *GlueDataCatalogObservation) DeepCopyInto(out *GlueDataCatalogObservation) { + *out = *in + if in.DatabaseName != nil { + in, out := &in.DatabaseName, &out.DatabaseName + *out = new(string) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.TablePrefix != nil { + in, out := &in.TablePrefix, &out.TablePrefix + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GlueDataCatalogObservation. +func (in *GlueDataCatalogObservation) DeepCopy() *GlueDataCatalogObservation { + if in == nil { + return nil + } + out := new(GlueDataCatalogObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *GlueDataCatalogParameters) DeepCopyInto(out *GlueDataCatalogParameters) { + *out = *in + if in.DatabaseName != nil { + in, out := &in.DatabaseName, &out.DatabaseName + *out = new(string) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.TablePrefix != nil { + in, out := &in.TablePrefix, &out.TablePrefix + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GlueDataCatalogParameters. +func (in *GlueDataCatalogParameters) DeepCopy() *GlueDataCatalogParameters { + if in == nil { + return nil + } + out := new(GlueDataCatalogParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *GoogleAnalyticsInitParameters) DeepCopyInto(out *GoogleAnalyticsInitParameters) { *out = *in @@ -2393,6 +2504,72 @@ func (in *MarketoParameters) DeepCopy() *MarketoParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MetadataCatalogConfigInitParameters) DeepCopyInto(out *MetadataCatalogConfigInitParameters) { + *out = *in + if in.GlueDataCatalog != nil { + in, out := &in.GlueDataCatalog, &out.GlueDataCatalog + *out = make([]GlueDataCatalogInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MetadataCatalogConfigInitParameters. +func (in *MetadataCatalogConfigInitParameters) DeepCopy() *MetadataCatalogConfigInitParameters { + if in == nil { + return nil + } + out := new(MetadataCatalogConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MetadataCatalogConfigObservation) DeepCopyInto(out *MetadataCatalogConfigObservation) { + *out = *in + if in.GlueDataCatalog != nil { + in, out := &in.GlueDataCatalog, &out.GlueDataCatalog + *out = make([]GlueDataCatalogObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MetadataCatalogConfigObservation. +func (in *MetadataCatalogConfigObservation) DeepCopy() *MetadataCatalogConfigObservation { + if in == nil { + return nil + } + out := new(MetadataCatalogConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MetadataCatalogConfigParameters) DeepCopyInto(out *MetadataCatalogConfigParameters) { + *out = *in + if in.GlueDataCatalog != nil { + in, out := &in.GlueDataCatalog, &out.GlueDataCatalog + *out = make([]GlueDataCatalogParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MetadataCatalogConfigParameters. +func (in *MetadataCatalogConfigParameters) DeepCopy() *MetadataCatalogConfigParameters { + if in == nil { + return nil + } + out := new(MetadataCatalogConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *PrefixConfigInitParameters) DeepCopyInto(out *PrefixConfigInitParameters) { *out = *in @@ -2401,6 +2578,17 @@ func (in *PrefixConfigInitParameters) DeepCopyInto(out *PrefixConfigInitParamete *out = new(string) **out = **in } + if in.PrefixHierarchy != nil { + in, out := &in.PrefixHierarchy, &out.PrefixHierarchy + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } if in.PrefixType != nil { in, out := &in.PrefixType, &out.PrefixType *out = new(string) @@ -2426,6 +2614,17 @@ func (in *PrefixConfigObservation) DeepCopyInto(out *PrefixConfigObservation) { *out = new(string) **out = **in } + if in.PrefixHierarchy != nil { + in, out := &in.PrefixHierarchy, &out.PrefixHierarchy + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } if in.PrefixType != nil { in, out := &in.PrefixType, &out.PrefixType *out = new(string) @@ -2451,6 +2650,17 @@ func (in *PrefixConfigParameters) DeepCopyInto(out *PrefixConfigParameters) { *out = new(string) **out = **in } + if in.PrefixHierarchy != nil { + in, out := &in.PrefixHierarchy, &out.PrefixHierarchy + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } if in.PrefixType != nil { in, out := &in.PrefixType, &out.PrefixType *out = new(string) @@ -2988,6 +3198,17 @@ func (in *S3OutputFormatConfigPrefixConfigInitParameters) DeepCopyInto(out *S3Ou *out = new(string) **out = **in } + if in.PrefixHierarchy != nil { + in, out := &in.PrefixHierarchy, &out.PrefixHierarchy + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } if in.PrefixType != nil { in, out := &in.PrefixType, &out.PrefixType *out = new(string) @@ -3013,6 +3234,17 @@ func (in *S3OutputFormatConfigPrefixConfigObservation) DeepCopyInto(out *S3Outpu *out = new(string) **out = **in } + if in.PrefixHierarchy != nil { + in, out := &in.PrefixHierarchy, &out.PrefixHierarchy + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } if in.PrefixType != nil { in, out := &in.PrefixType, &out.PrefixType *out = new(string) @@ -3038,6 +3270,17 @@ func (in *S3OutputFormatConfigPrefixConfigParameters) DeepCopyInto(out *S3Output *out = new(string) **out = **in } + if in.PrefixHierarchy != nil { + in, out := &in.PrefixHierarchy, &out.PrefixHierarchy + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } if in.PrefixType != nil { in, out := &in.PrefixType, &out.PrefixType *out = new(string) diff --git a/apis/appsync/v1beta1/zz_generated.deepcopy.go b/apis/appsync/v1beta1/zz_generated.deepcopy.go index 3f7ef23160..73f07a8f91 100644 --- a/apis/appsync/v1beta1/zz_generated.deepcopy.go +++ b/apis/appsync/v1beta1/zz_generated.deepcopy.go @@ -1535,6 +1535,96 @@ func (in *ElasticsearchConfigParameters) DeepCopy() *ElasticsearchConfigParamete return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EnhancedMetricsConfigInitParameters) DeepCopyInto(out *EnhancedMetricsConfigInitParameters) { + *out = *in + if in.DataSourceLevelMetricsBehavior != nil { + in, out := &in.DataSourceLevelMetricsBehavior, &out.DataSourceLevelMetricsBehavior + *out = new(string) + **out = **in + } + if in.OperationLevelMetricsConfig != nil { + in, out := &in.OperationLevelMetricsConfig, &out.OperationLevelMetricsConfig + *out = new(string) + **out = **in + } + if in.ResolverLevelMetricsBehavior != nil { + in, out := &in.ResolverLevelMetricsBehavior, &out.ResolverLevelMetricsBehavior + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EnhancedMetricsConfigInitParameters. +func (in *EnhancedMetricsConfigInitParameters) DeepCopy() *EnhancedMetricsConfigInitParameters { + if in == nil { + return nil + } + out := new(EnhancedMetricsConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EnhancedMetricsConfigObservation) DeepCopyInto(out *EnhancedMetricsConfigObservation) { + *out = *in + if in.DataSourceLevelMetricsBehavior != nil { + in, out := &in.DataSourceLevelMetricsBehavior, &out.DataSourceLevelMetricsBehavior + *out = new(string) + **out = **in + } + if in.OperationLevelMetricsConfig != nil { + in, out := &in.OperationLevelMetricsConfig, &out.OperationLevelMetricsConfig + *out = new(string) + **out = **in + } + if in.ResolverLevelMetricsBehavior != nil { + in, out := &in.ResolverLevelMetricsBehavior, &out.ResolverLevelMetricsBehavior + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EnhancedMetricsConfigObservation. +func (in *EnhancedMetricsConfigObservation) DeepCopy() *EnhancedMetricsConfigObservation { + if in == nil { + return nil + } + out := new(EnhancedMetricsConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EnhancedMetricsConfigParameters) DeepCopyInto(out *EnhancedMetricsConfigParameters) { + *out = *in + if in.DataSourceLevelMetricsBehavior != nil { + in, out := &in.DataSourceLevelMetricsBehavior, &out.DataSourceLevelMetricsBehavior + *out = new(string) + **out = **in + } + if in.OperationLevelMetricsConfig != nil { + in, out := &in.OperationLevelMetricsConfig, &out.OperationLevelMetricsConfig + *out = new(string) + **out = **in + } + if in.ResolverLevelMetricsBehavior != nil { + in, out := &in.ResolverLevelMetricsBehavior, &out.ResolverLevelMetricsBehavior + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EnhancedMetricsConfigParameters. +func (in *EnhancedMetricsConfigParameters) DeepCopy() *EnhancedMetricsConfigParameters { + if in == nil { + return nil + } + out := new(EnhancedMetricsConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *EventBridgeConfigInitParameters) DeepCopyInto(out *EventBridgeConfigInitParameters) { *out = *in @@ -2001,6 +2091,11 @@ func (in *GraphQLAPI) DeepCopyObject() runtime.Object { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *GraphQLAPIInitParameters) DeepCopyInto(out *GraphQLAPIInitParameters) { *out = *in + if in.APIType != nil { + in, out := &in.APIType, &out.APIType + *out = new(string) + **out = **in + } if in.AdditionalAuthenticationProvider != nil { in, out := &in.AdditionalAuthenticationProvider, &out.AdditionalAuthenticationProvider *out = make([]AdditionalAuthenticationProviderInitParameters, len(*in)) @@ -2013,6 +2108,13 @@ func (in *GraphQLAPIInitParameters) DeepCopyInto(out *GraphQLAPIInitParameters) *out = new(string) **out = **in } + if in.EnhancedMetricsConfig != nil { + in, out := &in.EnhancedMetricsConfig, &out.EnhancedMetricsConfig + *out = make([]EnhancedMetricsConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.IntrospectionConfig != nil { in, out := &in.IntrospectionConfig, &out.IntrospectionConfig *out = new(string) @@ -2032,6 +2134,21 @@ func (in *GraphQLAPIInitParameters) DeepCopyInto(out *GraphQLAPIInitParameters) (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.MergedAPIExecutionRoleArn != nil { + in, out := &in.MergedAPIExecutionRoleArn, &out.MergedAPIExecutionRoleArn + *out = new(string) + **out = **in + } + if in.MergedAPIExecutionRoleArnRef != nil { + in, out := &in.MergedAPIExecutionRoleArnRef, &out.MergedAPIExecutionRoleArnRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.MergedAPIExecutionRoleArnSelector != nil { + in, out := &in.MergedAPIExecutionRoleArnSelector, &out.MergedAPIExecutionRoleArnSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } if in.Name != nil { in, out := &in.Name, &out.Name *out = new(string) @@ -2229,6 +2346,11 @@ func (in *GraphQLAPIList) DeepCopyObject() runtime.Object { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *GraphQLAPIObservation) DeepCopyInto(out *GraphQLAPIObservation) { *out = *in + if in.APIType != nil { + in, out := &in.APIType, &out.APIType + *out = new(string) + **out = **in + } if in.AdditionalAuthenticationProvider != nil { in, out := &in.AdditionalAuthenticationProvider, &out.AdditionalAuthenticationProvider *out = make([]AdditionalAuthenticationProviderObservation, len(*in)) @@ -2246,6 +2368,13 @@ func (in *GraphQLAPIObservation) DeepCopyInto(out *GraphQLAPIObservation) { *out = new(string) **out = **in } + if in.EnhancedMetricsConfig != nil { + in, out := &in.EnhancedMetricsConfig, &out.EnhancedMetricsConfig + *out = make([]EnhancedMetricsConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ID != nil { in, out := &in.ID, &out.ID *out = new(string) @@ -2270,6 +2399,11 @@ func (in *GraphQLAPIObservation) DeepCopyInto(out *GraphQLAPIObservation) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.MergedAPIExecutionRoleArn != nil { + in, out := &in.MergedAPIExecutionRoleArn, &out.MergedAPIExecutionRoleArn + *out = new(string) + **out = **in + } if in.Name != nil { in, out := &in.Name, &out.Name *out = new(string) @@ -2482,6 +2616,11 @@ func (in *GraphQLAPIOpenIDConnectConfigParameters) DeepCopy() *GraphQLAPIOpenIDC // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *GraphQLAPIParameters) DeepCopyInto(out *GraphQLAPIParameters) { *out = *in + if in.APIType != nil { + in, out := &in.APIType, &out.APIType + *out = new(string) + **out = **in + } if in.AdditionalAuthenticationProvider != nil { in, out := &in.AdditionalAuthenticationProvider, &out.AdditionalAuthenticationProvider *out = make([]AdditionalAuthenticationProviderParameters, len(*in)) @@ -2494,6 +2633,13 @@ func (in *GraphQLAPIParameters) DeepCopyInto(out *GraphQLAPIParameters) { *out = new(string) **out = **in } + if in.EnhancedMetricsConfig != nil { + in, out := &in.EnhancedMetricsConfig, &out.EnhancedMetricsConfig + *out = make([]EnhancedMetricsConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.IntrospectionConfig != nil { in, out := &in.IntrospectionConfig, &out.IntrospectionConfig *out = new(string) @@ -2513,6 +2659,21 @@ func (in *GraphQLAPIParameters) DeepCopyInto(out *GraphQLAPIParameters) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.MergedAPIExecutionRoleArn != nil { + in, out := &in.MergedAPIExecutionRoleArn, &out.MergedAPIExecutionRoleArn + *out = new(string) + **out = **in + } + if in.MergedAPIExecutionRoleArnRef != nil { + in, out := &in.MergedAPIExecutionRoleArnRef, &out.MergedAPIExecutionRoleArnRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.MergedAPIExecutionRoleArnSelector != nil { + in, out := &in.MergedAPIExecutionRoleArnSelector, &out.MergedAPIExecutionRoleArnSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } if in.Name != nil { in, out := &in.Name, &out.Name *out = new(string) diff --git a/apis/appsync/v1beta1/zz_generated.resolvers.go b/apis/appsync/v1beta1/zz_generated.resolvers.go index 28b6b9f3c3..fd6ce94705 100644 --- a/apis/appsync/v1beta1/zz_generated.resolvers.go +++ b/apis/appsync/v1beta1/zz_generated.resolvers.go @@ -329,6 +329,25 @@ func (mg *GraphQLAPI) ResolveReferences(ctx context.Context, c client.Reader) er mg.Spec.ForProvider.LogConfig[i3].CloudwatchLogsRoleArnRef = rsp.ResolvedReference } + { + m, l, err = apisresolver.GetManagedResource("iam.aws.upbound.io", "v1beta1", "Role", "RoleList") + if err != nil { + return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") + } + rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ + CurrentValue: reference.FromPtrValue(mg.Spec.ForProvider.MergedAPIExecutionRoleArn), + Extract: common.ARNExtractor(), + Reference: mg.Spec.ForProvider.MergedAPIExecutionRoleArnRef, + Selector: mg.Spec.ForProvider.MergedAPIExecutionRoleArnSelector, + To: reference.To{List: l, Managed: m}, + }) + } + if err != nil { + return errors.Wrap(err, "mg.Spec.ForProvider.MergedAPIExecutionRoleArn") + } + mg.Spec.ForProvider.MergedAPIExecutionRoleArn = reference.ToPtrValue(rsp.ResolvedValue) + mg.Spec.ForProvider.MergedAPIExecutionRoleArnRef = rsp.ResolvedReference + for i3 := 0; i3 < len(mg.Spec.ForProvider.UserPoolConfig); i3++ { { m, l, err = apisresolver.GetManagedResource("cognitoidp.aws.upbound.io", "v1beta1", "UserPool", "UserPoolList") @@ -371,6 +390,25 @@ func (mg *GraphQLAPI) ResolveReferences(ctx context.Context, c client.Reader) er mg.Spec.InitProvider.LogConfig[i3].CloudwatchLogsRoleArnRef = rsp.ResolvedReference } + { + m, l, err = apisresolver.GetManagedResource("iam.aws.upbound.io", "v1beta1", "Role", "RoleList") + if err != nil { + return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") + } + rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ + CurrentValue: reference.FromPtrValue(mg.Spec.InitProvider.MergedAPIExecutionRoleArn), + Extract: common.ARNExtractor(), + Reference: mg.Spec.InitProvider.MergedAPIExecutionRoleArnRef, + Selector: mg.Spec.InitProvider.MergedAPIExecutionRoleArnSelector, + To: reference.To{List: l, Managed: m}, + }) + } + if err != nil { + return errors.Wrap(err, "mg.Spec.InitProvider.MergedAPIExecutionRoleArn") + } + mg.Spec.InitProvider.MergedAPIExecutionRoleArn = reference.ToPtrValue(rsp.ResolvedValue) + mg.Spec.InitProvider.MergedAPIExecutionRoleArnRef = rsp.ResolvedReference + for i3 := 0; i3 < len(mg.Spec.InitProvider.UserPoolConfig); i3++ { { m, l, err = apisresolver.GetManagedResource("cognitoidp.aws.upbound.io", "v1beta1", "UserPool", "UserPoolList") diff --git a/apis/batch/v1beta1/zz_generated.deepcopy.go b/apis/batch/v1beta1/zz_generated.deepcopy.go index b3aecbd950..27479bb0b3 100644 --- a/apis/batch/v1beta1/zz_generated.deepcopy.go +++ b/apis/batch/v1beta1/zz_generated.deepcopy.go @@ -729,6 +729,66 @@ func (in *HostPathParameters) DeepCopy() *HostPathParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePullSecretInitParameters) DeepCopyInto(out *ImagePullSecretInitParameters) { + *out = *in + if in.Name != nil { + in, out := &in.Name, &out.Name + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePullSecretInitParameters. +func (in *ImagePullSecretInitParameters) DeepCopy() *ImagePullSecretInitParameters { + if in == nil { + return nil + } + out := new(ImagePullSecretInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePullSecretObservation) DeepCopyInto(out *ImagePullSecretObservation) { + *out = *in + if in.Name != nil { + in, out := &in.Name, &out.Name + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePullSecretObservation. +func (in *ImagePullSecretObservation) DeepCopy() *ImagePullSecretObservation { + if in == nil { + return nil + } + out := new(ImagePullSecretObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePullSecretParameters) DeepCopyInto(out *ImagePullSecretParameters) { + *out = *in + if in.Name != nil { + in, out := &in.Name, &out.Name + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePullSecretParameters. +func (in *ImagePullSecretParameters) DeepCopy() *ImagePullSecretParameters { + if in == nil { + return nil + } + out := new(ImagePullSecretParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *JobDefinition) DeepCopyInto(out *JobDefinition) { *out = *in @@ -769,6 +829,11 @@ func (in *JobDefinitionInitParameters) DeepCopyInto(out *JobDefinitionInitParame *out = new(bool) **out = **in } + if in.EcsProperties != nil { + in, out := &in.EcsProperties, &out.EcsProperties + *out = new(string) + **out = **in + } if in.EksProperties != nil { in, out := &in.EksProperties, &out.EksProperties *out = make([]EksPropertiesInitParameters, len(*in)) @@ -925,6 +990,11 @@ func (in *JobDefinitionObservation) DeepCopyInto(out *JobDefinitionObservation) *out = new(bool) **out = **in } + if in.EcsProperties != nil { + in, out := &in.EcsProperties, &out.EcsProperties + *out = new(string) + **out = **in + } if in.EksProperties != nil { in, out := &in.EksProperties, &out.EksProperties *out = make([]EksPropertiesObservation, len(*in)) @@ -1065,6 +1135,11 @@ func (in *JobDefinitionParameters) DeepCopyInto(out *JobDefinitionParameters) { *out = new(bool) **out = **in } + if in.EcsProperties != nil { + in, out := &in.EcsProperties, &out.EcsProperties + *out = new(string) + **out = **in + } if in.EksProperties != nil { in, out := &in.EksProperties, &out.EksProperties *out = make([]EksPropertiesParameters, len(*in)) @@ -1319,6 +1394,13 @@ func (in *PodPropertiesInitParameters) DeepCopyInto(out *PodPropertiesInitParame *out = new(bool) **out = **in } + if in.ImagePullSecret != nil { + in, out := &in.ImagePullSecret, &out.ImagePullSecret + *out = make([]ImagePullSecretInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.Metadata != nil { in, out := &in.Metadata, &out.Metadata *out = make([]MetadataInitParameters, len(*in)) @@ -1370,6 +1452,13 @@ func (in *PodPropertiesObservation) DeepCopyInto(out *PodPropertiesObservation) *out = new(bool) **out = **in } + if in.ImagePullSecret != nil { + in, out := &in.ImagePullSecret, &out.ImagePullSecret + *out = make([]ImagePullSecretObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.Metadata != nil { in, out := &in.Metadata, &out.Metadata *out = make([]MetadataObservation, len(*in)) @@ -1421,6 +1510,13 @@ func (in *PodPropertiesParameters) DeepCopyInto(out *PodPropertiesParameters) { *out = new(bool) **out = **in } + if in.ImagePullSecret != nil { + in, out := &in.ImagePullSecret, &out.ImagePullSecret + *out = make([]ImagePullSecretParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.Metadata != nil { in, out := &in.Metadata, &out.Metadata *out = make([]MetadataParameters, len(*in)) diff --git a/apis/cloudformation/v1beta1/zz_generated.deepcopy.go b/apis/cloudformation/v1beta1/zz_generated.deepcopy.go index b15383eaa1..925fd61490 100644 --- a/apis/cloudformation/v1beta1/zz_generated.deepcopy.go +++ b/apis/cloudformation/v1beta1/zz_generated.deepcopy.go @@ -1335,6 +1335,11 @@ func (in *StackSetInstanceObservation) DeepCopy() *StackSetInstanceObservation { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *StackSetInstanceOperationPreferencesInitParameters) DeepCopyInto(out *StackSetInstanceOperationPreferencesInitParameters) { *out = *in + if in.ConcurrencyMode != nil { + in, out := &in.ConcurrencyMode, &out.ConcurrencyMode + *out = new(string) + **out = **in + } if in.FailureToleranceCount != nil { in, out := &in.FailureToleranceCount, &out.FailureToleranceCount *out = new(float64) @@ -1386,6 +1391,11 @@ func (in *StackSetInstanceOperationPreferencesInitParameters) DeepCopy() *StackS // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *StackSetInstanceOperationPreferencesObservation) DeepCopyInto(out *StackSetInstanceOperationPreferencesObservation) { *out = *in + if in.ConcurrencyMode != nil { + in, out := &in.ConcurrencyMode, &out.ConcurrencyMode + *out = new(string) + **out = **in + } if in.FailureToleranceCount != nil { in, out := &in.FailureToleranceCount, &out.FailureToleranceCount *out = new(float64) @@ -1437,6 +1447,11 @@ func (in *StackSetInstanceOperationPreferencesObservation) DeepCopy() *StackSetI // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *StackSetInstanceOperationPreferencesParameters) DeepCopyInto(out *StackSetInstanceOperationPreferencesParameters) { *out = *in + if in.ConcurrencyMode != nil { + in, out := &in.ConcurrencyMode, &out.ConcurrencyMode + *out = new(string) + **out = **in + } if in.FailureToleranceCount != nil { in, out := &in.FailureToleranceCount, &out.FailureToleranceCount *out = new(float64) diff --git a/apis/codepipeline/v1beta1/zz_generated.deepcopy.go b/apis/codepipeline/v1beta1/zz_generated.deepcopy.go index 1d127fdfaa..9ab8148274 100644 --- a/apis/codepipeline/v1beta1/zz_generated.deepcopy.go +++ b/apis/codepipeline/v1beta1/zz_generated.deepcopy.go @@ -89,6 +89,11 @@ func (in *ActionInitParameters) DeepCopyInto(out *ActionInitParameters) { *out = new(float64) **out = **in } + if in.TimeoutInMinutes != nil { + in, out := &in.TimeoutInMinutes, &out.TimeoutInMinutes + *out = new(float64) + **out = **in + } if in.Version != nil { in, out := &in.Version, &out.Version *out = new(string) @@ -187,6 +192,11 @@ func (in *ActionObservation) DeepCopyInto(out *ActionObservation) { *out = new(float64) **out = **in } + if in.TimeoutInMinutes != nil { + in, out := &in.TimeoutInMinutes, &out.TimeoutInMinutes + *out = new(float64) + **out = **in + } if in.Version != nil { in, out := &in.Version, &out.Version *out = new(string) @@ -285,6 +295,11 @@ func (in *ActionParameters) DeepCopyInto(out *ActionParameters) { *out = new(float64) **out = **in } + if in.TimeoutInMinutes != nil { + in, out := &in.TimeoutInMinutes, &out.TimeoutInMinutes + *out = new(float64) + **out = **in + } if in.Version != nil { in, out := &in.Version, &out.Version *out = new(string) diff --git a/apis/cognitoidp/v1beta1/zz_generated.deepcopy.go b/apis/cognitoidp/v1beta1/zz_generated.deepcopy.go index fa3bc398fc..d2cb0bd382 100644 --- a/apis/cognitoidp/v1beta1/zz_generated.deepcopy.go +++ b/apis/cognitoidp/v1beta1/zz_generated.deepcopy.go @@ -2716,6 +2716,11 @@ func (in *PasswordPolicyInitParameters) DeepCopyInto(out *PasswordPolicyInitPara *out = new(float64) **out = **in } + if in.PasswordHistorySize != nil { + in, out := &in.PasswordHistorySize, &out.PasswordHistorySize + *out = new(float64) + **out = **in + } if in.RequireLowercase != nil { in, out := &in.RequireLowercase, &out.RequireLowercase *out = new(bool) @@ -2761,6 +2766,11 @@ func (in *PasswordPolicyObservation) DeepCopyInto(out *PasswordPolicyObservation *out = new(float64) **out = **in } + if in.PasswordHistorySize != nil { + in, out := &in.PasswordHistorySize, &out.PasswordHistorySize + *out = new(float64) + **out = **in + } if in.RequireLowercase != nil { in, out := &in.RequireLowercase, &out.RequireLowercase *out = new(bool) @@ -2806,6 +2816,11 @@ func (in *PasswordPolicyParameters) DeepCopyInto(out *PasswordPolicyParameters) *out = new(float64) **out = **in } + if in.PasswordHistorySize != nil { + in, out := &in.PasswordHistorySize, &out.PasswordHistorySize + *out = new(float64) + **out = **in + } if in.RequireLowercase != nil { in, out := &in.RequireLowercase, &out.RequireLowercase *out = new(bool) diff --git a/apis/dynamodb/v1beta1/zz_generated.deepcopy.go b/apis/dynamodb/v1beta1/zz_generated.deepcopy.go index 40227ccea1..c2e5bf7d00 100644 --- a/apis/dynamodb/v1beta1/zz_generated.deepcopy.go +++ b/apis/dynamodb/v1beta1/zz_generated.deepcopy.go @@ -1919,6 +1919,11 @@ func (in *TableInitParameters) DeepCopyInto(out *TableInitParameters) { *out = new(string) **out = **in } + if in.RestoreSourceTableArn != nil { + in, out := &in.RestoreSourceTableArn, &out.RestoreSourceTableArn + *out = new(string) + **out = **in + } if in.RestoreToLatestTime != nil { in, out := &in.RestoreToLatestTime, &out.RestoreToLatestTime *out = new(bool) @@ -2337,6 +2342,11 @@ func (in *TableObservation) DeepCopyInto(out *TableObservation) { *out = new(string) **out = **in } + if in.RestoreSourceTableArn != nil { + in, out := &in.RestoreSourceTableArn, &out.RestoreSourceTableArn + *out = new(string) + **out = **in + } if in.RestoreToLatestTime != nil { in, out := &in.RestoreToLatestTime, &out.RestoreToLatestTime *out = new(bool) @@ -2515,6 +2525,11 @@ func (in *TableParameters) DeepCopyInto(out *TableParameters) { *out = new(string) **out = **in } + if in.RestoreSourceTableArn != nil { + in, out := &in.RestoreSourceTableArn, &out.RestoreSourceTableArn + *out = new(string) + **out = **in + } if in.RestoreToLatestTime != nil { in, out := &in.RestoreToLatestTime, &out.RestoreToLatestTime *out = new(bool) diff --git a/apis/ecs/v1beta1/zz_generated.deepcopy.go b/apis/ecs/v1beta1/zz_generated.deepcopy.go index 1a69262d6c..ee017c95f3 100644 --- a/apis/ecs/v1beta1/zz_generated.deepcopy.go +++ b/apis/ecs/v1beta1/zz_generated.deepcopy.go @@ -1444,6 +1444,13 @@ func (in *ConfigurationInitParameters) DeepCopyInto(out *ConfigurationInitParame (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.ManagedStorageConfiguration != nil { + in, out := &in.ManagedStorageConfiguration, &out.ManagedStorageConfiguration + *out = make([]ManagedStorageConfigurationInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigurationInitParameters. @@ -1466,6 +1473,13 @@ func (in *ConfigurationObservation) DeepCopyInto(out *ConfigurationObservation) (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.ManagedStorageConfiguration != nil { + in, out := &in.ManagedStorageConfiguration, &out.ManagedStorageConfiguration + *out = make([]ManagedStorageConfigurationObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigurationObservation. @@ -1488,6 +1502,13 @@ func (in *ConfigurationParameters) DeepCopyInto(out *ConfigurationParameters) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.ManagedStorageConfiguration != nil { + in, out := &in.ManagedStorageConfiguration, &out.ManagedStorageConfiguration + *out = make([]ManagedStorageConfigurationParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigurationParameters. @@ -3044,6 +3065,81 @@ func (in *ManagedScalingParameters) DeepCopy() *ManagedScalingParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ManagedStorageConfigurationInitParameters) DeepCopyInto(out *ManagedStorageConfigurationInitParameters) { + *out = *in + if in.FargateEphemeralStorageKMSKeyID != nil { + in, out := &in.FargateEphemeralStorageKMSKeyID, &out.FargateEphemeralStorageKMSKeyID + *out = new(string) + **out = **in + } + if in.KMSKeyID != nil { + in, out := &in.KMSKeyID, &out.KMSKeyID + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ManagedStorageConfigurationInitParameters. +func (in *ManagedStorageConfigurationInitParameters) DeepCopy() *ManagedStorageConfigurationInitParameters { + if in == nil { + return nil + } + out := new(ManagedStorageConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ManagedStorageConfigurationObservation) DeepCopyInto(out *ManagedStorageConfigurationObservation) { + *out = *in + if in.FargateEphemeralStorageKMSKeyID != nil { + in, out := &in.FargateEphemeralStorageKMSKeyID, &out.FargateEphemeralStorageKMSKeyID + *out = new(string) + **out = **in + } + if in.KMSKeyID != nil { + in, out := &in.KMSKeyID, &out.KMSKeyID + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ManagedStorageConfigurationObservation. +func (in *ManagedStorageConfigurationObservation) DeepCopy() *ManagedStorageConfigurationObservation { + if in == nil { + return nil + } + out := new(ManagedStorageConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ManagedStorageConfigurationParameters) DeepCopyInto(out *ManagedStorageConfigurationParameters) { + *out = *in + if in.FargateEphemeralStorageKMSKeyID != nil { + in, out := &in.FargateEphemeralStorageKMSKeyID, &out.FargateEphemeralStorageKMSKeyID + *out = new(string) + **out = **in + } + if in.KMSKeyID != nil { + in, out := &in.KMSKeyID, &out.KMSKeyID + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ManagedStorageConfigurationParameters. +func (in *ManagedStorageConfigurationParameters) DeepCopy() *ManagedStorageConfigurationParameters { + if in == nil { + return nil + } + out := new(ManagedStorageConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *NetworkConfigurationInitParameters) DeepCopyInto(out *NetworkConfigurationInitParameters) { *out = *in @@ -4198,6 +4294,11 @@ func (in *ServiceInitParameters) DeepCopyInto(out *ServiceInitParameters) { *out = new(bool) **out = **in } + if in.ForceDelete != nil { + in, out := &in.ForceDelete, &out.ForceDelete + *out = new(bool) + **out = **in + } if in.ForceNewDeployment != nil { in, out := &in.ForceNewDeployment, &out.ForceNewDeployment *out = new(bool) @@ -4447,6 +4548,11 @@ func (in *ServiceObservation) DeepCopyInto(out *ServiceObservation) { *out = new(bool) **out = **in } + if in.ForceDelete != nil { + in, out := &in.ForceDelete, &out.ForceDelete + *out = new(bool) + **out = **in + } if in.ForceNewDeployment != nil { in, out := &in.ForceNewDeployment, &out.ForceNewDeployment *out = new(bool) @@ -4675,6 +4781,11 @@ func (in *ServiceParameters) DeepCopyInto(out *ServiceParameters) { *out = new(bool) **out = **in } + if in.ForceDelete != nil { + in, out := &in.ForceDelete, &out.ForceDelete + *out = new(bool) + **out = **in + } if in.ForceNewDeployment != nil { in, out := &in.ForceNewDeployment, &out.ForceNewDeployment *out = new(bool) diff --git a/apis/eks/v1beta1/zz_generated.deepcopy.go b/apis/eks/v1beta1/zz_generated.deepcopy.go index c2060c8d08..252d76fd9a 100644 --- a/apis/eks/v1beta1/zz_generated.deepcopy.go +++ b/apis/eks/v1beta1/zz_generated.deepcopy.go @@ -1487,6 +1487,13 @@ func (in *ClusterInitParameters) DeepCopyInto(out *ClusterInitParameters) { (*out)[key] = outVal } } + if in.UpgradePolicy != nil { + in, out := &in.UpgradePolicy, &out.UpgradePolicy + *out = make([]UpgradePolicyInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.VPCConfig != nil { in, out := &in.VPCConfig, &out.VPCConfig *out = make([]VPCConfigInitParameters, len(*in)) @@ -1676,6 +1683,13 @@ func (in *ClusterObservation) DeepCopyInto(out *ClusterObservation) { (*out)[key] = outVal } } + if in.UpgradePolicy != nil { + in, out := &in.UpgradePolicy, &out.UpgradePolicy + *out = make([]UpgradePolicyObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.VPCConfig != nil { in, out := &in.VPCConfig, &out.VPCConfig *out = make([]VPCConfigObservation, len(*in)) @@ -1783,6 +1797,13 @@ func (in *ClusterParameters) DeepCopyInto(out *ClusterParameters) { (*out)[key] = outVal } } + if in.UpgradePolicy != nil { + in, out := &in.UpgradePolicy, &out.UpgradePolicy + *out = make([]UpgradePolicyParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.VPCConfig != nil { in, out := &in.VPCConfig, &out.VPCConfig *out = make([]VPCConfigParameters, len(*in)) @@ -4778,6 +4799,66 @@ func (in *UpdateConfigParameters) DeepCopy() *UpdateConfigParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UpgradePolicyInitParameters) DeepCopyInto(out *UpgradePolicyInitParameters) { + *out = *in + if in.SupportType != nil { + in, out := &in.SupportType, &out.SupportType + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UpgradePolicyInitParameters. +func (in *UpgradePolicyInitParameters) DeepCopy() *UpgradePolicyInitParameters { + if in == nil { + return nil + } + out := new(UpgradePolicyInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UpgradePolicyObservation) DeepCopyInto(out *UpgradePolicyObservation) { + *out = *in + if in.SupportType != nil { + in, out := &in.SupportType, &out.SupportType + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UpgradePolicyObservation. +func (in *UpgradePolicyObservation) DeepCopy() *UpgradePolicyObservation { + if in == nil { + return nil + } + out := new(UpgradePolicyObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UpgradePolicyParameters) DeepCopyInto(out *UpgradePolicyParameters) { + *out = *in + if in.SupportType != nil { + in, out := &in.SupportType, &out.SupportType + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UpgradePolicyParameters. +func (in *UpgradePolicyParameters) DeepCopy() *UpgradePolicyParameters { + if in == nil { + return nil + } + out := new(UpgradePolicyParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *VPCConfigInitParameters) DeepCopyInto(out *VPCConfigInitParameters) { *out = *in diff --git a/apis/elasticache/v1beta1/zz_generated.deepcopy.go b/apis/elasticache/v1beta1/zz_generated.deepcopy.go index 7225399545..2d94b88f84 100644 --- a/apis/elasticache/v1beta1/zz_generated.deepcopy.go +++ b/apis/elasticache/v1beta1/zz_generated.deepcopy.go @@ -2289,6 +2289,16 @@ func (in *ReplicationGroupInitParameters) DeepCopyInto(out *ReplicationGroupInit *out = new(bool) **out = **in } + if in.AuthTokenSecretRef != nil { + in, out := &in.AuthTokenSecretRef, &out.AuthTokenSecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } + if in.AuthTokenUpdateStrategy != nil { + in, out := &in.AuthTokenUpdateStrategy, &out.AuthTokenUpdateStrategy + *out = new(string) + **out = **in + } if in.AutoMinorVersionUpgrade != nil { in, out := &in.AutoMinorVersionUpgrade, &out.AutoMinorVersionUpgrade *out = new(string) @@ -2347,6 +2357,21 @@ func (in *ReplicationGroupInitParameters) DeepCopyInto(out *ReplicationGroupInit *out = new(string) **out = **in } + if in.GlobalReplicationGroupIDRef != nil { + in, out := &in.GlobalReplicationGroupIDRef, &out.GlobalReplicationGroupIDRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.GlobalReplicationGroupIDSelector != nil { + in, out := &in.GlobalReplicationGroupIDSelector, &out.GlobalReplicationGroupIDSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } + if in.IPDiscovery != nil { + in, out := &in.IPDiscovery, &out.IPDiscovery + *out = new(string) + **out = **in + } if in.KMSKeyID != nil { in, out := &in.KMSKeyID, &out.KMSKeyID *out = new(string) @@ -2379,6 +2404,11 @@ func (in *ReplicationGroupInitParameters) DeepCopyInto(out *ReplicationGroupInit *out = new(bool) **out = **in } + if in.NetworkType != nil { + in, out := &in.NetworkType, &out.NetworkType + *out = new(string) + **out = **in + } if in.NodeType != nil { in, out := &in.NodeType, &out.NodeType *out = new(string) @@ -2531,6 +2561,11 @@ func (in *ReplicationGroupInitParameters) DeepCopyInto(out *ReplicationGroupInit *out = new(bool) **out = **in } + if in.TransitEncryptionMode != nil { + in, out := &in.TransitEncryptionMode, &out.TransitEncryptionMode + *out = new(string) + **out = **in + } if in.UserGroupIds != nil { in, out := &in.UserGroupIds, &out.UserGroupIds *out = make([]*string, len(*in)) @@ -2709,6 +2744,11 @@ func (in *ReplicationGroupObservation) DeepCopyInto(out *ReplicationGroupObserva *out = new(bool) **out = **in } + if in.AuthTokenUpdateStrategy != nil { + in, out := &in.AuthTokenUpdateStrategy, &out.AuthTokenUpdateStrategy + *out = new(string) + **out = **in + } if in.AutoMinorVersionUpgrade != nil { in, out := &in.AutoMinorVersionUpgrade, &out.AutoMinorVersionUpgrade *out = new(string) @@ -2787,6 +2827,11 @@ func (in *ReplicationGroupObservation) DeepCopyInto(out *ReplicationGroupObserva *out = new(string) **out = **in } + if in.IPDiscovery != nil { + in, out := &in.IPDiscovery, &out.IPDiscovery + *out = new(string) + **out = **in + } if in.KMSKeyID != nil { in, out := &in.KMSKeyID, &out.KMSKeyID *out = new(string) @@ -2820,6 +2865,11 @@ func (in *ReplicationGroupObservation) DeepCopyInto(out *ReplicationGroupObserva *out = new(bool) **out = **in } + if in.NetworkType != nil { + in, out := &in.NetworkType, &out.NetworkType + *out = new(string) + **out = **in + } if in.NodeType != nil { in, out := &in.NodeType, &out.NodeType *out = new(string) @@ -2976,6 +3026,11 @@ func (in *ReplicationGroupObservation) DeepCopyInto(out *ReplicationGroupObserva *out = new(bool) **out = **in } + if in.TransitEncryptionMode != nil { + in, out := &in.TransitEncryptionMode, &out.TransitEncryptionMode + *out = new(string) + **out = **in + } if in.UserGroupIds != nil { in, out := &in.UserGroupIds, &out.UserGroupIds *out = make([]*string, len(*in)) @@ -3017,6 +3072,16 @@ func (in *ReplicationGroupParameters) DeepCopyInto(out *ReplicationGroupParamete *out = new(v1.SecretKeySelector) **out = **in } + if in.AuthTokenUpdateStrategy != nil { + in, out := &in.AuthTokenUpdateStrategy, &out.AuthTokenUpdateStrategy + *out = new(string) + **out = **in + } + if in.AutoGenerateAuthToken != nil { + in, out := &in.AutoGenerateAuthToken, &out.AutoGenerateAuthToken + *out = new(bool) + **out = **in + } if in.AutoMinorVersionUpgrade != nil { in, out := &in.AutoMinorVersionUpgrade, &out.AutoMinorVersionUpgrade *out = new(string) @@ -3075,6 +3140,21 @@ func (in *ReplicationGroupParameters) DeepCopyInto(out *ReplicationGroupParamete *out = new(string) **out = **in } + if in.GlobalReplicationGroupIDRef != nil { + in, out := &in.GlobalReplicationGroupIDRef, &out.GlobalReplicationGroupIDRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.GlobalReplicationGroupIDSelector != nil { + in, out := &in.GlobalReplicationGroupIDSelector, &out.GlobalReplicationGroupIDSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } + if in.IPDiscovery != nil { + in, out := &in.IPDiscovery, &out.IPDiscovery + *out = new(string) + **out = **in + } if in.KMSKeyID != nil { in, out := &in.KMSKeyID, &out.KMSKeyID *out = new(string) @@ -3107,6 +3187,11 @@ func (in *ReplicationGroupParameters) DeepCopyInto(out *ReplicationGroupParamete *out = new(bool) **out = **in } + if in.NetworkType != nil { + in, out := &in.NetworkType, &out.NetworkType + *out = new(string) + **out = **in + } if in.NodeType != nil { in, out := &in.NodeType, &out.NodeType *out = new(string) @@ -3264,6 +3349,11 @@ func (in *ReplicationGroupParameters) DeepCopyInto(out *ReplicationGroupParamete *out = new(bool) **out = **in } + if in.TransitEncryptionMode != nil { + in, out := &in.TransitEncryptionMode, &out.TransitEncryptionMode + *out = new(string) + **out = **in + } if in.UserGroupIds != nil { in, out := &in.UserGroupIds, &out.UserGroupIds *out = make([]*string, len(*in)) diff --git a/apis/elbv2/v1beta1/zz_generated.deepcopy.go b/apis/elbv2/v1beta1/zz_generated.deepcopy.go index 9137bef409..49b5a7937c 100644 --- a/apis/elbv2/v1beta1/zz_generated.deepcopy.go +++ b/apis/elbv2/v1beta1/zz_generated.deepcopy.go @@ -1827,6 +1827,81 @@ func (in *ConnectionLogsParameters) DeepCopy() *ConnectionLogsParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNSFailoverInitParameters) DeepCopyInto(out *DNSFailoverInitParameters) { + *out = *in + if in.MinimumHealthyTargetsCount != nil { + in, out := &in.MinimumHealthyTargetsCount, &out.MinimumHealthyTargetsCount + *out = new(string) + **out = **in + } + if in.MinimumHealthyTargetsPercentage != nil { + in, out := &in.MinimumHealthyTargetsPercentage, &out.MinimumHealthyTargetsPercentage + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNSFailoverInitParameters. +func (in *DNSFailoverInitParameters) DeepCopy() *DNSFailoverInitParameters { + if in == nil { + return nil + } + out := new(DNSFailoverInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNSFailoverObservation) DeepCopyInto(out *DNSFailoverObservation) { + *out = *in + if in.MinimumHealthyTargetsCount != nil { + in, out := &in.MinimumHealthyTargetsCount, &out.MinimumHealthyTargetsCount + *out = new(string) + **out = **in + } + if in.MinimumHealthyTargetsPercentage != nil { + in, out := &in.MinimumHealthyTargetsPercentage, &out.MinimumHealthyTargetsPercentage + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNSFailoverObservation. +func (in *DNSFailoverObservation) DeepCopy() *DNSFailoverObservation { + if in == nil { + return nil + } + out := new(DNSFailoverObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DNSFailoverParameters) DeepCopyInto(out *DNSFailoverParameters) { + *out = *in + if in.MinimumHealthyTargetsCount != nil { + in, out := &in.MinimumHealthyTargetsCount, &out.MinimumHealthyTargetsCount + *out = new(string) + **out = **in + } + if in.MinimumHealthyTargetsPercentage != nil { + in, out := &in.MinimumHealthyTargetsPercentage, &out.MinimumHealthyTargetsPercentage + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DNSFailoverParameters. +func (in *DNSFailoverParameters) DeepCopy() *DNSFailoverParameters { + if in == nil { + return nil + } + out := new(DNSFailoverParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DefaultActionInitParameters) DeepCopyInto(out *DefaultActionInitParameters) { *out = *in @@ -4783,6 +4858,11 @@ func (in *LBTargetGroupInitParameters) DeepCopyInto(out *LBTargetGroupInitParame (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.TargetGroupHealth != nil { + in, out := &in.TargetGroupHealth, &out.TargetGroupHealth + *out = new(TargetGroupHealthInitParameters) + (*in).DeepCopyInto(*out) + } if in.TargetHealthState != nil { in, out := &in.TargetHealthState, &out.TargetHealthState *out = make([]TargetHealthStateInitParameters, len(*in)) @@ -5006,6 +5086,11 @@ func (in *LBTargetGroupObservation) DeepCopyInto(out *LBTargetGroupObservation) (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.TargetGroupHealth != nil { + in, out := &in.TargetGroupHealth, &out.TargetGroupHealth + *out = new(TargetGroupHealthObservation) + (*in).DeepCopyInto(*out) + } if in.TargetHealthState != nil { in, out := &in.TargetHealthState, &out.TargetHealthState *out = make([]TargetHealthStateObservation, len(*in)) @@ -5150,6 +5235,11 @@ func (in *LBTargetGroupParameters) DeepCopyInto(out *LBTargetGroupParameters) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.TargetGroupHealth != nil { + in, out := &in.TargetGroupHealth, &out.TargetGroupHealth + *out = new(TargetGroupHealthParameters) + (*in).DeepCopyInto(*out) + } if in.TargetHealthState != nil { in, out := &in.TargetHealthState, &out.TargetHealthState *out = make([]TargetHealthStateParameters, len(*in)) @@ -6348,6 +6438,81 @@ func (in *TargetFailoverParameters) DeepCopy() *TargetFailoverParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *TargetGroupHealthInitParameters) DeepCopyInto(out *TargetGroupHealthInitParameters) { + *out = *in + if in.DNSFailover != nil { + in, out := &in.DNSFailover, &out.DNSFailover + *out = new(DNSFailoverInitParameters) + (*in).DeepCopyInto(*out) + } + if in.UnhealthyStateRouting != nil { + in, out := &in.UnhealthyStateRouting, &out.UnhealthyStateRouting + *out = new(UnhealthyStateRoutingInitParameters) + (*in).DeepCopyInto(*out) + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TargetGroupHealthInitParameters. +func (in *TargetGroupHealthInitParameters) DeepCopy() *TargetGroupHealthInitParameters { + if in == nil { + return nil + } + out := new(TargetGroupHealthInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *TargetGroupHealthObservation) DeepCopyInto(out *TargetGroupHealthObservation) { + *out = *in + if in.DNSFailover != nil { + in, out := &in.DNSFailover, &out.DNSFailover + *out = new(DNSFailoverObservation) + (*in).DeepCopyInto(*out) + } + if in.UnhealthyStateRouting != nil { + in, out := &in.UnhealthyStateRouting, &out.UnhealthyStateRouting + *out = new(UnhealthyStateRoutingObservation) + (*in).DeepCopyInto(*out) + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TargetGroupHealthObservation. +func (in *TargetGroupHealthObservation) DeepCopy() *TargetGroupHealthObservation { + if in == nil { + return nil + } + out := new(TargetGroupHealthObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *TargetGroupHealthParameters) DeepCopyInto(out *TargetGroupHealthParameters) { + *out = *in + if in.DNSFailover != nil { + in, out := &in.DNSFailover, &out.DNSFailover + *out = new(DNSFailoverParameters) + (*in).DeepCopyInto(*out) + } + if in.UnhealthyStateRouting != nil { + in, out := &in.UnhealthyStateRouting, &out.UnhealthyStateRouting + *out = new(UnhealthyStateRoutingParameters) + (*in).DeepCopyInto(*out) + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TargetGroupHealthParameters. +func (in *TargetGroupHealthParameters) DeepCopy() *TargetGroupHealthParameters { + if in == nil { + return nil + } + out := new(TargetGroupHealthParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *TargetGroupInitParameters) DeepCopyInto(out *TargetGroupInitParameters) { *out = *in @@ -6451,6 +6616,11 @@ func (in *TargetHealthStateInitParameters) DeepCopyInto(out *TargetHealthStateIn *out = new(bool) **out = **in } + if in.UnhealthyDrainingInterval != nil { + in, out := &in.UnhealthyDrainingInterval, &out.UnhealthyDrainingInterval + *out = new(float64) + **out = **in + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TargetHealthStateInitParameters. @@ -6471,6 +6641,11 @@ func (in *TargetHealthStateObservation) DeepCopyInto(out *TargetHealthStateObser *out = new(bool) **out = **in } + if in.UnhealthyDrainingInterval != nil { + in, out := &in.UnhealthyDrainingInterval, &out.UnhealthyDrainingInterval + *out = new(float64) + **out = **in + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TargetHealthStateObservation. @@ -6491,6 +6666,11 @@ func (in *TargetHealthStateParameters) DeepCopyInto(out *TargetHealthStateParame *out = new(bool) **out = **in } + if in.UnhealthyDrainingInterval != nil { + in, out := &in.UnhealthyDrainingInterval, &out.UnhealthyDrainingInterval + *out = new(float64) + **out = **in + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TargetHealthStateParameters. @@ -6502,3 +6682,78 @@ func (in *TargetHealthStateParameters) DeepCopy() *TargetHealthStateParameters { in.DeepCopyInto(out) return out } + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UnhealthyStateRoutingInitParameters) DeepCopyInto(out *UnhealthyStateRoutingInitParameters) { + *out = *in + if in.MinimumHealthyTargetsCount != nil { + in, out := &in.MinimumHealthyTargetsCount, &out.MinimumHealthyTargetsCount + *out = new(float64) + **out = **in + } + if in.MinimumHealthyTargetsPercentage != nil { + in, out := &in.MinimumHealthyTargetsPercentage, &out.MinimumHealthyTargetsPercentage + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UnhealthyStateRoutingInitParameters. +func (in *UnhealthyStateRoutingInitParameters) DeepCopy() *UnhealthyStateRoutingInitParameters { + if in == nil { + return nil + } + out := new(UnhealthyStateRoutingInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UnhealthyStateRoutingObservation) DeepCopyInto(out *UnhealthyStateRoutingObservation) { + *out = *in + if in.MinimumHealthyTargetsCount != nil { + in, out := &in.MinimumHealthyTargetsCount, &out.MinimumHealthyTargetsCount + *out = new(float64) + **out = **in + } + if in.MinimumHealthyTargetsPercentage != nil { + in, out := &in.MinimumHealthyTargetsPercentage, &out.MinimumHealthyTargetsPercentage + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UnhealthyStateRoutingObservation. +func (in *UnhealthyStateRoutingObservation) DeepCopy() *UnhealthyStateRoutingObservation { + if in == nil { + return nil + } + out := new(UnhealthyStateRoutingObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UnhealthyStateRoutingParameters) DeepCopyInto(out *UnhealthyStateRoutingParameters) { + *out = *in + if in.MinimumHealthyTargetsCount != nil { + in, out := &in.MinimumHealthyTargetsCount, &out.MinimumHealthyTargetsCount + *out = new(float64) + **out = **in + } + if in.MinimumHealthyTargetsPercentage != nil { + in, out := &in.MinimumHealthyTargetsPercentage, &out.MinimumHealthyTargetsPercentage + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UnhealthyStateRoutingParameters. +func (in *UnhealthyStateRoutingParameters) DeepCopy() *UnhealthyStateRoutingParameters { + if in == nil { + return nil + } + out := new(UnhealthyStateRoutingParameters) + in.DeepCopyInto(out) + return out +} diff --git a/apis/emrserverless/v1beta1/zz_generated.deepcopy.go b/apis/emrserverless/v1beta1/zz_generated.deepcopy.go index 3ad15705ad..fa1b2ae137 100644 --- a/apis/emrserverless/v1beta1/zz_generated.deepcopy.go +++ b/apis/emrserverless/v1beta1/zz_generated.deepcopy.go @@ -75,6 +75,13 @@ func (in *ApplicationInitParameters) DeepCopyInto(out *ApplicationInitParameters (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InteractiveConfiguration != nil { + in, out := &in.InteractiveConfiguration, &out.InteractiveConfiguration + *out = make([]InteractiveConfigurationInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.MaximumCapacity != nil { in, out := &in.MaximumCapacity, &out.MaximumCapacity *out = make([]MaximumCapacityInitParameters, len(*in)) @@ -210,6 +217,13 @@ func (in *ApplicationObservation) DeepCopyInto(out *ApplicationObservation) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InteractiveConfiguration != nil { + in, out := &in.InteractiveConfiguration, &out.InteractiveConfiguration + *out = make([]InteractiveConfigurationObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.MaximumCapacity != nil { in, out := &in.MaximumCapacity, &out.MaximumCapacity *out = make([]MaximumCapacityObservation, len(*in)) @@ -319,6 +333,13 @@ func (in *ApplicationParameters) DeepCopyInto(out *ApplicationParameters) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InteractiveConfiguration != nil { + in, out := &in.InteractiveConfiguration, &out.InteractiveConfiguration + *out = make([]InteractiveConfigurationParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.MaximumCapacity != nil { in, out := &in.MaximumCapacity, &out.MaximumCapacity *out = make([]MaximumCapacityParameters, len(*in)) @@ -773,6 +794,81 @@ func (in *InitialCapacityParameters) DeepCopy() *InitialCapacityParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *InteractiveConfigurationInitParameters) DeepCopyInto(out *InteractiveConfigurationInitParameters) { + *out = *in + if in.LivyEndpointEnabled != nil { + in, out := &in.LivyEndpointEnabled, &out.LivyEndpointEnabled + *out = new(bool) + **out = **in + } + if in.StudioEnabled != nil { + in, out := &in.StudioEnabled, &out.StudioEnabled + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new InteractiveConfigurationInitParameters. +func (in *InteractiveConfigurationInitParameters) DeepCopy() *InteractiveConfigurationInitParameters { + if in == nil { + return nil + } + out := new(InteractiveConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *InteractiveConfigurationObservation) DeepCopyInto(out *InteractiveConfigurationObservation) { + *out = *in + if in.LivyEndpointEnabled != nil { + in, out := &in.LivyEndpointEnabled, &out.LivyEndpointEnabled + *out = new(bool) + **out = **in + } + if in.StudioEnabled != nil { + in, out := &in.StudioEnabled, &out.StudioEnabled + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new InteractiveConfigurationObservation. +func (in *InteractiveConfigurationObservation) DeepCopy() *InteractiveConfigurationObservation { + if in == nil { + return nil + } + out := new(InteractiveConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *InteractiveConfigurationParameters) DeepCopyInto(out *InteractiveConfigurationParameters) { + *out = *in + if in.LivyEndpointEnabled != nil { + in, out := &in.LivyEndpointEnabled, &out.LivyEndpointEnabled + *out = new(bool) + **out = **in + } + if in.StudioEnabled != nil { + in, out := &in.StudioEnabled, &out.StudioEnabled + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new InteractiveConfigurationParameters. +func (in *InteractiveConfigurationParameters) DeepCopy() *InteractiveConfigurationParameters { + if in == nil { + return nil + } + out := new(InteractiveConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *MaximumCapacityInitParameters) DeepCopyInto(out *MaximumCapacityInitParameters) { *out = *in diff --git a/apis/firehose/v1beta1/zz_generated.deepcopy.go b/apis/firehose/v1beta1/zz_generated.deepcopy.go index 3eff7a99b1..4724c8868f 100644 --- a/apis/firehose/v1beta1/zz_generated.deepcopy.go +++ b/apis/firehose/v1beta1/zz_generated.deepcopy.go @@ -2118,6 +2118,11 @@ func (in *HTTPEndpointConfigurationInitParameters) DeepCopyInto(out *HTTPEndpoin (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SecretsManagerConfigurationInitParameters) + (*in).DeepCopyInto(*out) + } if in.URL != nil { in, out := &in.URL, &out.URL *out = new(string) @@ -2196,6 +2201,11 @@ func (in *HTTPEndpointConfigurationObservation) DeepCopyInto(out *HTTPEndpointCo (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SecretsManagerConfigurationObservation) + (*in).DeepCopyInto(*out) + } if in.URL != nil { in, out := &in.URL, &out.URL *out = new(string) @@ -2289,6 +2299,11 @@ func (in *HTTPEndpointConfigurationParameters) DeepCopyInto(out *HTTPEndpointCon (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SecretsManagerConfigurationParameters) + (*in).DeepCopyInto(*out) + } if in.URL != nil { in, out := &in.URL, &out.URL *out = new(string) @@ -6406,7 +6421,11 @@ func (in *RedshiftConfigurationInitParameters) DeepCopyInto(out *RedshiftConfigu *out = new(string) **out = **in } - out.PasswordSecretRef = in.PasswordSecretRef + if in.PasswordSecretRef != nil { + in, out := &in.PasswordSecretRef, &out.PasswordSecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.ProcessingConfiguration != nil { in, out := &in.ProcessingConfiguration, &out.ProcessingConfiguration *out = make([]RedshiftConfigurationProcessingConfigurationInitParameters, len(*in)) @@ -6453,6 +6472,11 @@ func (in *RedshiftConfigurationInitParameters) DeepCopyInto(out *RedshiftConfigu (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(RedshiftConfigurationSecretsManagerConfigurationInitParameters) + (*in).DeepCopyInto(*out) + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -6536,6 +6560,11 @@ func (in *RedshiftConfigurationObservation) DeepCopyInto(out *RedshiftConfigurat (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(RedshiftConfigurationSecretsManagerConfigurationObservation) + (*in).DeepCopyInto(*out) + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -6583,7 +6612,11 @@ func (in *RedshiftConfigurationParameters) DeepCopyInto(out *RedshiftConfigurati *out = new(string) **out = **in } - out.PasswordSecretRef = in.PasswordSecretRef + if in.PasswordSecretRef != nil { + in, out := &in.PasswordSecretRef, &out.PasswordSecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.ProcessingConfiguration != nil { in, out := &in.ProcessingConfiguration, &out.ProcessingConfiguration *out = make([]RedshiftConfigurationProcessingConfigurationParameters, len(*in)) @@ -6630,6 +6663,11 @@ func (in *RedshiftConfigurationParameters) DeepCopyInto(out *RedshiftConfigurati (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(RedshiftConfigurationSecretsManagerConfigurationParameters) + (*in).DeepCopyInto(*out) + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -7516,6 +7554,96 @@ func (in *RedshiftConfigurationS3ConfigurationParameters) DeepCopy() *RedshiftCo return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *RedshiftConfigurationSecretsManagerConfigurationInitParameters) DeepCopyInto(out *RedshiftConfigurationSecretsManagerConfigurationInitParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RedshiftConfigurationSecretsManagerConfigurationInitParameters. +func (in *RedshiftConfigurationSecretsManagerConfigurationInitParameters) DeepCopy() *RedshiftConfigurationSecretsManagerConfigurationInitParameters { + if in == nil { + return nil + } + out := new(RedshiftConfigurationSecretsManagerConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *RedshiftConfigurationSecretsManagerConfigurationObservation) DeepCopyInto(out *RedshiftConfigurationSecretsManagerConfigurationObservation) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RedshiftConfigurationSecretsManagerConfigurationObservation. +func (in *RedshiftConfigurationSecretsManagerConfigurationObservation) DeepCopy() *RedshiftConfigurationSecretsManagerConfigurationObservation { + if in == nil { + return nil + } + out := new(RedshiftConfigurationSecretsManagerConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *RedshiftConfigurationSecretsManagerConfigurationParameters) DeepCopyInto(out *RedshiftConfigurationSecretsManagerConfigurationParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RedshiftConfigurationSecretsManagerConfigurationParameters. +func (in *RedshiftConfigurationSecretsManagerConfigurationParameters) DeepCopy() *RedshiftConfigurationSecretsManagerConfigurationParameters { + if in == nil { + return nil + } + out := new(RedshiftConfigurationSecretsManagerConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *RequestConfigurationInitParameters) DeepCopyInto(out *RequestConfigurationInitParameters) { *out = *in @@ -8359,6 +8487,96 @@ func (in *SchemaConfigurationParameters) DeepCopy() *SchemaConfigurationParamete return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SecretsManagerConfigurationInitParameters) DeepCopyInto(out *SecretsManagerConfigurationInitParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretsManagerConfigurationInitParameters. +func (in *SecretsManagerConfigurationInitParameters) DeepCopy() *SecretsManagerConfigurationInitParameters { + if in == nil { + return nil + } + out := new(SecretsManagerConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SecretsManagerConfigurationObservation) DeepCopyInto(out *SecretsManagerConfigurationObservation) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretsManagerConfigurationObservation. +func (in *SecretsManagerConfigurationObservation) DeepCopy() *SecretsManagerConfigurationObservation { + if in == nil { + return nil + } + out := new(SecretsManagerConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SecretsManagerConfigurationParameters) DeepCopyInto(out *SecretsManagerConfigurationParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretsManagerConfigurationParameters. +func (in *SecretsManagerConfigurationParameters) DeepCopy() *SecretsManagerConfigurationParameters { + if in == nil { + return nil + } + out := new(SecretsManagerConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *SerializerInitParameters) DeepCopyInto(out *SerializerInitParameters) { *out = *in @@ -8634,6 +8852,16 @@ func (in *SnowflakeConfigurationInitParameters) DeepCopyInto(out *SnowflakeConfi *out = new(string) **out = **in } + if in.BufferingInterval != nil { + in, out := &in.BufferingInterval, &out.BufferingInterval + *out = new(float64) + **out = **in + } + if in.BufferingSize != nil { + in, out := &in.BufferingSize, &out.BufferingSize + *out = new(float64) + **out = **in + } if in.CloudwatchLoggingOptions != nil { in, out := &in.CloudwatchLoggingOptions, &out.CloudwatchLoggingOptions *out = make([]SnowflakeConfigurationCloudwatchLoggingOptionsInitParameters, len(*in)) @@ -8666,7 +8894,11 @@ func (in *SnowflakeConfigurationInitParameters) DeepCopyInto(out *SnowflakeConfi *out = new(string) **out = **in } - out.PrivateKeySecretRef = in.PrivateKeySecretRef + if in.PrivateKeySecretRef != nil { + in, out := &in.PrivateKeySecretRef, &out.PrivateKeySecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.ProcessingConfiguration != nil { in, out := &in.ProcessingConfiguration, &out.ProcessingConfiguration *out = make([]SnowflakeConfigurationProcessingConfigurationInitParameters, len(*in)) @@ -8711,6 +8943,11 @@ func (in *SnowflakeConfigurationInitParameters) DeepCopyInto(out *SnowflakeConfi *out = new(string) **out = **in } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SnowflakeConfigurationSecretsManagerConfigurationInitParameters) + (*in).DeepCopyInto(*out) + } if in.SnowflakeRoleConfiguration != nil { in, out := &in.SnowflakeRoleConfiguration, &out.SnowflakeRoleConfiguration *out = make([]SnowflakeRoleConfigurationInitParameters, len(*in)) @@ -8755,6 +8992,16 @@ func (in *SnowflakeConfigurationObservation) DeepCopyInto(out *SnowflakeConfigur *out = new(string) **out = **in } + if in.BufferingInterval != nil { + in, out := &in.BufferingInterval, &out.BufferingInterval + *out = new(float64) + **out = **in + } + if in.BufferingSize != nil { + in, out := &in.BufferingSize, &out.BufferingSize + *out = new(float64) + **out = **in + } if in.CloudwatchLoggingOptions != nil { in, out := &in.CloudwatchLoggingOptions, &out.CloudwatchLoggingOptions *out = make([]SnowflakeConfigurationCloudwatchLoggingOptionsObservation, len(*in)) @@ -8816,6 +9063,11 @@ func (in *SnowflakeConfigurationObservation) DeepCopyInto(out *SnowflakeConfigur *out = new(string) **out = **in } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SnowflakeConfigurationSecretsManagerConfigurationObservation) + (*in).DeepCopyInto(*out) + } if in.SnowflakeRoleConfiguration != nil { in, out := &in.SnowflakeRoleConfiguration, &out.SnowflakeRoleConfiguration *out = make([]SnowflakeRoleConfigurationObservation, len(*in)) @@ -8860,6 +9112,16 @@ func (in *SnowflakeConfigurationParameters) DeepCopyInto(out *SnowflakeConfigura *out = new(string) **out = **in } + if in.BufferingInterval != nil { + in, out := &in.BufferingInterval, &out.BufferingInterval + *out = new(float64) + **out = **in + } + if in.BufferingSize != nil { + in, out := &in.BufferingSize, &out.BufferingSize + *out = new(float64) + **out = **in + } if in.CloudwatchLoggingOptions != nil { in, out := &in.CloudwatchLoggingOptions, &out.CloudwatchLoggingOptions *out = make([]SnowflakeConfigurationCloudwatchLoggingOptionsParameters, len(*in)) @@ -8892,7 +9154,11 @@ func (in *SnowflakeConfigurationParameters) DeepCopyInto(out *SnowflakeConfigura *out = new(string) **out = **in } - out.PrivateKeySecretRef = in.PrivateKeySecretRef + if in.PrivateKeySecretRef != nil { + in, out := &in.PrivateKeySecretRef, &out.PrivateKeySecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.ProcessingConfiguration != nil { in, out := &in.ProcessingConfiguration, &out.ProcessingConfiguration *out = make([]SnowflakeConfigurationProcessingConfigurationParameters, len(*in)) @@ -8937,6 +9203,11 @@ func (in *SnowflakeConfigurationParameters) DeepCopyInto(out *SnowflakeConfigura *out = new(string) **out = **in } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SnowflakeConfigurationSecretsManagerConfigurationParameters) + (*in).DeepCopyInto(*out) + } if in.SnowflakeRoleConfiguration != nil { in, out := &in.SnowflakeRoleConfiguration, &out.SnowflakeRoleConfiguration *out = make([]SnowflakeRoleConfigurationParameters, len(*in)) @@ -9526,6 +9797,96 @@ func (in *SnowflakeConfigurationS3ConfigurationParameters) DeepCopy() *Snowflake return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SnowflakeConfigurationSecretsManagerConfigurationInitParameters) DeepCopyInto(out *SnowflakeConfigurationSecretsManagerConfigurationInitParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SnowflakeConfigurationSecretsManagerConfigurationInitParameters. +func (in *SnowflakeConfigurationSecretsManagerConfigurationInitParameters) DeepCopy() *SnowflakeConfigurationSecretsManagerConfigurationInitParameters { + if in == nil { + return nil + } + out := new(SnowflakeConfigurationSecretsManagerConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SnowflakeConfigurationSecretsManagerConfigurationObservation) DeepCopyInto(out *SnowflakeConfigurationSecretsManagerConfigurationObservation) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SnowflakeConfigurationSecretsManagerConfigurationObservation. +func (in *SnowflakeConfigurationSecretsManagerConfigurationObservation) DeepCopy() *SnowflakeConfigurationSecretsManagerConfigurationObservation { + if in == nil { + return nil + } + out := new(SnowflakeConfigurationSecretsManagerConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SnowflakeConfigurationSecretsManagerConfigurationParameters) DeepCopyInto(out *SnowflakeConfigurationSecretsManagerConfigurationParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SnowflakeConfigurationSecretsManagerConfigurationParameters. +func (in *SnowflakeConfigurationSecretsManagerConfigurationParameters) DeepCopy() *SnowflakeConfigurationSecretsManagerConfigurationParameters { + if in == nil { + return nil + } + out := new(SnowflakeConfigurationSecretsManagerConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *SnowflakeRoleConfigurationInitParameters) DeepCopyInto(out *SnowflakeRoleConfigurationInitParameters) { *out = *in @@ -9786,7 +10147,11 @@ func (in *SplunkConfigurationInitParameters) DeepCopyInto(out *SplunkConfigurati *out = new(string) **out = **in } - out.HecTokenSecretRef = in.HecTokenSecretRef + if in.HecTokenSecretRef != nil { + in, out := &in.HecTokenSecretRef, &out.HecTokenSecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.ProcessingConfiguration != nil { in, out := &in.ProcessingConfiguration, &out.ProcessingConfiguration *out = make([]SplunkConfigurationProcessingConfigurationInitParameters, len(*in)) @@ -9811,6 +10176,11 @@ func (in *SplunkConfigurationInitParameters) DeepCopyInto(out *SplunkConfigurati (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SplunkConfigurationSecretsManagerConfigurationInitParameters) + (*in).DeepCopyInto(*out) + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SplunkConfigurationInitParameters. @@ -9882,6 +10252,11 @@ func (in *SplunkConfigurationObservation) DeepCopyInto(out *SplunkConfigurationO (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SplunkConfigurationSecretsManagerConfigurationObservation) + (*in).DeepCopyInto(*out) + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SplunkConfigurationObservation. @@ -9929,7 +10304,11 @@ func (in *SplunkConfigurationParameters) DeepCopyInto(out *SplunkConfigurationPa *out = new(string) **out = **in } - out.HecTokenSecretRef = in.HecTokenSecretRef + if in.HecTokenSecretRef != nil { + in, out := &in.HecTokenSecretRef, &out.HecTokenSecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.ProcessingConfiguration != nil { in, out := &in.ProcessingConfiguration, &out.ProcessingConfiguration *out = make([]SplunkConfigurationProcessingConfigurationParameters, len(*in)) @@ -9954,6 +10333,11 @@ func (in *SplunkConfigurationParameters) DeepCopyInto(out *SplunkConfigurationPa (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SecretsManagerConfiguration != nil { + in, out := &in.SecretsManagerConfiguration, &out.SecretsManagerConfiguration + *out = new(SplunkConfigurationSecretsManagerConfigurationParameters) + (*in).DeepCopyInto(*out) + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SplunkConfigurationParameters. @@ -10519,6 +10903,96 @@ func (in *SplunkConfigurationS3ConfigurationParameters) DeepCopy() *SplunkConfig return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SplunkConfigurationSecretsManagerConfigurationInitParameters) DeepCopyInto(out *SplunkConfigurationSecretsManagerConfigurationInitParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SplunkConfigurationSecretsManagerConfigurationInitParameters. +func (in *SplunkConfigurationSecretsManagerConfigurationInitParameters) DeepCopy() *SplunkConfigurationSecretsManagerConfigurationInitParameters { + if in == nil { + return nil + } + out := new(SplunkConfigurationSecretsManagerConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SplunkConfigurationSecretsManagerConfigurationObservation) DeepCopyInto(out *SplunkConfigurationSecretsManagerConfigurationObservation) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SplunkConfigurationSecretsManagerConfigurationObservation. +func (in *SplunkConfigurationSecretsManagerConfigurationObservation) DeepCopy() *SplunkConfigurationSecretsManagerConfigurationObservation { + if in == nil { + return nil + } + out := new(SplunkConfigurationSecretsManagerConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SplunkConfigurationSecretsManagerConfigurationParameters) DeepCopyInto(out *SplunkConfigurationSecretsManagerConfigurationParameters) { + *out = *in + if in.Enabled != nil { + in, out := &in.Enabled, &out.Enabled + *out = new(bool) + **out = **in + } + if in.RoleArn != nil { + in, out := &in.RoleArn, &out.RoleArn + *out = new(string) + **out = **in + } + if in.SecretArn != nil { + in, out := &in.SecretArn, &out.SecretArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SplunkConfigurationSecretsManagerConfigurationParameters. +func (in *SplunkConfigurationSecretsManagerConfigurationParameters) DeepCopy() *SplunkConfigurationSecretsManagerConfigurationParameters { + if in == nil { + return nil + } + out := new(SplunkConfigurationSecretsManagerConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *VPCConfigInitParameters) DeepCopyInto(out *VPCConfigInitParameters) { *out = *in diff --git a/apis/fis/v1beta1/zz_generated.deepcopy.go b/apis/fis/v1beta1/zz_generated.deepcopy.go index b0fe07d360..82ade68251 100644 --- a/apis/fis/v1beta1/zz_generated.deepcopy.go +++ b/apis/fis/v1beta1/zz_generated.deepcopy.go @@ -238,6 +238,81 @@ func (in *CloudwatchLogsConfigurationParameters) DeepCopy() *CloudwatchLogsConfi return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ExperimentOptionsInitParameters) DeepCopyInto(out *ExperimentOptionsInitParameters) { + *out = *in + if in.AccountTargeting != nil { + in, out := &in.AccountTargeting, &out.AccountTargeting + *out = new(string) + **out = **in + } + if in.EmptyTargetResolutionMode != nil { + in, out := &in.EmptyTargetResolutionMode, &out.EmptyTargetResolutionMode + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExperimentOptionsInitParameters. +func (in *ExperimentOptionsInitParameters) DeepCopy() *ExperimentOptionsInitParameters { + if in == nil { + return nil + } + out := new(ExperimentOptionsInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ExperimentOptionsObservation) DeepCopyInto(out *ExperimentOptionsObservation) { + *out = *in + if in.AccountTargeting != nil { + in, out := &in.AccountTargeting, &out.AccountTargeting + *out = new(string) + **out = **in + } + if in.EmptyTargetResolutionMode != nil { + in, out := &in.EmptyTargetResolutionMode, &out.EmptyTargetResolutionMode + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExperimentOptionsObservation. +func (in *ExperimentOptionsObservation) DeepCopy() *ExperimentOptionsObservation { + if in == nil { + return nil + } + out := new(ExperimentOptionsObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ExperimentOptionsParameters) DeepCopyInto(out *ExperimentOptionsParameters) { + *out = *in + if in.AccountTargeting != nil { + in, out := &in.AccountTargeting, &out.AccountTargeting + *out = new(string) + **out = **in + } + if in.EmptyTargetResolutionMode != nil { + in, out := &in.EmptyTargetResolutionMode, &out.EmptyTargetResolutionMode + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExperimentOptionsParameters. +func (in *ExperimentOptionsParameters) DeepCopy() *ExperimentOptionsParameters { + if in == nil { + return nil + } + out := new(ExperimentOptionsParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *ExperimentTemplate) DeepCopyInto(out *ExperimentTemplate) { *out = *in @@ -280,6 +355,13 @@ func (in *ExperimentTemplateInitParameters) DeepCopyInto(out *ExperimentTemplate *out = new(string) **out = **in } + if in.ExperimentOptions != nil { + in, out := &in.ExperimentOptions, &out.ExperimentOptions + *out = make([]ExperimentOptionsInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.LogConfiguration != nil { in, out := &in.LogConfiguration, &out.LogConfiguration *out = make([]LogConfigurationInitParameters, len(*in)) @@ -391,6 +473,13 @@ func (in *ExperimentTemplateObservation) DeepCopyInto(out *ExperimentTemplateObs *out = new(string) **out = **in } + if in.ExperimentOptions != nil { + in, out := &in.ExperimentOptions, &out.ExperimentOptions + *out = make([]ExperimentOptionsObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ID != nil { in, out := &in.ID, &out.ID *out = new(string) @@ -481,6 +570,13 @@ func (in *ExperimentTemplateParameters) DeepCopyInto(out *ExperimentTemplatePara *out = new(string) **out = **in } + if in.ExperimentOptions != nil { + in, out := &in.ExperimentOptions, &out.ExperimentOptions + *out = make([]ExperimentOptionsParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.LogConfiguration != nil { in, out := &in.LogConfiguration, &out.LogConfiguration *out = make([]LogConfigurationParameters, len(*in)) diff --git a/apis/fsx/v1beta1/zz_generated.deepcopy.go b/apis/fsx/v1beta1/zz_generated.deepcopy.go index 0466f34bfd..49275040d8 100644 --- a/apis/fsx/v1beta1/zz_generated.deepcopy.go +++ b/apis/fsx/v1beta1/zz_generated.deepcopy.go @@ -1449,6 +1449,22 @@ func (in *LustreFileSystemInitParameters) DeepCopyInto(out *LustreFileSystemInit *out = new(string) **out = **in } + if in.FinalBackupTags != nil { + in, out := &in.FinalBackupTags, &out.FinalBackupTags + *out = make(map[string]*string, len(*in)) + for key, val := range *in { + var outVal *string + if val == nil { + (*out)[key] = nil + } else { + inVal := (*in)[key] + in, out := &inVal, &outVal + *out = new(string) + **out = **in + } + (*out)[key] = outVal + } + } if in.ImportPath != nil { in, out := &in.ImportPath, &out.ImportPath *out = new(string) @@ -1521,6 +1537,11 @@ func (in *LustreFileSystemInitParameters) DeepCopyInto(out *LustreFileSystemInit } } } + if in.SkipFinalBackup != nil { + in, out := &in.SkipFinalBackup, &out.SkipFinalBackup + *out = new(bool) + **out = **in + } if in.StorageCapacity != nil { in, out := &in.StorageCapacity, &out.StorageCapacity *out = new(float64) @@ -1682,6 +1703,22 @@ func (in *LustreFileSystemObservation) DeepCopyInto(out *LustreFileSystemObserva *out = new(string) **out = **in } + if in.FinalBackupTags != nil { + in, out := &in.FinalBackupTags, &out.FinalBackupTags + *out = make(map[string]*string, len(*in)) + for key, val := range *in { + var outVal *string + if val == nil { + (*out)[key] = nil + } else { + inVal := (*in)[key] + in, out := &inVal, &outVal + *out = new(string) + **out = **in + } + (*out)[key] = outVal + } + } if in.ID != nil { in, out := &in.ID, &out.ID *out = new(string) @@ -1758,6 +1795,11 @@ func (in *LustreFileSystemObservation) DeepCopyInto(out *LustreFileSystemObserva } } } + if in.SkipFinalBackup != nil { + in, out := &in.SkipFinalBackup, &out.SkipFinalBackup + *out = new(bool) + **out = **in + } if in.StorageCapacity != nil { in, out := &in.StorageCapacity, &out.StorageCapacity *out = new(float64) @@ -1886,6 +1928,22 @@ func (in *LustreFileSystemParameters) DeepCopyInto(out *LustreFileSystemParamete *out = new(string) **out = **in } + if in.FinalBackupTags != nil { + in, out := &in.FinalBackupTags, &out.FinalBackupTags + *out = make(map[string]*string, len(*in)) + for key, val := range *in { + var outVal *string + if val == nil { + (*out)[key] = nil + } else { + inVal := (*in)[key] + in, out := &inVal, &outVal + *out = new(string) + **out = **in + } + (*out)[key] = outVal + } + } if in.ImportPath != nil { in, out := &in.ImportPath, &out.ImportPath *out = new(string) @@ -1963,6 +2021,11 @@ func (in *LustreFileSystemParameters) DeepCopyInto(out *LustreFileSystemParamete } } } + if in.SkipFinalBackup != nil { + in, out := &in.SkipFinalBackup, &out.SkipFinalBackup + *out = new(bool) + **out = **in + } if in.StorageCapacity != nil { in, out := &in.StorageCapacity, &out.StorageCapacity *out = new(float64) @@ -4005,6 +4068,22 @@ func (in *WindowsFileSystemInitParameters) DeepCopyInto(out *WindowsFileSystemIn (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.FinalBackupTags != nil { + in, out := &in.FinalBackupTags, &out.FinalBackupTags + *out = make(map[string]*string, len(*in)) + for key, val := range *in { + var outVal *string + if val == nil { + (*out)[key] = nil + } else { + inVal := (*in)[key] + in, out := &inVal, &outVal + *out = new(string) + **out = **in + } + (*out)[key] = outVal + } + } if in.KMSKeyID != nil { in, out := &in.KMSKeyID, &out.KMSKeyID *out = new(string) @@ -4231,6 +4310,22 @@ func (in *WindowsFileSystemObservation) DeepCopyInto(out *WindowsFileSystemObser (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.FinalBackupTags != nil { + in, out := &in.FinalBackupTags, &out.FinalBackupTags + *out = make(map[string]*string, len(*in)) + for key, val := range *in { + var outVal *string + if val == nil { + (*out)[key] = nil + } else { + inVal := (*in)[key] + in, out := &inVal, &outVal + *out = new(string) + **out = **in + } + (*out)[key] = outVal + } + } if in.ID != nil { in, out := &in.ID, &out.ID *out = new(string) @@ -4443,6 +4538,22 @@ func (in *WindowsFileSystemParameters) DeepCopyInto(out *WindowsFileSystemParame (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.FinalBackupTags != nil { + in, out := &in.FinalBackupTags, &out.FinalBackupTags + *out = make(map[string]*string, len(*in)) + for key, val := range *in { + var outVal *string + if val == nil { + (*out)[key] = nil + } else { + inVal := (*in)[key] + in, out := &inVal, &outVal + *out = new(string) + **out = **in + } + (*out)[key] = outVal + } + } if in.KMSKeyID != nil { in, out := &in.KMSKeyID, &out.KMSKeyID *out = new(string) diff --git a/apis/glue/v1beta1/zz_generated.deepcopy.go b/apis/glue/v1beta1/zz_generated.deepcopy.go index c57b67ba64..e7a5bd3b66 100644 --- a/apis/glue/v1beta1/zz_generated.deepcopy.go +++ b/apis/glue/v1beta1/zz_generated.deepcopy.go @@ -5125,6 +5125,11 @@ func (in *JobInitParameters) DeepCopyInto(out *JobInitParameters) { *out = new(string) **out = **in } + if in.JobRunQueuingEnabled != nil { + in, out := &in.JobRunQueuingEnabled, &out.JobRunQueuingEnabled + *out = new(bool) + **out = **in + } if in.MaintenanceWindow != nil { in, out := &in.MaintenanceWindow, &out.MaintenanceWindow *out = new(string) @@ -5327,6 +5332,11 @@ func (in *JobObservation) DeepCopyInto(out *JobObservation) { *out = new(string) **out = **in } + if in.JobRunQueuingEnabled != nil { + in, out := &in.JobRunQueuingEnabled, &out.JobRunQueuingEnabled + *out = new(bool) + **out = **in + } if in.MaintenanceWindow != nil { in, out := &in.MaintenanceWindow, &out.MaintenanceWindow *out = new(string) @@ -5493,6 +5503,11 @@ func (in *JobParameters) DeepCopyInto(out *JobParameters) { *out = new(string) **out = **in } + if in.JobRunQueuingEnabled != nil { + in, out := &in.JobRunQueuingEnabled, &out.JobRunQueuingEnabled + *out = new(bool) + **out = **in + } if in.MaintenanceWindow != nil { in, out := &in.MaintenanceWindow, &out.MaintenanceWindow *out = new(string) diff --git a/apis/imagebuilder/v1beta1/zz_generated.deepcopy.go b/apis/imagebuilder/v1beta1/zz_generated.deepcopy.go index 66499d9d1f..768d9fa9ee 100644 --- a/apis/imagebuilder/v1beta1/zz_generated.deepcopy.go +++ b/apis/imagebuilder/v1beta1/zz_generated.deepcopy.go @@ -3349,6 +3349,11 @@ func (in *ImagePipelineInitParameters) DeepCopyInto(out *ImagePipelineInitParame *out = new(bool) **out = **in } + if in.ExecutionRole != nil { + in, out := &in.ExecutionRole, &out.ExecutionRole + *out = new(string) + **out = **in + } if in.ImageRecipeArn != nil { in, out := &in.ImageRecipeArn, &out.ImageRecipeArn *out = new(string) @@ -3426,6 +3431,13 @@ func (in *ImagePipelineInitParameters) DeepCopyInto(out *ImagePipelineInitParame (*out)[key] = outVal } } + if in.Workflow != nil { + in, out := &in.Workflow, &out.Workflow + *out = make([]ImagePipelineWorkflowInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineInitParameters. @@ -3518,6 +3530,11 @@ func (in *ImagePipelineObservation) DeepCopyInto(out *ImagePipelineObservation) *out = new(bool) **out = **in } + if in.ExecutionRole != nil { + in, out := &in.ExecutionRole, &out.ExecutionRole + *out = new(string) + **out = **in + } if in.ID != nil { in, out := &in.ID, &out.ID *out = new(string) @@ -3601,6 +3618,13 @@ func (in *ImagePipelineObservation) DeepCopyInto(out *ImagePipelineObservation) (*out)[key] = outVal } } + if in.Workflow != nil { + in, out := &in.Workflow, &out.Workflow + *out = make([]ImagePipelineWorkflowObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineObservation. @@ -3636,6 +3660,11 @@ func (in *ImagePipelineParameters) DeepCopyInto(out *ImagePipelineParameters) { *out = new(bool) **out = **in } + if in.ExecutionRole != nil { + in, out := &in.ExecutionRole, &out.ExecutionRole + *out = new(string) + **out = **in + } if in.ImageRecipeArn != nil { in, out := &in.ImageRecipeArn, &out.ImageRecipeArn *out = new(string) @@ -3718,6 +3747,13 @@ func (in *ImagePipelineParameters) DeepCopyInto(out *ImagePipelineParameters) { (*out)[key] = outVal } } + if in.Workflow != nil { + in, out := &in.Workflow, &out.Workflow + *out = make([]ImagePipelineWorkflowParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineParameters. @@ -3765,6 +3801,192 @@ func (in *ImagePipelineStatus) DeepCopy() *ImagePipelineStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePipelineWorkflowInitParameters) DeepCopyInto(out *ImagePipelineWorkflowInitParameters) { + *out = *in + if in.OnFailure != nil { + in, out := &in.OnFailure, &out.OnFailure + *out = new(string) + **out = **in + } + if in.ParallelGroup != nil { + in, out := &in.ParallelGroup, &out.ParallelGroup + *out = new(string) + **out = **in + } + if in.Parameter != nil { + in, out := &in.Parameter, &out.Parameter + *out = make([]ImagePipelineWorkflowParameterInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.WorkflowArn != nil { + in, out := &in.WorkflowArn, &out.WorkflowArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineWorkflowInitParameters. +func (in *ImagePipelineWorkflowInitParameters) DeepCopy() *ImagePipelineWorkflowInitParameters { + if in == nil { + return nil + } + out := new(ImagePipelineWorkflowInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePipelineWorkflowObservation) DeepCopyInto(out *ImagePipelineWorkflowObservation) { + *out = *in + if in.OnFailure != nil { + in, out := &in.OnFailure, &out.OnFailure + *out = new(string) + **out = **in + } + if in.ParallelGroup != nil { + in, out := &in.ParallelGroup, &out.ParallelGroup + *out = new(string) + **out = **in + } + if in.Parameter != nil { + in, out := &in.Parameter, &out.Parameter + *out = make([]ImagePipelineWorkflowParameterObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.WorkflowArn != nil { + in, out := &in.WorkflowArn, &out.WorkflowArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineWorkflowObservation. +func (in *ImagePipelineWorkflowObservation) DeepCopy() *ImagePipelineWorkflowObservation { + if in == nil { + return nil + } + out := new(ImagePipelineWorkflowObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePipelineWorkflowParameterInitParameters) DeepCopyInto(out *ImagePipelineWorkflowParameterInitParameters) { + *out = *in + if in.Name != nil { + in, out := &in.Name, &out.Name + *out = new(string) + **out = **in + } + if in.Value != nil { + in, out := &in.Value, &out.Value + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineWorkflowParameterInitParameters. +func (in *ImagePipelineWorkflowParameterInitParameters) DeepCopy() *ImagePipelineWorkflowParameterInitParameters { + if in == nil { + return nil + } + out := new(ImagePipelineWorkflowParameterInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePipelineWorkflowParameterObservation) DeepCopyInto(out *ImagePipelineWorkflowParameterObservation) { + *out = *in + if in.Name != nil { + in, out := &in.Name, &out.Name + *out = new(string) + **out = **in + } + if in.Value != nil { + in, out := &in.Value, &out.Value + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineWorkflowParameterObservation. +func (in *ImagePipelineWorkflowParameterObservation) DeepCopy() *ImagePipelineWorkflowParameterObservation { + if in == nil { + return nil + } + out := new(ImagePipelineWorkflowParameterObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePipelineWorkflowParameterParameters) DeepCopyInto(out *ImagePipelineWorkflowParameterParameters) { + *out = *in + if in.Name != nil { + in, out := &in.Name, &out.Name + *out = new(string) + **out = **in + } + if in.Value != nil { + in, out := &in.Value, &out.Value + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineWorkflowParameterParameters. +func (in *ImagePipelineWorkflowParameterParameters) DeepCopy() *ImagePipelineWorkflowParameterParameters { + if in == nil { + return nil + } + out := new(ImagePipelineWorkflowParameterParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ImagePipelineWorkflowParameters) DeepCopyInto(out *ImagePipelineWorkflowParameters) { + *out = *in + if in.OnFailure != nil { + in, out := &in.OnFailure, &out.OnFailure + *out = new(string) + **out = **in + } + if in.ParallelGroup != nil { + in, out := &in.ParallelGroup, &out.ParallelGroup + *out = new(string) + **out = **in + } + if in.Parameter != nil { + in, out := &in.Parameter, &out.Parameter + *out = make([]ImagePipelineWorkflowParameterParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.WorkflowArn != nil { + in, out := &in.WorkflowArn, &out.WorkflowArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ImagePipelineWorkflowParameters. +func (in *ImagePipelineWorkflowParameters) DeepCopy() *ImagePipelineWorkflowParameters { + if in == nil { + return nil + } + out := new(ImagePipelineWorkflowParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *ImageRecipe) DeepCopyInto(out *ImageRecipe) { *out = *in diff --git a/apis/lambda/v1beta1/zz_generated.deepcopy.go b/apis/lambda/v1beta1/zz_generated.deepcopy.go index 7c216d5417..5bad35c7e0 100644 --- a/apis/lambda/v1beta1/zz_generated.deepcopy.go +++ b/apis/lambda/v1beta1/zz_generated.deepcopy.go @@ -1386,6 +1386,21 @@ func (in *EventSourceMappingInitParameters) DeepCopyInto(out *EventSourceMapping } } } + if in.KMSKeyArn != nil { + in, out := &in.KMSKeyArn, &out.KMSKeyArn + *out = new(string) + **out = **in + } + if in.KMSKeyArnRef != nil { + in, out := &in.KMSKeyArnRef, &out.KMSKeyArnRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.KMSKeyArnSelector != nil { + in, out := &in.KMSKeyArnSelector, &out.KMSKeyArnSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } if in.MaximumBatchingWindowInSeconds != nil { in, out := &in.MaximumBatchingWindowInSeconds, &out.MaximumBatchingWindowInSeconds *out = new(float64) @@ -1592,6 +1607,11 @@ func (in *EventSourceMappingObservation) DeepCopyInto(out *EventSourceMappingObs *out = new(string) **out = **in } + if in.KMSKeyArn != nil { + in, out := &in.KMSKeyArn, &out.KMSKeyArn + *out = new(string) + **out = **in + } if in.LastModified != nil { in, out := &in.LastModified, &out.LastModified *out = new(string) @@ -1791,6 +1811,21 @@ func (in *EventSourceMappingParameters) DeepCopyInto(out *EventSourceMappingPara } } } + if in.KMSKeyArn != nil { + in, out := &in.KMSKeyArn, &out.KMSKeyArn + *out = new(string) + **out = **in + } + if in.KMSKeyArnRef != nil { + in, out := &in.KMSKeyArnRef, &out.KMSKeyArnRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.KMSKeyArnSelector != nil { + in, out := &in.KMSKeyArnSelector, &out.KMSKeyArnSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } if in.MaximumBatchingWindowInSeconds != nil { in, out := &in.MaximumBatchingWindowInSeconds, &out.MaximumBatchingWindowInSeconds *out = new(float64) diff --git a/apis/lambda/v1beta1/zz_generated.resolvers.go b/apis/lambda/v1beta1/zz_generated.resolvers.go index 25414b302b..82d9c75ace 100644 --- a/apis/lambda/v1beta1/zz_generated.resolvers.go +++ b/apis/lambda/v1beta1/zz_generated.resolvers.go @@ -132,6 +132,25 @@ func (mg *EventSourceMapping) ResolveReferences(ctx context.Context, c client.Re } mg.Spec.ForProvider.FunctionName = reference.ToPtrValue(rsp.ResolvedValue) mg.Spec.ForProvider.FunctionNameRef = rsp.ResolvedReference + { + m, l, err = apisresolver.GetManagedResource("kms.aws.upbound.io", "v1beta1", "Key", "KeyList") + if err != nil { + return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") + } + + rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ + CurrentValue: reference.FromPtrValue(mg.Spec.ForProvider.KMSKeyArn), + Extract: reference.ExternalName(), + Reference: mg.Spec.ForProvider.KMSKeyArnRef, + Selector: mg.Spec.ForProvider.KMSKeyArnSelector, + To: reference.To{List: l, Managed: m}, + }) + } + if err != nil { + return errors.Wrap(err, "mg.Spec.ForProvider.KMSKeyArn") + } + mg.Spec.ForProvider.KMSKeyArn = reference.ToPtrValue(rsp.ResolvedValue) + mg.Spec.ForProvider.KMSKeyArnRef = rsp.ResolvedReference { m, l, err = apisresolver.GetManagedResource("lambda.aws.upbound.io", "v1beta1", "Function", "FunctionList") if err != nil { @@ -151,6 +170,25 @@ func (mg *EventSourceMapping) ResolveReferences(ctx context.Context, c client.Re } mg.Spec.InitProvider.FunctionName = reference.ToPtrValue(rsp.ResolvedValue) mg.Spec.InitProvider.FunctionNameRef = rsp.ResolvedReference + { + m, l, err = apisresolver.GetManagedResource("kms.aws.upbound.io", "v1beta1", "Key", "KeyList") + if err != nil { + return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") + } + + rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ + CurrentValue: reference.FromPtrValue(mg.Spec.InitProvider.KMSKeyArn), + Extract: reference.ExternalName(), + Reference: mg.Spec.InitProvider.KMSKeyArnRef, + Selector: mg.Spec.InitProvider.KMSKeyArnSelector, + To: reference.To{List: l, Managed: m}, + }) + } + if err != nil { + return errors.Wrap(err, "mg.Spec.InitProvider.KMSKeyArn") + } + mg.Spec.InitProvider.KMSKeyArn = reference.ToPtrValue(rsp.ResolvedValue) + mg.Spec.InitProvider.KMSKeyArnRef = rsp.ResolvedReference return nil } diff --git a/apis/opensearch/v1beta1/zz_generated.deepcopy.go b/apis/opensearch/v1beta1/zz_generated.deepcopy.go index 5f67d8f049..61ae94b7bd 100644 --- a/apis/opensearch/v1beta1/zz_generated.deepcopy.go +++ b/apis/opensearch/v1beta1/zz_generated.deepcopy.go @@ -1031,6 +1031,11 @@ func (in *DomainObservation) DeepCopyInto(out *DomainObservation) { *out = new(string) **out = **in } + if in.DashboardEndpointV2 != nil { + in, out := &in.DashboardEndpointV2, &out.DashboardEndpointV2 + *out = new(string) + **out = **in + } if in.DomainEndpointOptions != nil { in, out := &in.DomainEndpointOptions, &out.DomainEndpointOptions *out = make([]DomainEndpointOptionsObservation, len(*in)) @@ -1038,6 +1043,11 @@ func (in *DomainObservation) DeepCopyInto(out *DomainObservation) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.DomainEndpointV2HostedZoneID != nil { + in, out := &in.DomainEndpointV2HostedZoneID, &out.DomainEndpointV2HostedZoneID + *out = new(string) + **out = **in + } if in.DomainID != nil { in, out := &in.DomainID, &out.DomainID *out = new(string) @@ -1067,6 +1077,11 @@ func (in *DomainObservation) DeepCopyInto(out *DomainObservation) { *out = new(string) **out = **in } + if in.EndpointV2 != nil { + in, out := &in.EndpointV2, &out.EndpointV2 + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) diff --git a/apis/rds/v1beta1/zz_generated.deepcopy.go b/apis/rds/v1beta1/zz_generated.deepcopy.go index f32891fdad..d34fa305e3 100644 --- a/apis/rds/v1beta1/zz_generated.deepcopy.go +++ b/apis/rds/v1beta1/zz_generated.deepcopy.go @@ -1019,6 +1019,11 @@ func (in *ClusterInitParameters) DeepCopyInto(out *ClusterInitParameters) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineMode != nil { in, out := &in.EngineMode, &out.EngineMode *out = new(string) @@ -1099,6 +1104,21 @@ func (in *ClusterInitParameters) DeepCopyInto(out *ClusterInitParameters) { *out = new(string) **out = **in } + if in.PerformanceInsightsEnabled != nil { + in, out := &in.PerformanceInsightsEnabled, &out.PerformanceInsightsEnabled + *out = new(bool) + **out = **in + } + if in.PerformanceInsightsKMSKeyID != nil { + in, out := &in.PerformanceInsightsKMSKeyID, &out.PerformanceInsightsKMSKeyID + *out = new(string) + **out = **in + } + if in.PerformanceInsightsRetentionPeriod != nil { + in, out := &in.PerformanceInsightsRetentionPeriod, &out.PerformanceInsightsRetentionPeriod + *out = new(float64) + **out = **in + } if in.Port != nil { in, out := &in.Port, &out.Port *out = new(float64) @@ -2149,6 +2169,11 @@ func (in *ClusterObservation) DeepCopyInto(out *ClusterObservation) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineMode != nil { in, out := &in.EngineMode, &out.EngineMode *out = new(string) @@ -2237,6 +2262,21 @@ func (in *ClusterObservation) DeepCopyInto(out *ClusterObservation) { *out = new(string) **out = **in } + if in.PerformanceInsightsEnabled != nil { + in, out := &in.PerformanceInsightsEnabled, &out.PerformanceInsightsEnabled + *out = new(bool) + **out = **in + } + if in.PerformanceInsightsKMSKeyID != nil { + in, out := &in.PerformanceInsightsKMSKeyID, &out.PerformanceInsightsKMSKeyID + *out = new(string) + **out = **in + } + if in.PerformanceInsightsRetentionPeriod != nil { + in, out := &in.PerformanceInsightsRetentionPeriod, &out.PerformanceInsightsRetentionPeriod + *out = new(float64) + **out = **in + } if in.Port != nil { in, out := &in.Port, &out.Port *out = new(float64) @@ -2905,6 +2945,11 @@ func (in *ClusterParameters) DeepCopyInto(out *ClusterParameters) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineMode != nil { in, out := &in.EngineMode, &out.EngineMode *out = new(string) @@ -2985,6 +3030,21 @@ func (in *ClusterParameters) DeepCopyInto(out *ClusterParameters) { *out = new(string) **out = **in } + if in.PerformanceInsightsEnabled != nil { + in, out := &in.PerformanceInsightsEnabled, &out.PerformanceInsightsEnabled + *out = new(bool) + **out = **in + } + if in.PerformanceInsightsKMSKeyID != nil { + in, out := &in.PerformanceInsightsKMSKeyID, &out.PerformanceInsightsKMSKeyID + *out = new(string) + **out = **in + } + if in.PerformanceInsightsRetentionPeriod != nil { + in, out := &in.PerformanceInsightsRetentionPeriod, &out.PerformanceInsightsRetentionPeriod + *out = new(float64) + **out = **in + } if in.Port != nil { in, out := &in.Port, &out.Port *out = new(float64) @@ -3142,6 +3202,11 @@ func (in *ClusterRestoreToPointInTimeInitParameters) DeepCopyInto(out *ClusterRe *out = new(v1.Selector) (*in).DeepCopyInto(*out) } + if in.SourceClusterResourceID != nil { + in, out := &in.SourceClusterResourceID, &out.SourceClusterResourceID + *out = new(string) + **out = **in + } if in.UseLatestRestorableTime != nil { in, out := &in.UseLatestRestorableTime, &out.UseLatestRestorableTime *out = new(bool) @@ -3177,6 +3242,11 @@ func (in *ClusterRestoreToPointInTimeObservation) DeepCopyInto(out *ClusterResto *out = new(string) **out = **in } + if in.SourceClusterResourceID != nil { + in, out := &in.SourceClusterResourceID, &out.SourceClusterResourceID + *out = new(string) + **out = **in + } if in.UseLatestRestorableTime != nil { in, out := &in.UseLatestRestorableTime, &out.UseLatestRestorableTime *out = new(bool) @@ -3222,6 +3292,11 @@ func (in *ClusterRestoreToPointInTimeParameters) DeepCopyInto(out *ClusterRestor *out = new(v1.Selector) (*in).DeepCopyInto(*out) } + if in.SourceClusterResourceID != nil { + in, out := &in.SourceClusterResourceID, &out.SourceClusterResourceID + *out = new(string) + **out = **in + } if in.UseLatestRestorableTime != nil { in, out := &in.UseLatestRestorableTime, &out.UseLatestRestorableTime *out = new(bool) @@ -5653,6 +5728,11 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(float64) **out = **in } + if in.BackupTarget != nil { + in, out := &in.BackupTarget, &out.BackupTarget + *out = new(string) + **out = **in + } if in.BackupWindow != nil { in, out := &in.BackupWindow, &out.BackupWindow *out = new(string) @@ -5710,6 +5790,11 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(v1.Selector) (*in).DeepCopyInto(*out) } + if in.DedicatedLogVolume != nil { + in, out := &in.DedicatedLogVolume, &out.DedicatedLogVolume + *out = new(bool) + **out = **in + } if in.DeleteAutomatedBackups != nil { in, out := &in.DeleteAutomatedBackups, &out.DeleteAutomatedBackups *out = new(bool) @@ -5725,11 +5810,37 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(string) **out = **in } + if in.DomainAuthSecretArn != nil { + in, out := &in.DomainAuthSecretArn, &out.DomainAuthSecretArn + *out = new(string) + **out = **in + } + if in.DomainDNSIps != nil { + in, out := &in.DomainDNSIps, &out.DomainDNSIps + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.DomainFqdn != nil { + in, out := &in.DomainFqdn, &out.DomainFqdn + *out = new(string) + **out = **in + } if in.DomainIAMRoleName != nil { in, out := &in.DomainIAMRoleName, &out.DomainIAMRoleName *out = new(string) **out = **in } + if in.DomainOu != nil { + in, out := &in.DomainOu, &out.DomainOu + *out = new(string) + **out = **in + } if in.EnabledCloudwatchLogsExports != nil { in, out := &in.EnabledCloudwatchLogsExports, &out.EnabledCloudwatchLogsExports *out = make([]*string, len(*in)) @@ -5746,6 +5857,11 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) @@ -5761,6 +5877,16 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(bool) **out = **in } + if in.Identifier != nil { + in, out := &in.Identifier, &out.Identifier + *out = new(string) + **out = **in + } + if in.IdentifierPrefix != nil { + in, out := &in.IdentifierPrefix, &out.IdentifierPrefix + *out = new(string) + **out = **in + } if in.InstanceClass != nil { in, out := &in.InstanceClass, &out.InstanceClass *out = new(string) @@ -5871,6 +5997,21 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(string) **out = **in } + if in.ParameterGroupNameRef != nil { + in, out := &in.ParameterGroupNameRef, &out.ParameterGroupNameRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.ParameterGroupNameSelector != nil { + in, out := &in.ParameterGroupNameSelector, &out.ParameterGroupNameSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } + if in.PasswordSecretRef != nil { + in, out := &in.PasswordSecretRef, &out.PasswordSecretRef + *out = new(v1.SecretKeySelector) + **out = **in + } if in.PerformanceInsightsEnabled != nil { in, out := &in.PerformanceInsightsEnabled, &out.PerformanceInsightsEnabled *out = new(bool) @@ -5987,6 +6128,11 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(string) **out = **in } + if in.UpgradeStorageConfig != nil { + in, out := &in.UpgradeStorageConfig, &out.UpgradeStorageConfig + *out = new(bool) + **out = **in + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -6102,6 +6248,11 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(float64) **out = **in } + if in.BackupTarget != nil { + in, out := &in.BackupTarget, &out.BackupTarget + *out = new(string) + **out = **in + } if in.BackupWindow != nil { in, out := &in.BackupWindow, &out.BackupWindow *out = new(string) @@ -6149,6 +6300,11 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.DedicatedLogVolume != nil { + in, out := &in.DedicatedLogVolume, &out.DedicatedLogVolume + *out = new(bool) + **out = **in + } if in.DeleteAutomatedBackups != nil { in, out := &in.DeleteAutomatedBackups, &out.DeleteAutomatedBackups *out = new(bool) @@ -6164,11 +6320,37 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.DomainAuthSecretArn != nil { + in, out := &in.DomainAuthSecretArn, &out.DomainAuthSecretArn + *out = new(string) + **out = **in + } + if in.DomainDNSIps != nil { + in, out := &in.DomainDNSIps, &out.DomainDNSIps + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.DomainFqdn != nil { + in, out := &in.DomainFqdn, &out.DomainFqdn + *out = new(string) + **out = **in + } if in.DomainIAMRoleName != nil { in, out := &in.DomainIAMRoleName, &out.DomainIAMRoleName *out = new(string) **out = **in } + if in.DomainOu != nil { + in, out := &in.DomainOu, &out.DomainOu + *out = new(string) + **out = **in + } if in.EnabledCloudwatchLogsExports != nil { in, out := &in.EnabledCloudwatchLogsExports, &out.EnabledCloudwatchLogsExports *out = make([]*string, len(*in)) @@ -6190,6 +6372,11 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) @@ -6220,6 +6407,16 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.Identifier != nil { + in, out := &in.Identifier, &out.Identifier + *out = new(string) + **out = **in + } + if in.IdentifierPrefix != nil { + in, out := &in.IdentifierPrefix, &out.IdentifierPrefix + *out = new(string) + **out = **in + } if in.InstanceClass != nil { in, out := &in.InstanceClass, &out.InstanceClass *out = new(string) @@ -6462,6 +6659,11 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.UpgradeStorageConfig != nil { + in, out := &in.UpgradeStorageConfig, &out.UpgradeStorageConfig + *out = new(bool) + **out = **in + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -6528,6 +6730,11 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(float64) **out = **in } + if in.BackupTarget != nil { + in, out := &in.BackupTarget, &out.BackupTarget + *out = new(string) + **out = **in + } if in.BackupWindow != nil { in, out := &in.BackupWindow, &out.BackupWindow *out = new(string) @@ -6585,6 +6792,11 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(v1.Selector) (*in).DeepCopyInto(*out) } + if in.DedicatedLogVolume != nil { + in, out := &in.DedicatedLogVolume, &out.DedicatedLogVolume + *out = new(bool) + **out = **in + } if in.DeleteAutomatedBackups != nil { in, out := &in.DeleteAutomatedBackups, &out.DeleteAutomatedBackups *out = new(bool) @@ -6600,11 +6812,37 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(string) **out = **in } + if in.DomainAuthSecretArn != nil { + in, out := &in.DomainAuthSecretArn, &out.DomainAuthSecretArn + *out = new(string) + **out = **in + } + if in.DomainDNSIps != nil { + in, out := &in.DomainDNSIps, &out.DomainDNSIps + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.DomainFqdn != nil { + in, out := &in.DomainFqdn, &out.DomainFqdn + *out = new(string) + **out = **in + } if in.DomainIAMRoleName != nil { in, out := &in.DomainIAMRoleName, &out.DomainIAMRoleName *out = new(string) **out = **in } + if in.DomainOu != nil { + in, out := &in.DomainOu, &out.DomainOu + *out = new(string) + **out = **in + } if in.EnabledCloudwatchLogsExports != nil { in, out := &in.EnabledCloudwatchLogsExports, &out.EnabledCloudwatchLogsExports *out = make([]*string, len(*in)) @@ -6621,6 +6859,11 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) @@ -6636,6 +6879,16 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(bool) **out = **in } + if in.Identifier != nil { + in, out := &in.Identifier, &out.Identifier + *out = new(string) + **out = **in + } + if in.IdentifierPrefix != nil { + in, out := &in.IdentifierPrefix, &out.IdentifierPrefix + *out = new(string) + **out = **in + } if in.InstanceClass != nil { in, out := &in.InstanceClass, &out.InstanceClass *out = new(string) @@ -6746,6 +6999,16 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(string) **out = **in } + if in.ParameterGroupNameRef != nil { + in, out := &in.ParameterGroupNameRef, &out.ParameterGroupNameRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.ParameterGroupNameSelector != nil { + in, out := &in.ParameterGroupNameSelector, &out.ParameterGroupNameSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } if in.PasswordSecretRef != nil { in, out := &in.PasswordSecretRef, &out.PasswordSecretRef *out = new(v1.SecretKeySelector) @@ -6872,6 +7135,11 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(string) **out = **in } + if in.UpgradeStorageConfig != nil { + in, out := &in.UpgradeStorageConfig, &out.UpgradeStorageConfig + *out = new(bool) + **out = **in + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -9765,6 +10033,11 @@ func (in *ScalingConfigurationInitParameters) DeepCopyInto(out *ScalingConfigura *out = new(float64) **out = **in } + if in.SecondsBeforeTimeout != nil { + in, out := &in.SecondsBeforeTimeout, &out.SecondsBeforeTimeout + *out = new(float64) + **out = **in + } if in.SecondsUntilAutoPause != nil { in, out := &in.SecondsUntilAutoPause, &out.SecondsUntilAutoPause *out = new(float64) @@ -9805,6 +10078,11 @@ func (in *ScalingConfigurationObservation) DeepCopyInto(out *ScalingConfiguratio *out = new(float64) **out = **in } + if in.SecondsBeforeTimeout != nil { + in, out := &in.SecondsBeforeTimeout, &out.SecondsBeforeTimeout + *out = new(float64) + **out = **in + } if in.SecondsUntilAutoPause != nil { in, out := &in.SecondsUntilAutoPause, &out.SecondsUntilAutoPause *out = new(float64) @@ -9845,6 +10123,11 @@ func (in *ScalingConfigurationParameters) DeepCopyInto(out *ScalingConfiguration *out = new(float64) **out = **in } + if in.SecondsBeforeTimeout != nil { + in, out := &in.SecondsBeforeTimeout, &out.SecondsBeforeTimeout + *out = new(float64) + **out = **in + } if in.SecondsUntilAutoPause != nil { in, out := &in.SecondsUntilAutoPause, &out.SecondsUntilAutoPause *out = new(float64) diff --git a/apis/rds/v1beta2/zz_generated.deepcopy.go b/apis/rds/v1beta2/zz_generated.deepcopy.go index 0a98d03e37..2d6ee6141d 100644 --- a/apis/rds/v1beta2/zz_generated.deepcopy.go +++ b/apis/rds/v1beta2/zz_generated.deepcopy.go @@ -2012,6 +2012,11 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) @@ -2262,6 +2267,11 @@ func (in *InstanceInitParameters) DeepCopyInto(out *InstanceInitParameters) { *out = new(string) **out = **in } + if in.UpgradeStorageConfig != nil { + in, out := &in.UpgradeStorageConfig, &out.UpgradeStorageConfig + *out = new(bool) + **out = **in + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -2501,6 +2511,11 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) @@ -2767,6 +2782,11 @@ func (in *InstanceObservation) DeepCopyInto(out *InstanceObservation) { *out = new(string) **out = **in } + if in.UpgradeStorageConfig != nil { + in, out := &in.UpgradeStorageConfig, &out.UpgradeStorageConfig + *out = new(bool) + **out = **in + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) @@ -2962,6 +2982,11 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(string) **out = **in } + if in.EngineLifecycleSupport != nil { + in, out := &in.EngineLifecycleSupport, &out.EngineLifecycleSupport + *out = new(string) + **out = **in + } if in.EngineVersion != nil { in, out := &in.EngineVersion, &out.EngineVersion *out = new(string) @@ -3217,6 +3242,11 @@ func (in *InstanceParameters) DeepCopyInto(out *InstanceParameters) { *out = new(string) **out = **in } + if in.UpgradeStorageConfig != nil { + in, out := &in.UpgradeStorageConfig, &out.UpgradeStorageConfig + *out = new(bool) + **out = **in + } if in.Username != nil { in, out := &in.Username, &out.Username *out = new(string) diff --git a/apis/sagemaker/v1beta1/zz_generated.deepcopy.go b/apis/sagemaker/v1beta1/zz_generated.deepcopy.go index a8f85aac7c..05bae68555 100644 --- a/apis/sagemaker/v1beta1/zz_generated.deepcopy.go +++ b/apis/sagemaker/v1beta1/zz_generated.deepcopy.go @@ -3171,6 +3171,11 @@ func (in *ContainerInitParameters) DeepCopyInto(out *ContainerInitParameters) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InferenceSpecificationName != nil { + in, out := &in.InferenceSpecificationName, &out.InferenceSpecificationName + *out = new(string) + **out = **in + } if in.Mode != nil { in, out := &in.Mode, &out.Mode *out = new(string) @@ -3193,6 +3198,13 @@ func (in *ContainerInitParameters) DeepCopyInto(out *ContainerInitParameters) { *out = new(string) **out = **in } + if in.MultiModelConfig != nil { + in, out := &in.MultiModelConfig, &out.MultiModelConfig + *out = make([]MultiModelConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerInitParameters. @@ -3241,6 +3253,11 @@ func (in *ContainerObservation) DeepCopyInto(out *ContainerObservation) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InferenceSpecificationName != nil { + in, out := &in.InferenceSpecificationName, &out.InferenceSpecificationName + *out = new(string) + **out = **in + } if in.Mode != nil { in, out := &in.Mode, &out.Mode *out = new(string) @@ -3263,6 +3280,13 @@ func (in *ContainerObservation) DeepCopyInto(out *ContainerObservation) { *out = new(string) **out = **in } + if in.MultiModelConfig != nil { + in, out := &in.MultiModelConfig, &out.MultiModelConfig + *out = make([]MultiModelConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerObservation. @@ -3311,6 +3335,11 @@ func (in *ContainerParameters) DeepCopyInto(out *ContainerParameters) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InferenceSpecificationName != nil { + in, out := &in.InferenceSpecificationName, &out.InferenceSpecificationName + *out = new(string) + **out = **in + } if in.Mode != nil { in, out := &in.Mode, &out.Mode *out = new(string) @@ -3333,6 +3362,13 @@ func (in *ContainerParameters) DeepCopyInto(out *ContainerParameters) { *out = new(string) **out = **in } + if in.MultiModelConfig != nil { + in, out := &in.MultiModelConfig, &out.MultiModelConfig + *out = make([]MultiModelConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerParameters. @@ -4227,11 +4263,32 @@ func (in *DefaultResourceSpecParameters) DeepCopy() *DefaultResourceSpecParamete // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DefaultSpaceSettingsInitParameters) DeepCopyInto(out *DefaultSpaceSettingsInitParameters) { *out = *in + if in.CustomFileSystemConfig != nil { + in, out := &in.CustomFileSystemConfig, &out.CustomFileSystemConfig + *out = make([]CustomFileSystemConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.CustomPosixUserConfig != nil { + in, out := &in.CustomPosixUserConfig, &out.CustomPosixUserConfig + *out = make([]CustomPosixUserConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ExecutionRole != nil { in, out := &in.ExecutionRole, &out.ExecutionRole *out = new(string) **out = **in } + if in.JupyterLabAppSettings != nil { + in, out := &in.JupyterLabAppSettings, &out.JupyterLabAppSettings + *out = make([]JupyterLabAppSettingsInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.JupyterServerAppSettings != nil { in, out := &in.JupyterServerAppSettings, &out.JupyterServerAppSettings *out = make([]JupyterServerAppSettingsInitParameters, len(*in)) @@ -4257,6 +4314,13 @@ func (in *DefaultSpaceSettingsInitParameters) DeepCopyInto(out *DefaultSpaceSett } } } + if in.SpaceStorageSettings != nil { + in, out := &in.SpaceStorageSettings, &out.SpaceStorageSettings + *out = make([]SpaceStorageSettingsInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DefaultSpaceSettingsInitParameters. @@ -4272,11 +4336,32 @@ func (in *DefaultSpaceSettingsInitParameters) DeepCopy() *DefaultSpaceSettingsIn // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DefaultSpaceSettingsObservation) DeepCopyInto(out *DefaultSpaceSettingsObservation) { *out = *in + if in.CustomFileSystemConfig != nil { + in, out := &in.CustomFileSystemConfig, &out.CustomFileSystemConfig + *out = make([]CustomFileSystemConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.CustomPosixUserConfig != nil { + in, out := &in.CustomPosixUserConfig, &out.CustomPosixUserConfig + *out = make([]CustomPosixUserConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ExecutionRole != nil { in, out := &in.ExecutionRole, &out.ExecutionRole *out = new(string) **out = **in } + if in.JupyterLabAppSettings != nil { + in, out := &in.JupyterLabAppSettings, &out.JupyterLabAppSettings + *out = make([]JupyterLabAppSettingsObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.JupyterServerAppSettings != nil { in, out := &in.JupyterServerAppSettings, &out.JupyterServerAppSettings *out = make([]JupyterServerAppSettingsObservation, len(*in)) @@ -4302,6 +4387,13 @@ func (in *DefaultSpaceSettingsObservation) DeepCopyInto(out *DefaultSpaceSetting } } } + if in.SpaceStorageSettings != nil { + in, out := &in.SpaceStorageSettings, &out.SpaceStorageSettings + *out = make([]SpaceStorageSettingsObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DefaultSpaceSettingsObservation. @@ -4317,11 +4409,32 @@ func (in *DefaultSpaceSettingsObservation) DeepCopy() *DefaultSpaceSettingsObser // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DefaultSpaceSettingsParameters) DeepCopyInto(out *DefaultSpaceSettingsParameters) { *out = *in + if in.CustomFileSystemConfig != nil { + in, out := &in.CustomFileSystemConfig, &out.CustomFileSystemConfig + *out = make([]CustomFileSystemConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.CustomPosixUserConfig != nil { + in, out := &in.CustomPosixUserConfig, &out.CustomPosixUserConfig + *out = make([]CustomPosixUserConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ExecutionRole != nil { in, out := &in.ExecutionRole, &out.ExecutionRole *out = new(string) **out = **in } + if in.JupyterLabAppSettings != nil { + in, out := &in.JupyterLabAppSettings, &out.JupyterLabAppSettings + *out = make([]JupyterLabAppSettingsParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.JupyterServerAppSettings != nil { in, out := &in.JupyterServerAppSettings, &out.JupyterServerAppSettings *out = make([]JupyterServerAppSettingsParameters, len(*in)) @@ -4347,6 +4460,13 @@ func (in *DefaultSpaceSettingsParameters) DeepCopyInto(out *DefaultSpaceSettings } } } + if in.SpaceStorageSettings != nil { + in, out := &in.SpaceStorageSettings, &out.SpaceStorageSettings + *out = make([]SpaceStorageSettingsParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DefaultSpaceSettingsParameters. @@ -4475,6 +4595,13 @@ func (in *DefaultUserSettingsInitParameters) DeepCopyInto(out *DefaultUserSettin *out = new(string) **out = **in } + if in.StudioWebPortalSettings != nil { + in, out := &in.StudioWebPortalSettings, &out.StudioWebPortalSettings + *out = make([]StudioWebPortalSettingsInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.TensorBoardAppSettings != nil { in, out := &in.TensorBoardAppSettings, &out.TensorBoardAppSettings *out = make([]TensorBoardAppSettingsInitParameters, len(*in)) @@ -5020,6 +5147,13 @@ func (in *DefaultUserSettingsObservation) DeepCopyInto(out *DefaultUserSettingsO *out = new(string) **out = **in } + if in.StudioWebPortalSettings != nil { + in, out := &in.StudioWebPortalSettings, &out.StudioWebPortalSettings + *out = make([]StudioWebPortalSettingsObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.TensorBoardAppSettings != nil { in, out := &in.TensorBoardAppSettings, &out.TensorBoardAppSettings *out = make([]TensorBoardAppSettingsObservation, len(*in)) @@ -5155,6 +5289,13 @@ func (in *DefaultUserSettingsParameters) DeepCopyInto(out *DefaultUserSettingsPa *out = new(string) **out = **in } + if in.StudioWebPortalSettings != nil { + in, out := &in.StudioWebPortalSettings, &out.StudioWebPortalSettings + *out = make([]StudioWebPortalSettingsParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.TensorBoardAppSettings != nil { in, out := &in.TensorBoardAppSettings, &out.TensorBoardAppSettings *out = make([]TensorBoardAppSettingsParameters, len(*in)) @@ -5956,6 +6097,99 @@ func (in *DirectDeploySettingsParameters) DeepCopy() *DirectDeploySettingsParame return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DockerSettingsInitParameters) DeepCopyInto(out *DockerSettingsInitParameters) { + *out = *in + if in.EnableDockerAccess != nil { + in, out := &in.EnableDockerAccess, &out.EnableDockerAccess + *out = new(string) + **out = **in + } + if in.VPCOnlyTrustedAccounts != nil { + in, out := &in.VPCOnlyTrustedAccounts, &out.VPCOnlyTrustedAccounts + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DockerSettingsInitParameters. +func (in *DockerSettingsInitParameters) DeepCopy() *DockerSettingsInitParameters { + if in == nil { + return nil + } + out := new(DockerSettingsInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DockerSettingsObservation) DeepCopyInto(out *DockerSettingsObservation) { + *out = *in + if in.EnableDockerAccess != nil { + in, out := &in.EnableDockerAccess, &out.EnableDockerAccess + *out = new(string) + **out = **in + } + if in.VPCOnlyTrustedAccounts != nil { + in, out := &in.VPCOnlyTrustedAccounts, &out.VPCOnlyTrustedAccounts + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DockerSettingsObservation. +func (in *DockerSettingsObservation) DeepCopy() *DockerSettingsObservation { + if in == nil { + return nil + } + out := new(DockerSettingsObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *DockerSettingsParameters) DeepCopyInto(out *DockerSettingsParameters) { + *out = *in + if in.EnableDockerAccess != nil { + in, out := &in.EnableDockerAccess, &out.EnableDockerAccess + *out = new(string) + **out = **in + } + if in.VPCOnlyTrustedAccounts != nil { + in, out := &in.VPCOnlyTrustedAccounts, &out.VPCOnlyTrustedAccounts + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DockerSettingsParameters. +func (in *DockerSettingsParameters) DeepCopy() *DockerSettingsParameters { + if in == nil { + return nil + } + out := new(DockerSettingsParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *Domain) DeepCopyInto(out *Domain) { *out = *in @@ -6438,6 +6672,13 @@ func (in *DomainParameters) DeepCopy() *DomainParameters { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DomainSettingsInitParameters) DeepCopyInto(out *DomainSettingsInitParameters) { *out = *in + if in.DockerSettings != nil { + in, out := &in.DockerSettings, &out.DockerSettings + *out = make([]DockerSettingsInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ExecutionRoleIdentityConfig != nil { in, out := &in.ExecutionRoleIdentityConfig, &out.ExecutionRoleIdentityConfig *out = new(string) @@ -6476,6 +6717,13 @@ func (in *DomainSettingsInitParameters) DeepCopy() *DomainSettingsInitParameters // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DomainSettingsObservation) DeepCopyInto(out *DomainSettingsObservation) { *out = *in + if in.DockerSettings != nil { + in, out := &in.DockerSettings, &out.DockerSettings + *out = make([]DockerSettingsObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ExecutionRoleIdentityConfig != nil { in, out := &in.ExecutionRoleIdentityConfig, &out.ExecutionRoleIdentityConfig *out = new(string) @@ -6514,6 +6762,13 @@ func (in *DomainSettingsObservation) DeepCopy() *DomainSettingsObservation { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DomainSettingsParameters) DeepCopyInto(out *DomainSettingsParameters) { *out = *in + if in.DockerSettings != nil { + in, out := &in.DockerSettings, &out.DockerSettings + *out = make([]DockerSettingsParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ExecutionRoleIdentityConfig != nil { in, out := &in.ExecutionRoleIdentityConfig, &out.ExecutionRoleIdentityConfig *out = new(string) @@ -10819,57 +11074,97 @@ func (in *LinearStepSizeParameters) DeepCopy() *LinearStepSizeParameters { } // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *MaximumBatchSizeInitParameters) DeepCopyInto(out *MaximumBatchSizeInitParameters) { +func (in *ManagedInstanceScalingInitParameters) DeepCopyInto(out *ManagedInstanceScalingInitParameters) { *out = *in - if in.Type != nil { - in, out := &in.Type, &out.Type - *out = new(string) + if in.MaxInstanceCount != nil { + in, out := &in.MaxInstanceCount, &out.MaxInstanceCount + *out = new(float64) **out = **in } - if in.Value != nil { - in, out := &in.Value, &out.Value + if in.MinInstanceCount != nil { + in, out := &in.MinInstanceCount, &out.MinInstanceCount *out = new(float64) **out = **in } + if in.Status != nil { + in, out := &in.Status, &out.Status + *out = new(string) + **out = **in + } } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MaximumBatchSizeInitParameters. -func (in *MaximumBatchSizeInitParameters) DeepCopy() *MaximumBatchSizeInitParameters { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ManagedInstanceScalingInitParameters. +func (in *ManagedInstanceScalingInitParameters) DeepCopy() *ManagedInstanceScalingInitParameters { if in == nil { return nil } - out := new(MaximumBatchSizeInitParameters) + out := new(ManagedInstanceScalingInitParameters) in.DeepCopyInto(out) return out } // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *MaximumBatchSizeObservation) DeepCopyInto(out *MaximumBatchSizeObservation) { +func (in *ManagedInstanceScalingObservation) DeepCopyInto(out *ManagedInstanceScalingObservation) { *out = *in - if in.Type != nil { - in, out := &in.Type, &out.Type + if in.MaxInstanceCount != nil { + in, out := &in.MaxInstanceCount, &out.MaxInstanceCount + *out = new(float64) + **out = **in + } + if in.MinInstanceCount != nil { + in, out := &in.MinInstanceCount, &out.MinInstanceCount + *out = new(float64) + **out = **in + } + if in.Status != nil { + in, out := &in.Status, &out.Status *out = new(string) **out = **in } - if in.Value != nil { - in, out := &in.Value, &out.Value +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ManagedInstanceScalingObservation. +func (in *ManagedInstanceScalingObservation) DeepCopy() *ManagedInstanceScalingObservation { + if in == nil { + return nil + } + out := new(ManagedInstanceScalingObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ManagedInstanceScalingParameters) DeepCopyInto(out *ManagedInstanceScalingParameters) { + *out = *in + if in.MaxInstanceCount != nil { + in, out := &in.MaxInstanceCount, &out.MaxInstanceCount + *out = new(float64) + **out = **in + } + if in.MinInstanceCount != nil { + in, out := &in.MinInstanceCount, &out.MinInstanceCount *out = new(float64) **out = **in } + if in.Status != nil { + in, out := &in.Status, &out.Status + *out = new(string) + **out = **in + } } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MaximumBatchSizeObservation. -func (in *MaximumBatchSizeObservation) DeepCopy() *MaximumBatchSizeObservation { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ManagedInstanceScalingParameters. +func (in *ManagedInstanceScalingParameters) DeepCopy() *ManagedInstanceScalingParameters { if in == nil { return nil } - out := new(MaximumBatchSizeObservation) + out := new(ManagedInstanceScalingParameters) in.DeepCopyInto(out) return out } // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *MaximumBatchSizeParameters) DeepCopyInto(out *MaximumBatchSizeParameters) { +func (in *MaximumBatchSizeInitParameters) DeepCopyInto(out *MaximumBatchSizeInitParameters) { *out = *in if in.Type != nil { in, out := &in.Type, &out.Type @@ -10883,12 +11178,62 @@ func (in *MaximumBatchSizeParameters) DeepCopyInto(out *MaximumBatchSizeParamete } } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MaximumBatchSizeParameters. -func (in *MaximumBatchSizeParameters) DeepCopy() *MaximumBatchSizeParameters { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MaximumBatchSizeInitParameters. +func (in *MaximumBatchSizeInitParameters) DeepCopy() *MaximumBatchSizeInitParameters { if in == nil { return nil } - out := new(MaximumBatchSizeParameters) + out := new(MaximumBatchSizeInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MaximumBatchSizeObservation) DeepCopyInto(out *MaximumBatchSizeObservation) { + *out = *in + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } + if in.Value != nil { + in, out := &in.Value, &out.Value + *out = new(float64) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MaximumBatchSizeObservation. +func (in *MaximumBatchSizeObservation) DeepCopy() *MaximumBatchSizeObservation { + if in == nil { + return nil + } + out := new(MaximumBatchSizeObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MaximumBatchSizeParameters) DeepCopyInto(out *MaximumBatchSizeParameters) { + *out = *in + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } + if in.Value != nil { + in, out := &in.Value, &out.Value + *out = new(float64) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MaximumBatchSizeParameters. +func (in *MaximumBatchSizeParameters) DeepCopy() *MaximumBatchSizeParameters { + if in == nil { + return nil + } + out := new(MaximumBatchSizeParameters) in.DeepCopyInto(out) return out } @@ -11007,6 +11352,66 @@ func (in *Model) DeepCopyObject() runtime.Object { return nil } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ModelAccessConfigInitParameters) DeepCopyInto(out *ModelAccessConfigInitParameters) { + *out = *in + if in.AcceptEula != nil { + in, out := &in.AcceptEula, &out.AcceptEula + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ModelAccessConfigInitParameters. +func (in *ModelAccessConfigInitParameters) DeepCopy() *ModelAccessConfigInitParameters { + if in == nil { + return nil + } + out := new(ModelAccessConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ModelAccessConfigObservation) DeepCopyInto(out *ModelAccessConfigObservation) { + *out = *in + if in.AcceptEula != nil { + in, out := &in.AcceptEula, &out.AcceptEula + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ModelAccessConfigObservation. +func (in *ModelAccessConfigObservation) DeepCopy() *ModelAccessConfigObservation { + if in == nil { + return nil + } + out := new(ModelAccessConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ModelAccessConfigParameters) DeepCopyInto(out *ModelAccessConfigParameters) { + *out = *in + if in.AcceptEula != nil { + in, out := &in.AcceptEula, &out.AcceptEula + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ModelAccessConfigParameters. +func (in *ModelAccessConfigParameters) DeepCopy() *ModelAccessConfigParameters { + if in == nil { + return nil + } + out := new(ModelAccessConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *ModelDataSourceInitParameters) DeepCopyInto(out *ModelDataSourceInitParameters) { *out = *in @@ -11081,6 +11486,13 @@ func (in *ModelDataSourceS3DataSourceInitParameters) DeepCopyInto(out *ModelData *out = new(string) **out = **in } + if in.ModelAccessConfig != nil { + in, out := &in.ModelAccessConfig, &out.ModelAccessConfig + *out = make([]S3DataSourceModelAccessConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.S3DataType != nil { in, out := &in.S3DataType, &out.S3DataType *out = new(string) @@ -11111,6 +11523,13 @@ func (in *ModelDataSourceS3DataSourceObservation) DeepCopyInto(out *ModelDataSou *out = new(string) **out = **in } + if in.ModelAccessConfig != nil { + in, out := &in.ModelAccessConfig, &out.ModelAccessConfig + *out = make([]S3DataSourceModelAccessConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.S3DataType != nil { in, out := &in.S3DataType, &out.S3DataType *out = new(string) @@ -11141,6 +11560,13 @@ func (in *ModelDataSourceS3DataSourceParameters) DeepCopyInto(out *ModelDataSour *out = new(string) **out = **in } + if in.ModelAccessConfig != nil { + in, out := &in.ModelAccessConfig, &out.ModelAccessConfig + *out = make([]S3DataSourceModelAccessConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.S3DataType != nil { in, out := &in.S3DataType, &out.S3DataType *out = new(string) @@ -11995,6 +12421,66 @@ func (in *ModelStatus) DeepCopy() *ModelStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MultiModelConfigInitParameters) DeepCopyInto(out *MultiModelConfigInitParameters) { + *out = *in + if in.ModelCacheSetting != nil { + in, out := &in.ModelCacheSetting, &out.ModelCacheSetting + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MultiModelConfigInitParameters. +func (in *MultiModelConfigInitParameters) DeepCopy() *MultiModelConfigInitParameters { + if in == nil { + return nil + } + out := new(MultiModelConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MultiModelConfigObservation) DeepCopyInto(out *MultiModelConfigObservation) { + *out = *in + if in.ModelCacheSetting != nil { + in, out := &in.ModelCacheSetting, &out.ModelCacheSetting + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MultiModelConfigObservation. +func (in *MultiModelConfigObservation) DeepCopy() *MultiModelConfigObservation { + if in == nil { + return nil + } + out := new(MultiModelConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *MultiModelConfigParameters) DeepCopyInto(out *MultiModelConfigParameters) { + *out = *in + if in.ModelCacheSetting != nil { + in, out := &in.ModelCacheSetting, &out.ModelCacheSetting + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MultiModelConfigParameters. +func (in *MultiModelConfigParameters) DeepCopy() *MultiModelConfigParameters { + if in == nil { + return nil + } + out := new(MultiModelConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *NotebookInstance) DeepCopyInto(out *NotebookInstance) { *out = *in @@ -13704,6 +14190,11 @@ func (in *PrimaryContainerInitParameters) DeepCopyInto(out *PrimaryContainerInit (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InferenceSpecificationName != nil { + in, out := &in.InferenceSpecificationName, &out.InferenceSpecificationName + *out = new(string) + **out = **in + } if in.Mode != nil { in, out := &in.Mode, &out.Mode *out = new(string) @@ -13726,6 +14217,13 @@ func (in *PrimaryContainerInitParameters) DeepCopyInto(out *PrimaryContainerInit *out = new(string) **out = **in } + if in.MultiModelConfig != nil { + in, out := &in.MultiModelConfig, &out.MultiModelConfig + *out = make([]PrimaryContainerMultiModelConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PrimaryContainerInitParameters. @@ -13804,6 +14302,66 @@ func (in *PrimaryContainerModelDataSourceParameters) DeepCopy() *PrimaryContaine return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *PrimaryContainerMultiModelConfigInitParameters) DeepCopyInto(out *PrimaryContainerMultiModelConfigInitParameters) { + *out = *in + if in.ModelCacheSetting != nil { + in, out := &in.ModelCacheSetting, &out.ModelCacheSetting + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PrimaryContainerMultiModelConfigInitParameters. +func (in *PrimaryContainerMultiModelConfigInitParameters) DeepCopy() *PrimaryContainerMultiModelConfigInitParameters { + if in == nil { + return nil + } + out := new(PrimaryContainerMultiModelConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *PrimaryContainerMultiModelConfigObservation) DeepCopyInto(out *PrimaryContainerMultiModelConfigObservation) { + *out = *in + if in.ModelCacheSetting != nil { + in, out := &in.ModelCacheSetting, &out.ModelCacheSetting + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PrimaryContainerMultiModelConfigObservation. +func (in *PrimaryContainerMultiModelConfigObservation) DeepCopy() *PrimaryContainerMultiModelConfigObservation { + if in == nil { + return nil + } + out := new(PrimaryContainerMultiModelConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *PrimaryContainerMultiModelConfigParameters) DeepCopyInto(out *PrimaryContainerMultiModelConfigParameters) { + *out = *in + if in.ModelCacheSetting != nil { + in, out := &in.ModelCacheSetting, &out.ModelCacheSetting + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PrimaryContainerMultiModelConfigParameters. +func (in *PrimaryContainerMultiModelConfigParameters) DeepCopy() *PrimaryContainerMultiModelConfigParameters { + if in == nil { + return nil + } + out := new(PrimaryContainerMultiModelConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *PrimaryContainerObservation) DeepCopyInto(out *PrimaryContainerObservation) { *out = *in @@ -13840,6 +14398,11 @@ func (in *PrimaryContainerObservation) DeepCopyInto(out *PrimaryContainerObserva (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InferenceSpecificationName != nil { + in, out := &in.InferenceSpecificationName, &out.InferenceSpecificationName + *out = new(string) + **out = **in + } if in.Mode != nil { in, out := &in.Mode, &out.Mode *out = new(string) @@ -13862,6 +14425,13 @@ func (in *PrimaryContainerObservation) DeepCopyInto(out *PrimaryContainerObserva *out = new(string) **out = **in } + if in.MultiModelConfig != nil { + in, out := &in.MultiModelConfig, &out.MultiModelConfig + *out = make([]PrimaryContainerMultiModelConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PrimaryContainerObservation. @@ -13910,6 +14480,11 @@ func (in *PrimaryContainerParameters) DeepCopyInto(out *PrimaryContainerParamete (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.InferenceSpecificationName != nil { + in, out := &in.InferenceSpecificationName, &out.InferenceSpecificationName + *out = new(string) + **out = **in + } if in.Mode != nil { in, out := &in.Mode, &out.Mode *out = new(string) @@ -13932,6 +14507,13 @@ func (in *PrimaryContainerParameters) DeepCopyInto(out *PrimaryContainerParamete *out = new(string) **out = **in } + if in.MultiModelConfig != nil { + in, out := &in.MultiModelConfig, &out.MultiModelConfig + *out = make([]PrimaryContainerMultiModelConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } } // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PrimaryContainerParameters. @@ -13989,6 +14571,13 @@ func (in *ProductionVariantsInitParameters) DeepCopyInto(out *ProductionVariants *out = new(string) **out = **in } + if in.ManagedInstanceScaling != nil { + in, out := &in.ManagedInstanceScaling, &out.ManagedInstanceScaling + *out = make([]ManagedInstanceScalingInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ModelDataDownloadTimeoutInSeconds != nil { in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds *out = new(float64) @@ -14090,6 +14679,13 @@ func (in *ProductionVariantsObservation) DeepCopyInto(out *ProductionVariantsObs *out = new(string) **out = **in } + if in.ManagedInstanceScaling != nil { + in, out := &in.ManagedInstanceScaling, &out.ManagedInstanceScaling + *out = make([]ManagedInstanceScalingObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ModelDataDownloadTimeoutInSeconds != nil { in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds *out = new(float64) @@ -14181,6 +14777,13 @@ func (in *ProductionVariantsParameters) DeepCopyInto(out *ProductionVariantsPara *out = new(string) **out = **in } + if in.ManagedInstanceScaling != nil { + in, out := &in.ManagedInstanceScaling, &out.ManagedInstanceScaling + *out = make([]ManagedInstanceScalingParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ModelDataDownloadTimeoutInSeconds != nil { in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds *out = new(float64) @@ -15340,6 +15943,13 @@ func (in *S3DataSourceInitParameters) DeepCopyInto(out *S3DataSourceInitParamete *out = new(string) **out = **in } + if in.ModelAccessConfig != nil { + in, out := &in.ModelAccessConfig, &out.ModelAccessConfig + *out = make([]ModelAccessConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.S3DataType != nil { in, out := &in.S3DataType, &out.S3DataType *out = new(string) @@ -15362,6 +15972,66 @@ func (in *S3DataSourceInitParameters) DeepCopy() *S3DataSourceInitParameters { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *S3DataSourceModelAccessConfigInitParameters) DeepCopyInto(out *S3DataSourceModelAccessConfigInitParameters) { + *out = *in + if in.AcceptEula != nil { + in, out := &in.AcceptEula, &out.AcceptEula + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new S3DataSourceModelAccessConfigInitParameters. +func (in *S3DataSourceModelAccessConfigInitParameters) DeepCopy() *S3DataSourceModelAccessConfigInitParameters { + if in == nil { + return nil + } + out := new(S3DataSourceModelAccessConfigInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *S3DataSourceModelAccessConfigObservation) DeepCopyInto(out *S3DataSourceModelAccessConfigObservation) { + *out = *in + if in.AcceptEula != nil { + in, out := &in.AcceptEula, &out.AcceptEula + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new S3DataSourceModelAccessConfigObservation. +func (in *S3DataSourceModelAccessConfigObservation) DeepCopy() *S3DataSourceModelAccessConfigObservation { + if in == nil { + return nil + } + out := new(S3DataSourceModelAccessConfigObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *S3DataSourceModelAccessConfigParameters) DeepCopyInto(out *S3DataSourceModelAccessConfigParameters) { + *out = *in + if in.AcceptEula != nil { + in, out := &in.AcceptEula, &out.AcceptEula + *out = new(bool) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new S3DataSourceModelAccessConfigParameters. +func (in *S3DataSourceModelAccessConfigParameters) DeepCopy() *S3DataSourceModelAccessConfigParameters { + if in == nil { + return nil + } + out := new(S3DataSourceModelAccessConfigParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *S3DataSourceObservation) DeepCopyInto(out *S3DataSourceObservation) { *out = *in @@ -15370,6 +16040,13 @@ func (in *S3DataSourceObservation) DeepCopyInto(out *S3DataSourceObservation) { *out = new(string) **out = **in } + if in.ModelAccessConfig != nil { + in, out := &in.ModelAccessConfig, &out.ModelAccessConfig + *out = make([]ModelAccessConfigObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.S3DataType != nil { in, out := &in.S3DataType, &out.S3DataType *out = new(string) @@ -15400,6 +16077,13 @@ func (in *S3DataSourceParameters) DeepCopyInto(out *S3DataSourceParameters) { *out = new(string) **out = **in } + if in.ModelAccessConfig != nil { + in, out := &in.ModelAccessConfig, &out.ModelAccessConfig + *out = make([]ModelAccessConfigParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.S3DataType != nil { in, out := &in.S3DataType, &out.S3DataType *out = new(string) @@ -15991,63 +16675,160 @@ func (in *ShadowProductionVariantsInitParameters) DeepCopyInto(out *ShadowProduc *out = new(string) **out = **in } - if in.InitialInstanceCount != nil { - in, out := &in.InitialInstanceCount, &out.InitialInstanceCount + if in.InitialInstanceCount != nil { + in, out := &in.InitialInstanceCount, &out.InitialInstanceCount + *out = new(float64) + **out = **in + } + if in.InitialVariantWeight != nil { + in, out := &in.InitialVariantWeight, &out.InitialVariantWeight + *out = new(float64) + **out = **in + } + if in.InstanceType != nil { + in, out := &in.InstanceType, &out.InstanceType + *out = new(string) + **out = **in + } + if in.ManagedInstanceScaling != nil { + in, out := &in.ManagedInstanceScaling, &out.ManagedInstanceScaling + *out = make([]ShadowProductionVariantsManagedInstanceScalingInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.ModelDataDownloadTimeoutInSeconds != nil { + in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds + *out = new(float64) + **out = **in + } + if in.ModelName != nil { + in, out := &in.ModelName, &out.ModelName + *out = new(string) + **out = **in + } + if in.RoutingConfig != nil { + in, out := &in.RoutingConfig, &out.RoutingConfig + *out = make([]ShadowProductionVariantsRoutingConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.ServerlessConfig != nil { + in, out := &in.ServerlessConfig, &out.ServerlessConfig + *out = make([]ShadowProductionVariantsServerlessConfigInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.VariantName != nil { + in, out := &in.VariantName, &out.VariantName + *out = new(string) + **out = **in + } + if in.VolumeSizeInGb != nil { + in, out := &in.VolumeSizeInGb, &out.VolumeSizeInGb + *out = new(float64) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ShadowProductionVariantsInitParameters. +func (in *ShadowProductionVariantsInitParameters) DeepCopy() *ShadowProductionVariantsInitParameters { + if in == nil { + return nil + } + out := new(ShadowProductionVariantsInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ShadowProductionVariantsManagedInstanceScalingInitParameters) DeepCopyInto(out *ShadowProductionVariantsManagedInstanceScalingInitParameters) { + *out = *in + if in.MaxInstanceCount != nil { + in, out := &in.MaxInstanceCount, &out.MaxInstanceCount + *out = new(float64) + **out = **in + } + if in.MinInstanceCount != nil { + in, out := &in.MinInstanceCount, &out.MinInstanceCount + *out = new(float64) + **out = **in + } + if in.Status != nil { + in, out := &in.Status, &out.Status + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ShadowProductionVariantsManagedInstanceScalingInitParameters. +func (in *ShadowProductionVariantsManagedInstanceScalingInitParameters) DeepCopy() *ShadowProductionVariantsManagedInstanceScalingInitParameters { + if in == nil { + return nil + } + out := new(ShadowProductionVariantsManagedInstanceScalingInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ShadowProductionVariantsManagedInstanceScalingObservation) DeepCopyInto(out *ShadowProductionVariantsManagedInstanceScalingObservation) { + *out = *in + if in.MaxInstanceCount != nil { + in, out := &in.MaxInstanceCount, &out.MaxInstanceCount *out = new(float64) **out = **in } - if in.InitialVariantWeight != nil { - in, out := &in.InitialVariantWeight, &out.InitialVariantWeight + if in.MinInstanceCount != nil { + in, out := &in.MinInstanceCount, &out.MinInstanceCount *out = new(float64) **out = **in } - if in.InstanceType != nil { - in, out := &in.InstanceType, &out.InstanceType + if in.Status != nil { + in, out := &in.Status, &out.Status *out = new(string) **out = **in } - if in.ModelDataDownloadTimeoutInSeconds != nil { - in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ShadowProductionVariantsManagedInstanceScalingObservation. +func (in *ShadowProductionVariantsManagedInstanceScalingObservation) DeepCopy() *ShadowProductionVariantsManagedInstanceScalingObservation { + if in == nil { + return nil + } + out := new(ShadowProductionVariantsManagedInstanceScalingObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ShadowProductionVariantsManagedInstanceScalingParameters) DeepCopyInto(out *ShadowProductionVariantsManagedInstanceScalingParameters) { + *out = *in + if in.MaxInstanceCount != nil { + in, out := &in.MaxInstanceCount, &out.MaxInstanceCount *out = new(float64) **out = **in } - if in.ModelName != nil { - in, out := &in.ModelName, &out.ModelName - *out = new(string) + if in.MinInstanceCount != nil { + in, out := &in.MinInstanceCount, &out.MinInstanceCount + *out = new(float64) **out = **in } - if in.RoutingConfig != nil { - in, out := &in.RoutingConfig, &out.RoutingConfig - *out = make([]ShadowProductionVariantsRoutingConfigInitParameters, len(*in)) - for i := range *in { - (*in)[i].DeepCopyInto(&(*out)[i]) - } - } - if in.ServerlessConfig != nil { - in, out := &in.ServerlessConfig, &out.ServerlessConfig - *out = make([]ShadowProductionVariantsServerlessConfigInitParameters, len(*in)) - for i := range *in { - (*in)[i].DeepCopyInto(&(*out)[i]) - } - } - if in.VariantName != nil { - in, out := &in.VariantName, &out.VariantName + if in.Status != nil { + in, out := &in.Status, &out.Status *out = new(string) **out = **in } - if in.VolumeSizeInGb != nil { - in, out := &in.VolumeSizeInGb, &out.VolumeSizeInGb - *out = new(float64) - **out = **in - } } -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ShadowProductionVariantsInitParameters. -func (in *ShadowProductionVariantsInitParameters) DeepCopy() *ShadowProductionVariantsInitParameters { +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ShadowProductionVariantsManagedInstanceScalingParameters. +func (in *ShadowProductionVariantsManagedInstanceScalingParameters) DeepCopy() *ShadowProductionVariantsManagedInstanceScalingParameters { if in == nil { return nil } - out := new(ShadowProductionVariantsInitParameters) + out := new(ShadowProductionVariantsManagedInstanceScalingParameters) in.DeepCopyInto(out) return out } @@ -16097,6 +16878,13 @@ func (in *ShadowProductionVariantsObservation) DeepCopyInto(out *ShadowProductio *out = new(string) **out = **in } + if in.ManagedInstanceScaling != nil { + in, out := &in.ManagedInstanceScaling, &out.ManagedInstanceScaling + *out = make([]ShadowProductionVariantsManagedInstanceScalingObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ModelDataDownloadTimeoutInSeconds != nil { in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds *out = new(float64) @@ -16188,6 +16976,13 @@ func (in *ShadowProductionVariantsParameters) DeepCopyInto(out *ShadowProduction *out = new(string) **out = **in } + if in.ManagedInstanceScaling != nil { + in, out := &in.ManagedInstanceScaling, &out.ManagedInstanceScaling + *out = make([]ShadowProductionVariantsManagedInstanceScalingParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ModelDataDownloadTimeoutInSeconds != nil { in, out := &in.ModelDataDownloadTimeoutInSeconds, &out.ModelDataDownloadTimeoutInSeconds *out = new(float64) @@ -18692,6 +19487,117 @@ func (in *StudioLifecycleConfigStatus) DeepCopy() *StudioLifecycleConfigStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *StudioWebPortalSettingsInitParameters) DeepCopyInto(out *StudioWebPortalSettingsInitParameters) { + *out = *in + if in.HiddenAppTypes != nil { + in, out := &in.HiddenAppTypes, &out.HiddenAppTypes + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.HiddenMLTools != nil { + in, out := &in.HiddenMLTools, &out.HiddenMLTools + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StudioWebPortalSettingsInitParameters. +func (in *StudioWebPortalSettingsInitParameters) DeepCopy() *StudioWebPortalSettingsInitParameters { + if in == nil { + return nil + } + out := new(StudioWebPortalSettingsInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *StudioWebPortalSettingsObservation) DeepCopyInto(out *StudioWebPortalSettingsObservation) { + *out = *in + if in.HiddenAppTypes != nil { + in, out := &in.HiddenAppTypes, &out.HiddenAppTypes + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.HiddenMLTools != nil { + in, out := &in.HiddenMLTools, &out.HiddenMLTools + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StudioWebPortalSettingsObservation. +func (in *StudioWebPortalSettingsObservation) DeepCopy() *StudioWebPortalSettingsObservation { + if in == nil { + return nil + } + out := new(StudioWebPortalSettingsObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *StudioWebPortalSettingsParameters) DeepCopyInto(out *StudioWebPortalSettingsParameters) { + *out = *in + if in.HiddenAppTypes != nil { + in, out := &in.HiddenAppTypes, &out.HiddenAppTypes + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.HiddenMLTools != nil { + in, out := &in.HiddenMLTools, &out.HiddenMLTools + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StudioWebPortalSettingsParameters. +func (in *StudioWebPortalSettingsParameters) DeepCopy() *StudioWebPortalSettingsParameters { + if in == nil { + return nil + } + out := new(StudioWebPortalSettingsParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *TTLDurationInitParameters) DeepCopyInto(out *TTLDurationInitParameters) { *out = *in @@ -20232,6 +21138,13 @@ func (in *UserSettingsInitParameters) DeepCopyInto(out *UserSettingsInitParamete *out = new(string) **out = **in } + if in.StudioWebPortalSettings != nil { + in, out := &in.StudioWebPortalSettings, &out.StudioWebPortalSettings + *out = make([]UserSettingsStudioWebPortalSettingsInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.TensorBoardAppSettings != nil { in, out := &in.TensorBoardAppSettings, &out.TensorBoardAppSettings *out = make([]UserSettingsTensorBoardAppSettingsInitParameters, len(*in)) @@ -21398,6 +22311,13 @@ func (in *UserSettingsObservation) DeepCopyInto(out *UserSettingsObservation) { *out = new(string) **out = **in } + if in.StudioWebPortalSettings != nil { + in, out := &in.StudioWebPortalSettings, &out.StudioWebPortalSettings + *out = make([]UserSettingsStudioWebPortalSettingsObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.TensorBoardAppSettings != nil { in, out := &in.TensorBoardAppSettings, &out.TensorBoardAppSettings *out = make([]UserSettingsTensorBoardAppSettingsObservation, len(*in)) @@ -21523,6 +22443,13 @@ func (in *UserSettingsParameters) DeepCopyInto(out *UserSettingsParameters) { *out = new(string) **out = **in } + if in.StudioWebPortalSettings != nil { + in, out := &in.StudioWebPortalSettings, &out.StudioWebPortalSettings + *out = make([]UserSettingsStudioWebPortalSettingsParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.TensorBoardAppSettings != nil { in, out := &in.TensorBoardAppSettings, &out.TensorBoardAppSettings *out = make([]UserSettingsTensorBoardAppSettingsParameters, len(*in)) @@ -22070,6 +22997,117 @@ func (in *UserSettingsSpaceStorageSettingsParameters) DeepCopy() *UserSettingsSp return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UserSettingsStudioWebPortalSettingsInitParameters) DeepCopyInto(out *UserSettingsStudioWebPortalSettingsInitParameters) { + *out = *in + if in.HiddenAppTypes != nil { + in, out := &in.HiddenAppTypes, &out.HiddenAppTypes + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.HiddenMLTools != nil { + in, out := &in.HiddenMLTools, &out.HiddenMLTools + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UserSettingsStudioWebPortalSettingsInitParameters. +func (in *UserSettingsStudioWebPortalSettingsInitParameters) DeepCopy() *UserSettingsStudioWebPortalSettingsInitParameters { + if in == nil { + return nil + } + out := new(UserSettingsStudioWebPortalSettingsInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UserSettingsStudioWebPortalSettingsObservation) DeepCopyInto(out *UserSettingsStudioWebPortalSettingsObservation) { + *out = *in + if in.HiddenAppTypes != nil { + in, out := &in.HiddenAppTypes, &out.HiddenAppTypes + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.HiddenMLTools != nil { + in, out := &in.HiddenMLTools, &out.HiddenMLTools + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UserSettingsStudioWebPortalSettingsObservation. +func (in *UserSettingsStudioWebPortalSettingsObservation) DeepCopy() *UserSettingsStudioWebPortalSettingsObservation { + if in == nil { + return nil + } + out := new(UserSettingsStudioWebPortalSettingsObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *UserSettingsStudioWebPortalSettingsParameters) DeepCopyInto(out *UserSettingsStudioWebPortalSettingsParameters) { + *out = *in + if in.HiddenAppTypes != nil { + in, out := &in.HiddenAppTypes, &out.HiddenAppTypes + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } + if in.HiddenMLTools != nil { + in, out := &in.HiddenMLTools, &out.HiddenMLTools + *out = make([]*string, len(*in)) + for i := range *in { + if (*in)[i] != nil { + in, out := &(*in)[i], &(*out)[i] + *out = new(string) + **out = **in + } + } + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UserSettingsStudioWebPortalSettingsParameters. +func (in *UserSettingsStudioWebPortalSettingsParameters) DeepCopy() *UserSettingsStudioWebPortalSettingsParameters { + if in == nil { + return nil + } + out := new(UserSettingsStudioWebPortalSettingsParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *UserSettingsTensorBoardAppSettingsDefaultResourceSpecInitParameters) DeepCopyInto(out *UserSettingsTensorBoardAppSettingsDefaultResourceSpecInitParameters) { *out = *in diff --git a/apis/sesv2/v1beta1/zz_generated.deepcopy.go b/apis/sesv2/v1beta1/zz_generated.deepcopy.go index 1e58fac18c..895da5ecb6 100644 --- a/apis/sesv2/v1beta1/zz_generated.deepcopy.go +++ b/apis/sesv2/v1beta1/zz_generated.deepcopy.go @@ -1844,6 +1844,66 @@ func (in *EmailIdentityStatus) DeepCopy() *EmailIdentityStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EventBridgeDestinationInitParameters) DeepCopyInto(out *EventBridgeDestinationInitParameters) { + *out = *in + if in.EventBusArn != nil { + in, out := &in.EventBusArn, &out.EventBusArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventBridgeDestinationInitParameters. +func (in *EventBridgeDestinationInitParameters) DeepCopy() *EventBridgeDestinationInitParameters { + if in == nil { + return nil + } + out := new(EventBridgeDestinationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EventBridgeDestinationObservation) DeepCopyInto(out *EventBridgeDestinationObservation) { + *out = *in + if in.EventBusArn != nil { + in, out := &in.EventBusArn, &out.EventBusArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventBridgeDestinationObservation. +func (in *EventBridgeDestinationObservation) DeepCopy() *EventBridgeDestinationObservation { + if in == nil { + return nil + } + out := new(EventBridgeDestinationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EventBridgeDestinationParameters) DeepCopyInto(out *EventBridgeDestinationParameters) { + *out = *in + if in.EventBusArn != nil { + in, out := &in.EventBusArn, &out.EventBusArn + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventBridgeDestinationParameters. +func (in *EventBridgeDestinationParameters) DeepCopy() *EventBridgeDestinationParameters { + if in == nil { + return nil + } + out := new(EventBridgeDestinationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *EventDestinationInitParameters) DeepCopyInto(out *EventDestinationInitParameters) { *out = *in @@ -1859,6 +1919,13 @@ func (in *EventDestinationInitParameters) DeepCopyInto(out *EventDestinationInit *out = new(bool) **out = **in } + if in.EventBridgeDestination != nil { + in, out := &in.EventBridgeDestination, &out.EventBridgeDestination + *out = make([]EventBridgeDestinationInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.KinesisFirehoseDestination != nil { in, out := &in.KinesisFirehoseDestination, &out.KinesisFirehoseDestination *out = make([]KinesisFirehoseDestinationInitParameters, len(*in)) @@ -1918,6 +1985,13 @@ func (in *EventDestinationObservation) DeepCopyInto(out *EventDestinationObserva *out = new(bool) **out = **in } + if in.EventBridgeDestination != nil { + in, out := &in.EventBridgeDestination, &out.EventBridgeDestination + *out = make([]EventBridgeDestinationObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.KinesisFirehoseDestination != nil { in, out := &in.KinesisFirehoseDestination, &out.KinesisFirehoseDestination *out = make([]KinesisFirehoseDestinationObservation, len(*in)) @@ -1977,6 +2051,13 @@ func (in *EventDestinationParameters) DeepCopyInto(out *EventDestinationParamete *out = new(bool) **out = **in } + if in.EventBridgeDestination != nil { + in, out := &in.EventBridgeDestination, &out.EventBridgeDestination + *out = make([]EventBridgeDestinationParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.KinesisFirehoseDestination != nil { in, out := &in.KinesisFirehoseDestination, &out.KinesisFirehoseDestination *out = make([]KinesisFirehoseDestinationParameters, len(*in)) diff --git a/apis/sfn/v1beta1/zz_activity_types.go b/apis/sfn/v1beta1/zz_activity_types.go index 6e41e83a85..42206ee3e8 100755 --- a/apis/sfn/v1beta1/zz_activity_types.go +++ b/apis/sfn/v1beta1/zz_activity_types.go @@ -13,54 +13,7 @@ import ( v1 "github.com/crossplane/crossplane-runtime/apis/common/v1" ) -type ActivityInitParameters struct { - - // Defines what encryption configuration is used to encrypt data in the Activity. For more information see the section Data at rest encyption in the AWS Step Functions User Guide. - EncryptionConfiguration *EncryptionConfigurationInitParameters `json:"encryptionConfiguration,omitempty" tf:"encryption_configuration,omitempty"` - - // Key-value map of resource tags. - // +mapType=granular - Tags map[string]*string `json:"tags,omitempty" tf:"tags,omitempty"` -} - -type ActivityObservation struct { - - // The date the activity was created. - CreationDate *string `json:"creationDate,omitempty" tf:"creation_date,omitempty"` - - // Defines what encryption configuration is used to encrypt data in the Activity. For more information see the section Data at rest encyption in the AWS Step Functions User Guide. - EncryptionConfiguration *EncryptionConfigurationObservation `json:"encryptionConfiguration,omitempty" tf:"encryption_configuration,omitempty"` - - // The Amazon Resource Name (ARN) that identifies the created activity. - ID *string `json:"id,omitempty" tf:"id,omitempty"` - - // Key-value map of resource tags. - // +mapType=granular - Tags map[string]*string `json:"tags,omitempty" tf:"tags,omitempty"` - - // A map of tags assigned to the resource, including those inherited from the provider default_tags configuration block. - // +mapType=granular - TagsAll map[string]*string `json:"tagsAll,omitempty" tf:"tags_all,omitempty"` -} - -type ActivityParameters struct { - - // Defines what encryption configuration is used to encrypt data in the Activity. For more information see the section Data at rest encyption in the AWS Step Functions User Guide. - // +kubebuilder:validation:Optional - EncryptionConfiguration *EncryptionConfigurationParameters `json:"encryptionConfiguration,omitempty" tf:"encryption_configuration,omitempty"` - - // Region is the region you'd like your resource to be created in. - // +upjet:crd:field:TFTag=- - // +kubebuilder:validation:Required - Region *string `json:"region" tf:"-"` - - // Key-value map of resource tags. - // +kubebuilder:validation:Optional - // +mapType=granular - Tags map[string]*string `json:"tags,omitempty" tf:"tags,omitempty"` -} - -type EncryptionConfigurationInitParameters struct { +type ActivityEncryptionConfigurationInitParameters struct { // Maximum duration for which Activities will reuse data keys. When the period expires, Activities will call GenerateDataKey. This setting only applies to customer managed KMS key and does not apply to AWS owned KMS key. KMSDataKeyReusePeriodSeconds *float64 `json:"kmsDataKeyReusePeriodSeconds,omitempty" tf:"kms_data_key_reuse_period_seconds,omitempty"` @@ -82,7 +35,7 @@ type EncryptionConfigurationInitParameters struct { Type *string `json:"type,omitempty" tf:"type,omitempty"` } -type EncryptionConfigurationObservation struct { +type ActivityEncryptionConfigurationObservation struct { // Maximum duration for which Activities will reuse data keys. When the period expires, Activities will call GenerateDataKey. This setting only applies to customer managed KMS key and does not apply to AWS owned KMS key. KMSDataKeyReusePeriodSeconds *float64 `json:"kmsDataKeyReusePeriodSeconds,omitempty" tf:"kms_data_key_reuse_period_seconds,omitempty"` @@ -94,7 +47,7 @@ type EncryptionConfigurationObservation struct { Type *string `json:"type,omitempty" tf:"type,omitempty"` } -type EncryptionConfigurationParameters struct { +type ActivityEncryptionConfigurationParameters struct { // Maximum duration for which Activities will reuse data keys. When the period expires, Activities will call GenerateDataKey. This setting only applies to customer managed KMS key and does not apply to AWS owned KMS key. // +kubebuilder:validation:Optional @@ -119,6 +72,53 @@ type EncryptionConfigurationParameters struct { Type *string `json:"type,omitempty" tf:"type,omitempty"` } +type ActivityInitParameters struct { + + // Defines what encryption configuration is used to encrypt data in the Activity. For more information see the section Data at rest encyption in the AWS Step Functions User Guide. + EncryptionConfiguration *ActivityEncryptionConfigurationInitParameters `json:"encryptionConfiguration,omitempty" tf:"encryption_configuration,omitempty"` + + // Key-value map of resource tags. + // +mapType=granular + Tags map[string]*string `json:"tags,omitempty" tf:"tags,omitempty"` +} + +type ActivityObservation struct { + + // The date the activity was created. + CreationDate *string `json:"creationDate,omitempty" tf:"creation_date,omitempty"` + + // Defines what encryption configuration is used to encrypt data in the Activity. For more information see the section Data at rest encyption in the AWS Step Functions User Guide. + EncryptionConfiguration *ActivityEncryptionConfigurationObservation `json:"encryptionConfiguration,omitempty" tf:"encryption_configuration,omitempty"` + + // The Amazon Resource Name (ARN) that identifies the created activity. + ID *string `json:"id,omitempty" tf:"id,omitempty"` + + // Key-value map of resource tags. + // +mapType=granular + Tags map[string]*string `json:"tags,omitempty" tf:"tags,omitempty"` + + // A map of tags assigned to the resource, including those inherited from the provider default_tags configuration block. + // +mapType=granular + TagsAll map[string]*string `json:"tagsAll,omitempty" tf:"tags_all,omitempty"` +} + +type ActivityParameters struct { + + // Defines what encryption configuration is used to encrypt data in the Activity. For more information see the section Data at rest encyption in the AWS Step Functions User Guide. + // +kubebuilder:validation:Optional + EncryptionConfiguration *ActivityEncryptionConfigurationParameters `json:"encryptionConfiguration,omitempty" tf:"encryption_configuration,omitempty"` + + // Region is the region you'd like your resource to be created in. + // +upjet:crd:field:TFTag=- + // +kubebuilder:validation:Required + Region *string `json:"region" tf:"-"` + + // Key-value map of resource tags. + // +kubebuilder:validation:Optional + // +mapType=granular + Tags map[string]*string `json:"tags,omitempty" tf:"tags,omitempty"` +} + // ActivitySpec defines the desired state of Activity type ActivitySpec struct { v1.ResourceSpec `json:",inline"` diff --git a/apis/sfn/v1beta1/zz_generated.deepcopy.go b/apis/sfn/v1beta1/zz_generated.deepcopy.go index 4101da50cb..83fc32499e 100644 --- a/apis/sfn/v1beta1/zz_generated.deepcopy.go +++ b/apis/sfn/v1beta1/zz_generated.deepcopy.go @@ -40,12 +40,122 @@ func (in *Activity) DeepCopyObject() runtime.Object { return nil } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ActivityEncryptionConfigurationInitParameters) DeepCopyInto(out *ActivityEncryptionConfigurationInitParameters) { + *out = *in + if in.KMSDataKeyReusePeriodSeconds != nil { + in, out := &in.KMSDataKeyReusePeriodSeconds, &out.KMSDataKeyReusePeriodSeconds + *out = new(float64) + **out = **in + } + if in.KMSKeyID != nil { + in, out := &in.KMSKeyID, &out.KMSKeyID + *out = new(string) + **out = **in + } + if in.KMSKeyIDRef != nil { + in, out := &in.KMSKeyIDRef, &out.KMSKeyIDRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.KMSKeyIDSelector != nil { + in, out := &in.KMSKeyIDSelector, &out.KMSKeyIDSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ActivityEncryptionConfigurationInitParameters. +func (in *ActivityEncryptionConfigurationInitParameters) DeepCopy() *ActivityEncryptionConfigurationInitParameters { + if in == nil { + return nil + } + out := new(ActivityEncryptionConfigurationInitParameters) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ActivityEncryptionConfigurationObservation) DeepCopyInto(out *ActivityEncryptionConfigurationObservation) { + *out = *in + if in.KMSDataKeyReusePeriodSeconds != nil { + in, out := &in.KMSDataKeyReusePeriodSeconds, &out.KMSDataKeyReusePeriodSeconds + *out = new(float64) + **out = **in + } + if in.KMSKeyID != nil { + in, out := &in.KMSKeyID, &out.KMSKeyID + *out = new(string) + **out = **in + } + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ActivityEncryptionConfigurationObservation. +func (in *ActivityEncryptionConfigurationObservation) DeepCopy() *ActivityEncryptionConfigurationObservation { + if in == nil { + return nil + } + out := new(ActivityEncryptionConfigurationObservation) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ActivityEncryptionConfigurationParameters) DeepCopyInto(out *ActivityEncryptionConfigurationParameters) { + *out = *in + if in.KMSDataKeyReusePeriodSeconds != nil { + in, out := &in.KMSDataKeyReusePeriodSeconds, &out.KMSDataKeyReusePeriodSeconds + *out = new(float64) + **out = **in + } + if in.KMSKeyID != nil { + in, out := &in.KMSKeyID, &out.KMSKeyID + *out = new(string) + **out = **in + } + if in.KMSKeyIDRef != nil { + in, out := &in.KMSKeyIDRef, &out.KMSKeyIDRef + *out = new(v1.Reference) + (*in).DeepCopyInto(*out) + } + if in.KMSKeyIDSelector != nil { + in, out := &in.KMSKeyIDSelector, &out.KMSKeyIDSelector + *out = new(v1.Selector) + (*in).DeepCopyInto(*out) + } + if in.Type != nil { + in, out := &in.Type, &out.Type + *out = new(string) + **out = **in + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ActivityEncryptionConfigurationParameters. +func (in *ActivityEncryptionConfigurationParameters) DeepCopy() *ActivityEncryptionConfigurationParameters { + if in == nil { + return nil + } + out := new(ActivityEncryptionConfigurationParameters) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *ActivityInitParameters) DeepCopyInto(out *ActivityInitParameters) { *out = *in if in.EncryptionConfiguration != nil { in, out := &in.EncryptionConfiguration, &out.EncryptionConfiguration - *out = new(EncryptionConfigurationInitParameters) + *out = new(ActivityEncryptionConfigurationInitParameters) (*in).DeepCopyInto(*out) } if in.Tags != nil { @@ -118,7 +228,7 @@ func (in *ActivityObservation) DeepCopyInto(out *ActivityObservation) { } if in.EncryptionConfiguration != nil { in, out := &in.EncryptionConfiguration, &out.EncryptionConfiguration - *out = new(EncryptionConfigurationObservation) + *out = new(ActivityEncryptionConfigurationObservation) (*in).DeepCopyInto(*out) } if in.ID != nil { @@ -175,7 +285,7 @@ func (in *ActivityParameters) DeepCopyInto(out *ActivityParameters) { *out = *in if in.EncryptionConfiguration != nil { in, out := &in.EncryptionConfiguration, &out.EncryptionConfiguration - *out = new(EncryptionConfigurationParameters) + *out = new(ActivityEncryptionConfigurationParameters) (*in).DeepCopyInto(*out) } if in.Region != nil { @@ -481,6 +591,13 @@ func (in *StateMachineInitParameters) DeepCopyInto(out *StateMachineInitParamete *out = new(string) **out = **in } + if in.EncryptionConfiguration != nil { + in, out := &in.EncryptionConfiguration, &out.EncryptionConfiguration + *out = make([]EncryptionConfigurationInitParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.LoggingConfiguration != nil { in, out := &in.LoggingConfiguration, &out.LoggingConfiguration *out = make([]LoggingConfigurationInitParameters, len(*in)) @@ -603,6 +720,13 @@ func (in *StateMachineObservation) DeepCopyInto(out *StateMachineObservation) { *out = new(string) **out = **in } + if in.EncryptionConfiguration != nil { + in, out := &in.EncryptionConfiguration, &out.EncryptionConfiguration + *out = make([]EncryptionConfigurationObservation, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.ID != nil { in, out := &in.ID, &out.ID *out = new(string) @@ -709,6 +833,13 @@ func (in *StateMachineParameters) DeepCopyInto(out *StateMachineParameters) { *out = new(string) **out = **in } + if in.EncryptionConfiguration != nil { + in, out := &in.EncryptionConfiguration, &out.EncryptionConfiguration + *out = make([]EncryptionConfigurationParameters, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } if in.LoggingConfiguration != nil { in, out := &in.LoggingConfiguration, &out.LoggingConfiguration *out = make([]LoggingConfigurationParameters, len(*in)) diff --git a/apis/sfn/v1beta1/zz_generated.resolvers.go b/apis/sfn/v1beta1/zz_generated.resolvers.go index baeb9fe251..dc5678af5d 100644 --- a/apis/sfn/v1beta1/zz_generated.resolvers.go +++ b/apis/sfn/v1beta1/zz_generated.resolvers.go @@ -82,12 +82,33 @@ func (mg *StateMachine) ResolveReferences(ctx context.Context, c client.Reader) var rsp reference.ResolutionResponse var err error + + for i3 := 0; i3 < len(mg.Spec.ForProvider.EncryptionConfiguration); i3++ { + { + m, l, err = apisresolver.GetManagedResource("kms.aws.upbound.io", "v1beta1", "Key", "KeyList") + if err != nil { + return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") + } + rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ + CurrentValue: reference.FromPtrValue(mg.Spec.ForProvider.EncryptionConfiguration[i3].KMSKeyID), + Extract: resource.ExtractParamPath("arn", true), + Reference: mg.Spec.ForProvider.EncryptionConfiguration[i3].KMSKeyIDRef, + Selector: mg.Spec.ForProvider.EncryptionConfiguration[i3].KMSKeyIDSelector, + To: reference.To{List: l, Managed: m}, + }) + } + if err != nil { + return errors.Wrap(err, "mg.Spec.ForProvider.EncryptionConfiguration[i3].KMSKeyID") + } + mg.Spec.ForProvider.EncryptionConfiguration[i3].KMSKeyID = reference.ToPtrValue(rsp.ResolvedValue) + mg.Spec.ForProvider.EncryptionConfiguration[i3].KMSKeyIDRef = rsp.ResolvedReference + + } { m, l, err = apisresolver.GetManagedResource("iam.aws.upbound.io", "v1beta1", "Role", "RoleList") if err != nil { return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") } - rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ CurrentValue: reference.FromPtrValue(mg.Spec.ForProvider.RoleArn), Extract: common.ARNExtractor(), @@ -101,12 +122,33 @@ func (mg *StateMachine) ResolveReferences(ctx context.Context, c client.Reader) } mg.Spec.ForProvider.RoleArn = reference.ToPtrValue(rsp.ResolvedValue) mg.Spec.ForProvider.RoleArnRef = rsp.ResolvedReference + + for i3 := 0; i3 < len(mg.Spec.InitProvider.EncryptionConfiguration); i3++ { + { + m, l, err = apisresolver.GetManagedResource("kms.aws.upbound.io", "v1beta1", "Key", "KeyList") + if err != nil { + return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") + } + rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ + CurrentValue: reference.FromPtrValue(mg.Spec.InitProvider.EncryptionConfiguration[i3].KMSKeyID), + Extract: resource.ExtractParamPath("arn", true), + Reference: mg.Spec.InitProvider.EncryptionConfiguration[i3].KMSKeyIDRef, + Selector: mg.Spec.InitProvider.EncryptionConfiguration[i3].KMSKeyIDSelector, + To: reference.To{List: l, Managed: m}, + }) + } + if err != nil { + return errors.Wrap(err, "mg.Spec.InitProvider.EncryptionConfiguration[i3].KMSKeyID") + } + mg.Spec.InitProvider.EncryptionConfiguration[i3].KMSKeyID = reference.ToPtrValue(rsp.ResolvedValue) + mg.Spec.InitProvider.EncryptionConfiguration[i3].KMSKeyIDRef = rsp.ResolvedReference + + } { m, l, err = apisresolver.GetManagedResource("iam.aws.upbound.io", "v1beta1", "Role", "RoleList") if err != nil { return errors.Wrap(err, "failed to get the reference target managed resource and its list for reference resolution") } - rsp, err = r.Resolve(ctx, reference.ResolutionRequest{ CurrentValue: reference.FromPtrValue(mg.Spec.InitProvider.RoleArn), Extract: common.ARNExtractor(), diff --git a/package/crds/amplify.aws.upbound.io_apps.yaml b/package/crds/amplify.aws.upbound.io_apps.yaml index 2d6fff7953..a6b640ef75 100644 --- a/package/crds/amplify.aws.upbound.io_apps.yaml +++ b/package/crds/amplify.aws.upbound.io_apps.yaml @@ -182,6 +182,17 @@ spec: description: The build specification (build spec) for an Amplify app. type: string + cacheConfig: + description: Cache configuration for the Amplify app. See cache_config + Block for details. + items: + properties: + type: + description: 'Type of cache configuration to use for an + Amplify app. Valid values: AMPLIFY_MANAGED, AMPLIFY_MANAGED_NO_COOKIES.' + type: string + type: object + type: array customHeaders: description: The custom HTTP headers for an Amplify app. type: string @@ -476,6 +487,17 @@ spec: description: The build specification (build spec) for an Amplify app. type: string + cacheConfig: + description: Cache configuration for the Amplify app. See cache_config + Block for details. + items: + properties: + type: + description: 'Type of cache configuration to use for an + Amplify app. Valid values: AMPLIFY_MANAGED, AMPLIFY_MANAGED_NO_COOKIES.' + type: string + type: object + type: array customHeaders: description: The custom HTTP headers for an Amplify app. type: string @@ -874,6 +896,17 @@ spec: description: The build specification (build spec) for an Amplify app. type: string + cacheConfig: + description: Cache configuration for the Amplify app. See cache_config + Block for details. + items: + properties: + type: + description: 'Type of cache configuration to use for an + Amplify app. Valid values: AMPLIFY_MANAGED, AMPLIFY_MANAGED_NO_COOKIES.' + type: string + type: object + type: array customHeaders: description: The custom HTTP headers for an Amplify app. type: string diff --git a/package/crds/appflow.aws.upbound.io_flows.yaml b/package/crds/appflow.aws.upbound.io_flows.yaml index 189c3be839..4afafe8e0f 100644 --- a/package/crds/appflow.aws.upbound.io_flows.yaml +++ b/package/crds/appflow.aws.upbound.io_flows.yaml @@ -467,6 +467,15 @@ spec: in the prefix. Valid values are YEAR, MONTH, DAY, HOUR, and MINUTE. type: string + prefixHierarchy: + description: Determines whether + the destination file path includes + either or both of the selected + elements. Valid values are EXECUTION_ID + and SCHEMA_VERSION + items: + type: string + type: array prefixType: description: Determines the format of the prefix, and whether it @@ -701,6 +710,15 @@ spec: in the prefix. Valid values are YEAR, MONTH, DAY, HOUR, and MINUTE. type: string + prefixHierarchy: + description: Determines whether + the destination file path includes + either or both of the selected + elements. Valid values are EXECUTION_ID + and SCHEMA_VERSION + items: + type: string + type: array prefixType: description: Determines the format of the prefix, and whether it @@ -773,6 +791,34 @@ spec: If you don't provide anything here, Amazon AppFlow uses the Amazon AppFlow-managed KMS key. type: string + metadataCatalogConfig: + description: A Catalog that determines the configuration that + Amazon AppFlow uses when it catalogs the data that’s transferred + by the associated flow. When Amazon AppFlow catalogs the data + from a flow, it stores metadata in a data catalog. + items: + properties: + glueDataCatalog: + items: + properties: + databaseName: + description: The name of an existing Glue database + to store the metadata tables that Amazon AppFlow + creates. + type: string + roleArn: + description: The ARN of an IAM role that grants AppFlow + the permissions it needs to create Data Catalog + tables, databases, and partitions. + type: string + tablePrefix: + description: A naming prefix for each Data Catalog + table that Amazon AppFlow creates + type: string + type: object + type: array + type: object + type: array region: description: Region is the region you'd like your resource to be created in. @@ -1734,6 +1780,15 @@ spec: in the prefix. Valid values are YEAR, MONTH, DAY, HOUR, and MINUTE. type: string + prefixHierarchy: + description: Determines whether + the destination file path includes + either or both of the selected + elements. Valid values are EXECUTION_ID + and SCHEMA_VERSION + items: + type: string + type: array prefixType: description: Determines the format of the prefix, and whether it @@ -1968,6 +2023,15 @@ spec: in the prefix. Valid values are YEAR, MONTH, DAY, HOUR, and MINUTE. type: string + prefixHierarchy: + description: Determines whether + the destination file path includes + either or both of the selected + elements. Valid values are EXECUTION_ID + and SCHEMA_VERSION + items: + type: string + type: array prefixType: description: Determines the format of the prefix, and whether it @@ -2040,6 +2104,34 @@ spec: If you don't provide anything here, Amazon AppFlow uses the Amazon AppFlow-managed KMS key. type: string + metadataCatalogConfig: + description: A Catalog that determines the configuration that + Amazon AppFlow uses when it catalogs the data that’s transferred + by the associated flow. When Amazon AppFlow catalogs the data + from a flow, it stores metadata in a data catalog. + items: + properties: + glueDataCatalog: + items: + properties: + databaseName: + description: The name of an existing Glue database + to store the metadata tables that Amazon AppFlow + creates. + type: string + roleArn: + description: The ARN of an IAM role that grants AppFlow + the permissions it needs to create Data Catalog + tables, databases, and partitions. + type: string + tablePrefix: + description: A naming prefix for each Data Catalog + table that Amazon AppFlow creates + type: string + type: object + type: array + type: object + type: array sourceFlowConfig: description: The Source Flow Config that controls how Amazon AppFlow retrieves data from the source connector. @@ -3098,6 +3190,15 @@ spec: in the prefix. Valid values are YEAR, MONTH, DAY, HOUR, and MINUTE. type: string + prefixHierarchy: + description: Determines whether + the destination file path includes + either or both of the selected + elements. Valid values are EXECUTION_ID + and SCHEMA_VERSION + items: + type: string + type: array prefixType: description: Determines the format of the prefix, and whether it @@ -3332,6 +3433,15 @@ spec: in the prefix. Valid values are YEAR, MONTH, DAY, HOUR, and MINUTE. type: string + prefixHierarchy: + description: Determines whether + the destination file path includes + either or both of the selected + elements. Valid values are EXECUTION_ID + and SCHEMA_VERSION + items: + type: string + type: array prefixType: description: Determines the format of the prefix, and whether it @@ -3409,6 +3519,34 @@ spec: If you don't provide anything here, Amazon AppFlow uses the Amazon AppFlow-managed KMS key. type: string + metadataCatalogConfig: + description: A Catalog that determines the configuration that + Amazon AppFlow uses when it catalogs the data that’s transferred + by the associated flow. When Amazon AppFlow catalogs the data + from a flow, it stores metadata in a data catalog. + items: + properties: + glueDataCatalog: + items: + properties: + databaseName: + description: The name of an existing Glue database + to store the metadata tables that Amazon AppFlow + creates. + type: string + roleArn: + description: The ARN of an IAM role that grants AppFlow + the permissions it needs to create Data Catalog + tables, databases, and partitions. + type: string + tablePrefix: + description: A naming prefix for each Data Catalog + table that Amazon AppFlow creates + type: string + type: object + type: array + type: object + type: array sourceFlowConfig: description: The Source Flow Config that controls how Amazon AppFlow retrieves data from the source connector. diff --git a/package/crds/appsync.aws.upbound.io_graphqlapis.yaml b/package/crds/appsync.aws.upbound.io_graphqlapis.yaml index 470298a79d..bcd74ad9ca 100644 --- a/package/crds/appsync.aws.upbound.io_graphqlapis.yaml +++ b/package/crds/appsync.aws.upbound.io_graphqlapis.yaml @@ -157,10 +157,34 @@ spec: type: array type: object type: array + apiType: + description: API type. Valid values are GRAPHQL or MERGED. A MERGED + type requires merged_api_execution_role_arn to be set. + type: string authenticationType: description: 'Authentication type. Valid values: API_KEY, AWS_IAM, AMAZON_COGNITO_USER_POOLS, OPENID_CONNECT, AWS_LAMBDA' type: string + enhancedMetricsConfig: + description: Enables and controls the enhanced metrics feature. + See enhanced_metrics_config Block for details. + items: + properties: + dataSourceLevelMetricsBehavior: + description: 'How data source metrics will be emitted to + CloudWatch. Valid values: FULL_REQUEST_DATA_SOURCE_METRICS, + PER_DATA_SOURCE_METRICS' + type: string + operationLevelMetricsConfig: + description: 'How operation metrics will be emitted to CloudWatch. + Valid values: ENABLED, DISABLED' + type: string + resolverLevelMetricsBehavior: + description: 'How resolver metrics will be emitted to CloudWatch. + Valid values: FULL_REQUEST_RESOLVER_METRICS, PER_RESOLVER_METRICS' + type: string + type: object + type: array introspectionConfig: description: Sets the value of the GraphQL API to enable (ENABLED) or disable (DISABLED) introspection. If no value is provided, @@ -289,6 +313,84 @@ spec: type: string type: object type: array + mergedApiExecutionRoleArn: + description: ARN of the execution role when api_type is set to + MERGED. + type: string + mergedApiExecutionRoleArnRef: + description: Reference to a Role in iam to populate mergedApiExecutionRoleArn. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + mergedApiExecutionRoleArnSelector: + description: Selector for a Role in iam to populate mergedApiExecutionRoleArn. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object name: description: User-supplied name for the GraphqlApi. type: string @@ -556,10 +658,34 @@ spec: type: array type: object type: array + apiType: + description: API type. Valid values are GRAPHQL or MERGED. A MERGED + type requires merged_api_execution_role_arn to be set. + type: string authenticationType: description: 'Authentication type. Valid values: API_KEY, AWS_IAM, AMAZON_COGNITO_USER_POOLS, OPENID_CONNECT, AWS_LAMBDA' type: string + enhancedMetricsConfig: + description: Enables and controls the enhanced metrics feature. + See enhanced_metrics_config Block for details. + items: + properties: + dataSourceLevelMetricsBehavior: + description: 'How data source metrics will be emitted to + CloudWatch. Valid values: FULL_REQUEST_DATA_SOURCE_METRICS, + PER_DATA_SOURCE_METRICS' + type: string + operationLevelMetricsConfig: + description: 'How operation metrics will be emitted to CloudWatch. + Valid values: ENABLED, DISABLED' + type: string + resolverLevelMetricsBehavior: + description: 'How resolver metrics will be emitted to CloudWatch. + Valid values: FULL_REQUEST_RESOLVER_METRICS, PER_RESOLVER_METRICS' + type: string + type: object + type: array introspectionConfig: description: Sets the value of the GraphQL API to enable (ENABLED) or disable (DISABLED) introspection. If no value is provided, @@ -688,6 +814,84 @@ spec: type: string type: object type: array + mergedApiExecutionRoleArn: + description: ARN of the execution role when api_type is set to + MERGED. + type: string + mergedApiExecutionRoleArnRef: + description: Reference to a Role in iam to populate mergedApiExecutionRoleArn. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + mergedApiExecutionRoleArnSelector: + description: Selector for a Role in iam to populate mergedApiExecutionRoleArn. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object name: description: User-supplied name for the GraphqlApi. type: string @@ -1117,6 +1321,10 @@ spec: type: array type: object type: array + apiType: + description: API type. Valid values are GRAPHQL or MERGED. A MERGED + type requires merged_api_execution_role_arn to be set. + type: string arn: description: ARN type: string @@ -1124,6 +1332,26 @@ spec: description: 'Authentication type. Valid values: API_KEY, AWS_IAM, AMAZON_COGNITO_USER_POOLS, OPENID_CONNECT, AWS_LAMBDA' type: string + enhancedMetricsConfig: + description: Enables and controls the enhanced metrics feature. + See enhanced_metrics_config Block for details. + items: + properties: + dataSourceLevelMetricsBehavior: + description: 'How data source metrics will be emitted to + CloudWatch. Valid values: FULL_REQUEST_DATA_SOURCE_METRICS, + PER_DATA_SOURCE_METRICS' + type: string + operationLevelMetricsConfig: + description: 'How operation metrics will be emitted to CloudWatch. + Valid values: ENABLED, DISABLED' + type: string + resolverLevelMetricsBehavior: + description: 'How resolver metrics will be emitted to CloudWatch. + Valid values: FULL_REQUEST_RESOLVER_METRICS, PER_RESOLVER_METRICS' + type: string + type: object + type: array id: description: API ID type: string @@ -1181,6 +1409,10 @@ spec: type: string type: object type: array + mergedApiExecutionRoleArn: + description: ARN of the execution role when api_type is set to + MERGED. + type: string name: description: User-supplied name for the GraphqlApi. type: string diff --git a/package/crds/batch.aws.upbound.io_jobdefinitions.yaml b/package/crds/batch.aws.upbound.io_jobdefinitions.yaml index cf90aec464..6372b92cc3 100644 --- a/package/crds/batch.aws.upbound.io_jobdefinitions.yaml +++ b/package/crds/batch.aws.upbound.io_jobdefinitions.yaml @@ -83,6 +83,11 @@ spec: created. This parameter determines if the previous version is deregistered (INACTIVE) or left ACTIVE. Defaults to true. type: boolean + ecsProperties: + description: Valid ECS properties provided as a single valid JSON + document. This parameter is only valid if the type parameter + is container. + type: string eksProperties: description: A valid eks properties. This parameter is only valid if the type parameter is container. @@ -218,6 +223,16 @@ spec: and don't require the overhead of IP allocation for each pod for incoming connections. type: boolean + imagePullSecret: + description: List of Kubernetes secret resources. + See image_pull_secret below. + items: + properties: + name: + description: Unique identifier. + type: string + type: object + type: array metadata: description: Metadata about the Kubernetes pod. items: @@ -415,6 +430,11 @@ spec: created. This parameter determines if the previous version is deregistered (INACTIVE) or left ACTIVE. Defaults to true. type: boolean + ecsProperties: + description: Valid ECS properties provided as a single valid JSON + document. This parameter is only valid if the type parameter + is container. + type: string eksProperties: description: A valid eks properties. This parameter is only valid if the type parameter is container. @@ -550,6 +570,16 @@ spec: and don't require the overhead of IP allocation for each pod for incoming connections. type: boolean + imagePullSecret: + description: List of Kubernetes secret resources. + See image_pull_secret below. + items: + properties: + name: + description: Unique identifier. + type: string + type: object + type: array metadata: description: Metadata about the Kubernetes pod. items: @@ -916,6 +946,11 @@ spec: created. This parameter determines if the previous version is deregistered (INACTIVE) or left ACTIVE. Defaults to true. type: boolean + ecsProperties: + description: Valid ECS properties provided as a single valid JSON + document. This parameter is only valid if the type parameter + is container. + type: string eksProperties: description: A valid eks properties. This parameter is only valid if the type parameter is container. @@ -1051,6 +1086,16 @@ spec: and don't require the overhead of IP allocation for each pod for incoming connections. type: boolean + imagePullSecret: + description: List of Kubernetes secret resources. + See image_pull_secret below. + items: + properties: + name: + description: Unique identifier. + type: string + type: object + type: array metadata: description: Metadata about the Kubernetes pod. items: diff --git a/package/crds/cloudformation.aws.upbound.io_stacksetinstances.yaml b/package/crds/cloudformation.aws.upbound.io_stacksetinstances.yaml index f53285bc8e..a3a26e7f91 100644 --- a/package/crds/cloudformation.aws.upbound.io_stacksetinstances.yaml +++ b/package/crds/cloudformation.aws.upbound.io_stacksetinstances.yaml @@ -120,6 +120,11 @@ spec: stack set operation. items: properties: + concurrencyMode: + description: Specifies how the concurrency level behaves + during the operation execution. Valid values are STRICT_FAILURE_TOLERANCE + and SOFT_FAILURE_TOLERANCE. + type: string failureToleranceCount: description: The number of accounts, per Region, for which this operation can fail before AWS CloudFormation stops @@ -309,6 +314,11 @@ spec: stack set operation. items: properties: + concurrencyMode: + description: Specifies how the concurrency level behaves + during the operation execution. Valid values are STRICT_FAILURE_TOLERANCE + and SOFT_FAILURE_TOLERANCE. + type: string failureToleranceCount: description: The number of accounts, per Region, for which this operation can fail before AWS CloudFormation stops @@ -658,6 +668,11 @@ spec: stack set operation. items: properties: + concurrencyMode: + description: Specifies how the concurrency level behaves + during the operation execution. Valid values are STRICT_FAILURE_TOLERANCE + and SOFT_FAILURE_TOLERANCE. + type: string failureToleranceCount: description: The number of accounts, per Region, for which this operation can fail before AWS CloudFormation stops diff --git a/package/crds/codepipeline.aws.upbound.io_codepipelines.yaml b/package/crds/codepipeline.aws.upbound.io_codepipelines.yaml index d3522f4658..75ccabaa01 100644 --- a/package/crds/codepipeline.aws.upbound.io_codepipelines.yaml +++ b/package/crds/codepipeline.aws.upbound.io_codepipelines.yaml @@ -343,6 +343,8 @@ spec: runOrder: description: The order in which actions are run. type: number + timeoutInMinutes: + type: number version: description: A string that identifies the action type. type: string @@ -807,6 +809,8 @@ spec: runOrder: description: The order in which actions are run. type: number + timeoutInMinutes: + type: number version: description: A string that identifies the action type. type: string @@ -1303,6 +1307,8 @@ spec: runOrder: description: The order in which actions are run. type: number + timeoutInMinutes: + type: number version: description: A string that identifies the action type. type: string diff --git a/package/crds/cognitoidp.aws.upbound.io_userpools.yaml b/package/crds/cognitoidp.aws.upbound.io_userpools.yaml index d755c286fd..47a13a4796 100644 --- a/package/crds/cognitoidp.aws.upbound.io_userpools.yaml +++ b/package/crds/cognitoidp.aws.upbound.io_userpools.yaml @@ -1316,6 +1316,14 @@ spec: description: Minimum length of the password policy that you have set. type: number + passwordHistorySize: + description: Number of previous passwords that you want + Amazon Cognito to restrict each user from reusing. Users + can't set a password that matches any of number of previous + passwords specified by this argument. A value of 0 means + that password history is not enforced. Valid values are + between 0 and 24. + type: number requireLowercase: description: Whether you have required users to use at least one lowercase letter in their password. @@ -2877,6 +2885,14 @@ spec: description: Minimum length of the password policy that you have set. type: number + passwordHistorySize: + description: Number of previous passwords that you want + Amazon Cognito to restrict each user from reusing. Users + can't set a password that matches any of number of previous + passwords specified by this argument. A value of 0 means + that password history is not enforced. Valid values are + between 0 and 24. + type: number requireLowercase: description: Whether you have required users to use at least one lowercase letter in their password. @@ -3639,6 +3655,14 @@ spec: description: Minimum length of the password policy that you have set. type: number + passwordHistorySize: + description: Number of previous passwords that you want + Amazon Cognito to restrict each user from reusing. Users + can't set a password that matches any of number of previous + passwords specified by this argument. A value of 0 means + that password history is not enforced. Valid values are + between 0 and 24. + type: number requireLowercase: description: Whether you have required users to use at least one lowercase letter in their password. diff --git a/package/crds/dynamodb.aws.upbound.io_tables.yaml b/package/crds/dynamodb.aws.upbound.io_tables.yaml index ab0fd32629..f514dfd497 100644 --- a/package/crds/dynamodb.aws.upbound.io_tables.yaml +++ b/package/crds/dynamodb.aws.upbound.io_tables.yaml @@ -297,6 +297,10 @@ spec: description: Name of the table to restore. Must match the name of an existing table. type: string + restoreSourceTableArn: + description: ARN of the source table to restore. Must be supplied + for cross-region restores. + type: string restoreToLatestTime: description: If set, restores table to the most recent point-in-time recovery point. @@ -600,6 +604,10 @@ spec: description: Name of the table to restore. Must match the name of an existing table. type: string + restoreSourceTableArn: + description: ARN of the source table to restore. Must be supplied + for cross-region restores. + type: string restoreToLatestTime: description: If set, restores table to the most recent point-in-time recovery point. @@ -1081,6 +1089,10 @@ spec: description: Name of the table to restore. Must match the name of an existing table. type: string + restoreSourceTableArn: + description: ARN of the source table to restore. Must be supplied + for cross-region restores. + type: string restoreToLatestTime: description: If set, restores table to the most recent point-in-time recovery point. diff --git a/package/crds/ecs.aws.upbound.io_clusters.yaml b/package/crds/ecs.aws.upbound.io_clusters.yaml index 1bf24e1bc9..bffae0e794 100644 --- a/package/crds/ecs.aws.upbound.io_clusters.yaml +++ b/package/crds/ecs.aws.upbound.io_clusters.yaml @@ -124,6 +124,21 @@ spec: type: string type: object type: array + managedStorageConfiguration: + description: Details of the managed storage configuration. + See managed_storage_configuration Block for details. + items: + properties: + fargateEphemeralStorageKmsKeyId: + description: AWS Key Management Service key ID for + the Fargate ephemeral storage. + type: string + kmsKeyId: + description: AWS Key Management Service key ID to + encrypt the managed storage. + type: string + type: object + type: array type: object type: array region: @@ -232,6 +247,21 @@ spec: type: string type: object type: array + managedStorageConfiguration: + description: Details of the managed storage configuration. + See managed_storage_configuration Block for details. + items: + properties: + fargateEphemeralStorageKmsKeyId: + description: AWS Key Management Service key ID for + the Fargate ephemeral storage. + type: string + kmsKeyId: + description: AWS Key Management Service key ID to + encrypt the managed storage. + type: string + type: object + type: array type: object type: array serviceConnectDefaults: @@ -496,6 +526,21 @@ spec: type: string type: object type: array + managedStorageConfiguration: + description: Details of the managed storage configuration. + See managed_storage_configuration Block for details. + items: + properties: + fargateEphemeralStorageKmsKeyId: + description: AWS Key Management Service key ID for + the Fargate ephemeral storage. + type: string + kmsKeyId: + description: AWS Key Management Service key ID to + encrypt the managed storage. + type: string + type: object + type: array type: object type: array id: diff --git a/package/crds/ecs.aws.upbound.io_services.yaml b/package/crds/ecs.aws.upbound.io_services.yaml index 88517772c9..45610d5876 100644 --- a/package/crds/ecs.aws.upbound.io_services.yaml +++ b/package/crds/ecs.aws.upbound.io_services.yaml @@ -247,6 +247,11 @@ spec: description: Specifies whether to enable Amazon ECS Exec for the tasks within the service. type: boolean + forceDelete: + description: Enable to delete a service even if it wasn't scaled + down to zero tasks. It's only necessary to use this if the service + uses the REPLICA scheduling strategy. + type: boolean forceNewDeployment: description: Enable to force a new task deployment of the service. This can be used to update tasks to use a newer Docker image @@ -1202,6 +1207,11 @@ spec: description: Specifies whether to enable Amazon ECS Exec for the tasks within the service. type: boolean + forceDelete: + description: Enable to delete a service even if it wasn't scaled + down to zero tasks. It's only necessary to use this if the service + uses the REPLICA scheduling strategy. + type: boolean forceNewDeployment: description: Enable to force a new task deployment of the service. This can be used to update tasks to use a newer Docker image @@ -2236,6 +2246,11 @@ spec: description: Specifies whether to enable Amazon ECS Exec for the tasks within the service. type: boolean + forceDelete: + description: Enable to delete a service even if it wasn't scaled + down to zero tasks. It's only necessary to use this if the service + uses the REPLICA scheduling strategy. + type: boolean forceNewDeployment: description: Enable to force a new task deployment of the service. This can be used to update tasks to use a newer Docker image diff --git a/package/crds/eks.aws.upbound.io_clusters.yaml b/package/crds/eks.aws.upbound.io_clusters.yaml index 7e8806c27a..fdb2e2d4b5 100644 --- a/package/crds/eks.aws.upbound.io_clusters.yaml +++ b/package/crds/eks.aws.upbound.io_clusters.yaml @@ -289,6 +289,20 @@ spec: description: Key-value map of resource tags. type: object x-kubernetes-map-type: granular + upgradePolicy: + description: Configuration block for the support policy to use + for the cluster. See upgrade_policy for details. + items: + properties: + supportType: + description: Support type to use for the cluster. If the + cluster is set to EXTENDED, it will enter extended support + at the end of standard support. If the cluster is set + to STANDARD, it will be automatically upgraded at the + end of standard support. Valid values are EXTENDED, STANDARD + type: string + type: object + type: array version: description: – Desired Kubernetes master version. If you do not specify a value, the latest available version at resource creation @@ -734,6 +748,20 @@ spec: description: Key-value map of resource tags. type: object x-kubernetes-map-type: granular + upgradePolicy: + description: Configuration block for the support policy to use + for the cluster. See upgrade_policy for details. + items: + properties: + supportType: + description: Support type to use for the cluster. If the + cluster is set to EXTENDED, it will enter extended support + at the end of standard support. If the cluster is set + to STANDARD, it will be automatically upgraded at the + end of standard support. Valid values are EXTENDED, STANDARD + type: string + type: object + type: array version: description: – Desired Kubernetes master version. If you do not specify a value, the latest available version at resource creation @@ -1339,6 +1367,20 @@ spec: inherited from the provider default_tags configuration block. type: object x-kubernetes-map-type: granular + upgradePolicy: + description: Configuration block for the support policy to use + for the cluster. See upgrade_policy for details. + items: + properties: + supportType: + description: Support type to use for the cluster. If the + cluster is set to EXTENDED, it will enter extended support + at the end of standard support. If the cluster is set + to STANDARD, it will be automatically upgraded at the + end of standard support. Valid values are EXTENDED, STANDARD + type: string + type: object + type: array version: description: – Desired Kubernetes master version. If you do not specify a value, the latest available version at resource creation diff --git a/package/crds/elasticache.aws.upbound.io_replicationgroups.yaml b/package/crds/elasticache.aws.upbound.io_replicationgroups.yaml index 1c5f326c2f..b85ef04913 100644 --- a/package/crds/elasticache.aws.upbound.io_replicationgroups.yaml +++ b/package/crds/elasticache.aws.upbound.io_replicationgroups.yaml @@ -98,6 +98,15 @@ spec: - name - namespace type: object + authTokenUpdateStrategy: + description: Strategy to use when updating the auth_token. Valid + values are SET, ROTATE, and DELETE. Defaults to ROTATE. + type: string + autoGenerateAuthToken: + description: |- + Password used to access a password protected server. Can be specified only if transit_encryption_enabled = true. + If true, the auth token will be auto-generated and stored in the Secret referenced by the authTokenSecretRef field. + type: boolean autoMinorVersionUpgrade: description: |- Specifies whether minor version engine upgrades will be applied automatically to the underlying Cache Cluster instances during the maintenance window. @@ -176,6 +185,86 @@ spec: is set, the num_node_groups parameter (or the num_node_groups parameter of the deprecated cluster_mode block) cannot be set. type: string + globalReplicationGroupIdRef: + description: Reference to a GlobalReplicationGroup in elasticache + to populate globalReplicationGroupId. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + globalReplicationGroupIdSelector: + description: Selector for a GlobalReplicationGroup in elasticache + to populate globalReplicationGroupId. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object + ipDiscovery: + description: The IP version to advertise in the discovery protocol. + Valid values are ipv4 or ipv6. + type: string kmsKeyId: description: The ARN of the key that you wish to use if encrypting at rest. If not supplied, uses service managed encryption. Can @@ -287,6 +376,10 @@ spec: the replication group. If true, automatic_failover_enabled must also be enabled. Defaults to false. type: boolean + networkType: + description: The IP versions for cache cluster connections. Valid + values are ipv4, ipv6 or dual_stack. + type: string nodeType: description: Instance class to be used. See AWS documentation for information on supported node types and guidance on selecting @@ -554,6 +647,13 @@ spec: transitEncryptionEnabled: description: Whether to enable encryption in transit. type: boolean + transitEncryptionMode: + description: |- + A setting that enables clients to migrate to in-transit encryption with no downtime. + Valid values are preferred and required. + When enabling encryption on an existing replication group, this must first be set to preferred before setting it to required in a subsequent apply. + See the TransitEncryptionMode field in the CreateReplicationGroup API documentation for additional details. + type: string userGroupIds: description: 'User Group ID to associate with the replication group. Only a maximum of one (1) user group ID is valid. NOTE: @@ -587,6 +687,29 @@ spec: atRestEncryptionEnabled: description: Whether to enable encryption at rest. type: boolean + authTokenSecretRef: + description: |- + Password used to access a password protected server. Can be specified only if transit_encryption_enabled = true. + If you set autoGenerateAuthToken to true, the Secret referenced here will be created or updated with generated auth token if it does not already contain one. + properties: + key: + description: The key to select. + type: string + name: + description: Name of the secret. + type: string + namespace: + description: Namespace of the secret. + type: string + required: + - key + - name + - namespace + type: object + authTokenUpdateStrategy: + description: Strategy to use when updating the auth_token. Valid + values are SET, ROTATE, and DELETE. Defaults to ROTATE. + type: string autoMinorVersionUpgrade: description: |- Specifies whether minor version engine upgrades will be applied automatically to the underlying Cache Cluster instances during the maintenance window. @@ -665,6 +788,86 @@ spec: is set, the num_node_groups parameter (or the num_node_groups parameter of the deprecated cluster_mode block) cannot be set. type: string + globalReplicationGroupIdRef: + description: Reference to a GlobalReplicationGroup in elasticache + to populate globalReplicationGroupId. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + globalReplicationGroupIdSelector: + description: Selector for a GlobalReplicationGroup in elasticache + to populate globalReplicationGroupId. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object + ipDiscovery: + description: The IP version to advertise in the discovery protocol. + Valid values are ipv4 or ipv6. + type: string kmsKeyId: description: The ARN of the key that you wish to use if encrypting at rest. If not supplied, uses service managed encryption. Can @@ -776,6 +979,10 @@ spec: the replication group. If true, automatic_failover_enabled must also be enabled. Defaults to false. type: boolean + networkType: + description: The IP versions for cache cluster connections. Valid + values are ipv4, ipv6 or dual_stack. + type: string nodeType: description: Instance class to be used. See AWS documentation for information on supported node types and guidance on selecting @@ -1039,6 +1246,13 @@ spec: transitEncryptionEnabled: description: Whether to enable encryption in transit. type: boolean + transitEncryptionMode: + description: |- + A setting that enables clients to migrate to in-transit encryption with no downtime. + Valid values are preferred and required. + When enabling encryption on an existing replication group, this must first be set to preferred before setting it to required in a subsequent apply. + See the TransitEncryptionMode field in the CreateReplicationGroup API documentation for additional details. + type: string userGroupIds: description: 'User Group ID to associate with the replication group. Only a maximum of one (1) user group ID is valid. NOTE: @@ -1232,6 +1446,10 @@ spec: atRestEncryptionEnabled: description: Whether to enable encryption at rest. type: boolean + authTokenUpdateStrategy: + description: Strategy to use when updating the auth_token. Valid + values are SET, ROTATE, and DELETE. Defaults to ROTATE. + type: string autoMinorVersionUpgrade: description: |- Specifies whether minor version engine upgrades will be applied automatically to the underlying Cache Cluster instances during the maintenance window. @@ -1325,6 +1543,10 @@ spec: id: description: ID of the ElastiCache Replication Group. type: string + ipDiscovery: + description: The IP version to advertise in the discovery protocol. + Valid values are ipv4 or ipv6. + type: string kmsKeyId: description: The ARN of the key that you wish to use if encrypting at rest. If not supplied, uses service managed encryption. Can @@ -1369,6 +1591,10 @@ spec: the replication group. If true, automatic_failover_enabled must also be enabled. Defaults to false. type: boolean + networkType: + description: The IP versions for cache cluster connections. Valid + values are ipv4, ipv6 or dual_stack. + type: string nodeType: description: Instance class to be used. See AWS documentation for information on supported node types and guidance on selecting @@ -1495,6 +1721,13 @@ spec: transitEncryptionEnabled: description: Whether to enable encryption in transit. type: boolean + transitEncryptionMode: + description: |- + A setting that enables clients to migrate to in-transit encryption with no downtime. + Valid values are preferred and required. + When enabling encryption on an existing replication group, this must first be set to preferred before setting it to required in a subsequent apply. + See the TransitEncryptionMode field in the CreateReplicationGroup API documentation for additional details. + type: string userGroupIds: description: 'User Group ID to associate with the replication group. Only a maximum of one (1) user group ID is valid. NOTE: diff --git a/package/crds/elbv2.aws.upbound.io_lbtargetgroups.yaml b/package/crds/elbv2.aws.upbound.io_lbtargetgroups.yaml index 0e5a69282c..924b113c86 100644 --- a/package/crds/elbv2.aws.upbound.io_lbtargetgroups.yaml +++ b/package/crds/elbv2.aws.upbound.io_lbtargetgroups.yaml @@ -269,6 +269,51 @@ spec: type: string type: object type: array + targetGroupHealth: + description: Target health requirements block. See target_group_health + for more information. + properties: + dnsFailover: + description: Block to configure DNS Failover requirements. + See DNS Failover below for details on attributes. + properties: + minimumHealthyTargetsCount: + description: The minimum number of targets that must be + healthy. If the number of healthy targets is below this + value, mark the zone as unhealthy in DNS, so that traffic + is routed only to healthy zones. The possible values + are off or an integer from 1 to the maximum number of + targets. The default is off. + type: string + minimumHealthyTargetsPercentage: + description: The minimum percentage of targets that must + be healthy. If the percentage of healthy targets is + below this value, mark the zone as unhealthy in DNS, + so that traffic is routed only to healthy zones. The + possible values are off or an integer from 1 to 100. + The default is off. + type: string + type: object + unhealthyStateRouting: + description: Block to configure Unhealthy State Routing requirements. + See Unhealthy State Routing below for details on attributes. + properties: + minimumHealthyTargetsCount: + description: The minimum number of targets that must be + healthy. If the number of healthy targets is below this + value, send traffic to all targets, including unhealthy + targets. The possible values are 1 to the maximum number + of targets. The default is 1. + type: number + minimumHealthyTargetsPercentage: + description: The minimum percentage of targets that must + be healthy. If the percentage of healthy targets is + below this value, send traffic to all targets, including + unhealthy targets. The possible values are off or an + integer from 1 to 100. The default is off. + type: string + type: object + type: object targetHealthState: description: Target health state block. Only applicable for Network Load Balancer target groups when protocol is TCP or TLS. See @@ -280,6 +325,12 @@ spec: connections to unhealthy targets. Possible values are true or false. Default: true.' type: boolean + unhealthyDrainingInterval: + description: 'Indicates the time to wait for in-flight requests + to complete when a target becomes unhealthy. The range + is 0-360000. This value has to be set only if enable_unhealthy_connection_termination + is set to false. Default: 0.' + type: number type: object type: array targetType: @@ -575,6 +626,51 @@ spec: type: string type: object type: array + targetGroupHealth: + description: Target health requirements block. See target_group_health + for more information. + properties: + dnsFailover: + description: Block to configure DNS Failover requirements. + See DNS Failover below for details on attributes. + properties: + minimumHealthyTargetsCount: + description: The minimum number of targets that must be + healthy. If the number of healthy targets is below this + value, mark the zone as unhealthy in DNS, so that traffic + is routed only to healthy zones. The possible values + are off or an integer from 1 to the maximum number of + targets. The default is off. + type: string + minimumHealthyTargetsPercentage: + description: The minimum percentage of targets that must + be healthy. If the percentage of healthy targets is + below this value, mark the zone as unhealthy in DNS, + so that traffic is routed only to healthy zones. The + possible values are off or an integer from 1 to 100. + The default is off. + type: string + type: object + unhealthyStateRouting: + description: Block to configure Unhealthy State Routing requirements. + See Unhealthy State Routing below for details on attributes. + properties: + minimumHealthyTargetsCount: + description: The minimum number of targets that must be + healthy. If the number of healthy targets is below this + value, send traffic to all targets, including unhealthy + targets. The possible values are 1 to the maximum number + of targets. The default is 1. + type: number + minimumHealthyTargetsPercentage: + description: The minimum percentage of targets that must + be healthy. If the percentage of healthy targets is + below this value, send traffic to all targets, including + unhealthy targets. The possible values are off or an + integer from 1 to 100. The default is off. + type: string + type: object + type: object targetHealthState: description: Target health state block. Only applicable for Network Load Balancer target groups when protocol is TCP or TLS. See @@ -586,6 +682,12 @@ spec: connections to unhealthy targets. Possible values are true or false. Default: true.' type: boolean + unhealthyDrainingInterval: + description: 'Indicates the time to wait for in-flight requests + to complete when a target becomes unhealthy. The range + is 0-360000. This value has to be set only if enable_unhealthy_connection_termination + is set to false. Default: 0.' + type: number type: object type: array targetType: @@ -1067,6 +1169,51 @@ spec: type: string type: object type: array + targetGroupHealth: + description: Target health requirements block. See target_group_health + for more information. + properties: + dnsFailover: + description: Block to configure DNS Failover requirements. + See DNS Failover below for details on attributes. + properties: + minimumHealthyTargetsCount: + description: The minimum number of targets that must be + healthy. If the number of healthy targets is below this + value, mark the zone as unhealthy in DNS, so that traffic + is routed only to healthy zones. The possible values + are off or an integer from 1 to the maximum number of + targets. The default is off. + type: string + minimumHealthyTargetsPercentage: + description: The minimum percentage of targets that must + be healthy. If the percentage of healthy targets is + below this value, mark the zone as unhealthy in DNS, + so that traffic is routed only to healthy zones. The + possible values are off or an integer from 1 to 100. + The default is off. + type: string + type: object + unhealthyStateRouting: + description: Block to configure Unhealthy State Routing requirements. + See Unhealthy State Routing below for details on attributes. + properties: + minimumHealthyTargetsCount: + description: The minimum number of targets that must be + healthy. If the number of healthy targets is below this + value, send traffic to all targets, including unhealthy + targets. The possible values are 1 to the maximum number + of targets. The default is 1. + type: number + minimumHealthyTargetsPercentage: + description: The minimum percentage of targets that must + be healthy. If the percentage of healthy targets is + below this value, send traffic to all targets, including + unhealthy targets. The possible values are off or an + integer from 1 to 100. The default is off. + type: string + type: object + type: object targetHealthState: description: Target health state block. Only applicable for Network Load Balancer target groups when protocol is TCP or TLS. See @@ -1078,6 +1225,12 @@ spec: connections to unhealthy targets. Possible values are true or false. Default: true.' type: boolean + unhealthyDrainingInterval: + description: 'Indicates the time to wait for in-flight requests + to complete when a target becomes unhealthy. The range + is 0-360000. This value has to be set only if enable_unhealthy_connection_termination + is set to false. Default: 0.' + type: number type: object type: array targetType: diff --git a/package/crds/emrserverless.aws.upbound.io_applications.yaml b/package/crds/emrserverless.aws.upbound.io_applications.yaml index 41b4177c78..26de0dea8e 100644 --- a/package/crds/emrserverless.aws.upbound.io_applications.yaml +++ b/package/crds/emrserverless.aws.upbound.io_applications.yaml @@ -156,6 +156,21 @@ spec: type: string type: object type: array + interactiveConfiguration: + description: – Enables the interactive use cases to use when + running an application. + items: + properties: + livyEndpointEnabled: + description: Enables an Apache Livy endpoint that you can + connect to and run interactive jobs. + type: boolean + studioEnabled: + description: Enables you to connect an application to Amazon + EMR Studio to run interactive workloads in a notebook. + type: boolean + type: object + type: array maximumCapacity: description: – The maximum capacity to allocate when the application is created. This is cumulative across all workers at any given @@ -313,6 +328,21 @@ spec: type: string type: object type: array + interactiveConfiguration: + description: – Enables the interactive use cases to use when + running an application. + items: + properties: + livyEndpointEnabled: + description: Enables an Apache Livy endpoint that you can + connect to and run interactive jobs. + type: boolean + studioEnabled: + description: Enables you to connect an application to Amazon + EMR Studio to run interactive workloads in a notebook. + type: boolean + type: object + type: array maximumCapacity: description: – The maximum capacity to allocate when the application is created. This is cumulative across all workers at any given @@ -642,6 +672,21 @@ spec: type: string type: object type: array + interactiveConfiguration: + description: – Enables the interactive use cases to use when + running an application. + items: + properties: + livyEndpointEnabled: + description: Enables an Apache Livy endpoint that you can + connect to and run interactive jobs. + type: boolean + studioEnabled: + description: Enables you to connect an application to Amazon + EMR Studio to run interactive workloads in a notebook. + type: boolean + type: object + type: array maximumCapacity: description: – The maximum capacity to allocate when the application is created. This is cumulative across all workers at any given diff --git a/package/crds/firehose.aws.upbound.io_deliverystreams.yaml b/package/crds/firehose.aws.upbound.io_deliverystreams.yaml index 217c8f82e5..cf25992e7b 100644 --- a/package/crds/firehose.aws.upbound.io_deliverystreams.yaml +++ b/package/crds/firehose.aws.upbound.io_deliverystreams.yaml @@ -1985,6 +1985,25 @@ spec: type: object type: object type: array + secretsManagerConfiguration: + description: The Secret Manager Configuration. See secrets_manager_configuration + block below for details. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object url: description: The HTTP endpoint URL to which Kinesis Firehose sends your data. @@ -3226,7 +3245,8 @@ spec: that the s3 bucket will copy to. type: string passwordSecretRef: - description: The password for the username above. + description: The password for the username above. This value + is required if secrets_manager_configuration is not provided. properties: key: description: The key to select. @@ -3832,6 +3852,26 @@ spec: type: object type: object type: array + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object username: description: The username that the firehose delivery stream will assume. It is strongly recommended that the username @@ -3874,6 +3914,16 @@ spec: description: 'The URL of the Snowflake account. Format: https://[account_identifier].snowflakecomputing.com.' type: string + bufferingInterval: + description: Buffer incoming data for the specified period + of time, in seconds between 0 to 900, before delivering + it to the destination. The default value is 0s. + type: number + bufferingSize: + description: Buffer incoming data to the specified size, + in MBs between 1 to 128, before delivering it to the destination. The + default value is 1MB. + type: number cloudwatchLoggingOptions: description: The CloudWatch Logging Options for the delivery stream. See cloudwatch_logging_options block below for @@ -3924,7 +3974,8 @@ spec: description: The name of the metadata column. type: string privateKeySecretRef: - description: The private key for authentication. + description: The private key for authentication. This value + is required if secrets_manager_configuration is not provided. properties: key: description: The key to select. @@ -4306,6 +4357,26 @@ spec: schema: description: The Snowflake schema name. type: string + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object snowflakeRoleConfiguration: description: The configuration for Snowflake role. items: @@ -4387,7 +4458,8 @@ spec: type: string hecTokenSecretRef: description: The GUID that you obtain from your Splunk cluster - when you create a new HEC endpoint. + when you create a new HEC endpoint. This value is required + if secrets_manager_configuration is not provided. properties: key: description: The key to select. @@ -4691,6 +4763,26 @@ spec: type: object type: object type: array + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object type: object type: array tags: @@ -6627,6 +6719,25 @@ spec: type: object type: object type: array + secretsManagerConfiguration: + description: The Secret Manager Configuration. See secrets_manager_configuration + block below for details. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object url: description: The HTTP endpoint URL to which Kinesis Firehose sends your data. @@ -7868,7 +7979,8 @@ spec: that the s3 bucket will copy to. type: string passwordSecretRef: - description: The password for the username above. + description: The password for the username above. This value + is required if secrets_manager_configuration is not provided. properties: key: description: The key to select. @@ -8474,6 +8586,26 @@ spec: type: object type: object type: array + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object username: description: The username that the firehose delivery stream will assume. It is strongly recommended that the username @@ -8481,8 +8613,6 @@ spec: Firehose purposes, and that the permissions for the account are restricted for Amazon Redshift INSERT permissions. type: string - required: - - passwordSecretRef type: object type: array serverSideEncryption: @@ -8513,6 +8643,16 @@ spec: description: 'The URL of the Snowflake account. Format: https://[account_identifier].snowflakecomputing.com.' type: string + bufferingInterval: + description: Buffer incoming data for the specified period + of time, in seconds between 0 to 900, before delivering + it to the destination. The default value is 0s. + type: number + bufferingSize: + description: Buffer incoming data to the specified size, + in MBs between 1 to 128, before delivering it to the destination. The + default value is 1MB. + type: number cloudwatchLoggingOptions: description: The CloudWatch Logging Options for the delivery stream. See cloudwatch_logging_options block below for @@ -8563,7 +8703,8 @@ spec: description: The name of the metadata column. type: string privateKeySecretRef: - description: The private key for authentication. + description: The private key for authentication. This value + is required if secrets_manager_configuration is not provided. properties: key: description: The key to select. @@ -8945,6 +9086,26 @@ spec: schema: description: The Snowflake schema name. type: string + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object snowflakeRoleConfiguration: description: The configuration for Snowflake role. items: @@ -8974,8 +9135,6 @@ spec: user: description: The user for authentication. type: string - required: - - privateKeySecretRef type: object type: array splunkConfiguration: @@ -9028,7 +9187,8 @@ spec: type: string hecTokenSecretRef: description: The GUID that you obtain from your Splunk cluster - when you create a new HEC endpoint. + when you create a new HEC endpoint. This value is required + if secrets_manager_configuration is not provided. properties: key: description: The key to select. @@ -9332,8 +9492,26 @@ spec: type: object type: object type: array - required: - - hecTokenSecretRef + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object type: object type: array tags: @@ -10522,6 +10700,25 @@ spec: type: string type: object type: array + secretsManagerConfiguration: + description: The Secret Manager Configuration. See secrets_manager_configuration + block below for details. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object url: description: The HTTP endpoint URL to which Kinesis Firehose sends your data. @@ -11300,6 +11497,26 @@ spec: type: string type: object type: array + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object username: description: The username that the firehose delivery stream will assume. It is strongly recommended that the username @@ -11337,6 +11554,16 @@ spec: description: 'The URL of the Snowflake account. Format: https://[account_identifier].snowflakecomputing.com.' type: string + bufferingInterval: + description: Buffer incoming data for the specified period + of time, in seconds between 0 to 900, before delivering + it to the destination. The default value is 0s. + type: number + bufferingSize: + description: Buffer incoming data to the specified size, + in MBs between 1 to 128, before delivering it to the destination. The + default value is 1MB. + type: number cloudwatchLoggingOptions: description: The CloudWatch Logging Options for the delivery stream. See cloudwatch_logging_options block below for @@ -11509,6 +11736,26 @@ spec: schema: description: The Snowflake schema name. type: string + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object snowflakeRoleConfiguration: description: The configuration for Snowflake role. items: @@ -11724,6 +11971,26 @@ spec: type: string type: object type: array + secretsManagerConfiguration: + description: The Secrets Manager configuration. See secrets_manager_configuration + block below for details. This value is required if user + and private_key are not provided. + properties: + enabled: + description: Enables or disables the logging. Defaults + to false. + type: boolean + roleArn: + description: The ARN of the IAM role to be assumed by + Firehose for calling the Amazon EC2 configuration + API and for creating network interfaces. Make sure + role has necessary IAM permissions + type: string + secretArn: + description: The ARN of the Secrets Manager secret. + This value is required if enabled is true. + type: string + type: object type: object type: array tags: diff --git a/package/crds/fis.aws.upbound.io_experimenttemplates.yaml b/package/crds/fis.aws.upbound.io_experimenttemplates.yaml index c161d82420..c33fbf13a0 100644 --- a/package/crds/fis.aws.upbound.io_experimenttemplates.yaml +++ b/package/crds/fis.aws.upbound.io_experimenttemplates.yaml @@ -126,6 +126,21 @@ spec: description: description: Description for the experiment template. type: string + experimentOptions: + description: The experiment options for the experiment template. + See experiment_options below for more details! + items: + properties: + accountTargeting: + description: Specifies the account targeting setting for + experiment options. Supports single-account and multi-account. + type: string + emptyTargetResolutionMode: + description: Specifies the empty target resolution mode + for experiment options. Supports fail and skip. + type: string + type: object + type: array logConfiguration: description: The configuration for experiment logging. See below. items: @@ -396,6 +411,21 @@ spec: description: description: Description for the experiment template. type: string + experimentOptions: + description: The experiment options for the experiment template. + See experiment_options below for more details! + items: + properties: + accountTargeting: + description: Specifies the account targeting setting for + experiment options. Supports single-account and multi-account. + type: string + emptyTargetResolutionMode: + description: Specifies the empty target resolution mode + for experiment options. Supports fail and skip. + type: string + type: object + type: array logConfiguration: description: The configuration for experiment logging. See below. items: @@ -832,6 +862,21 @@ spec: description: description: Description for the experiment template. type: string + experimentOptions: + description: The experiment options for the experiment template. + See experiment_options below for more details! + items: + properties: + accountTargeting: + description: Specifies the account targeting setting for + experiment options. Supports single-account and multi-account. + type: string + emptyTargetResolutionMode: + description: Specifies the empty target resolution mode + for experiment options. Supports fail and skip. + type: string + type: object + type: array id: description: Experiment Template ID. type: string diff --git a/package/crds/fsx.aws.upbound.io_lustrefilesystems.yaml b/package/crds/fsx.aws.upbound.io_lustrefilesystems.yaml index 86f2e4c8f5..160b452226 100644 --- a/package/crds/fsx.aws.upbound.io_lustrefilesystems.yaml +++ b/package/crds/fsx.aws.upbound.io_lustrefilesystems.yaml @@ -129,6 +129,13 @@ spec: and PERSISTENT_1 deployment types. Valid values for 2.12 include all deployment types. type: string + finalBackupTags: + additionalProperties: + type: string + description: A map of tags to apply to the file system's final + backup. + type: object + x-kubernetes-map-type: granular importPath: description: S3 URI (with optional prefix) that you're using as the data repository for your FSx for Lustre file system. For @@ -392,6 +399,12 @@ spec: type: string type: array x-kubernetes-list-type: set + skipFinalBackup: + description: When enabled, will skip the default final backup + taken when the file system is deleted. This configuration must + be applied separately before attempting to delete the resource + to have the desired behavior. Defaults to true. + type: boolean storageCapacity: description: The storage capacity (GiB) of the file system. Minimum of 1200. See more details at Allowed values for Fsx storage @@ -573,6 +586,13 @@ spec: and PERSISTENT_1 deployment types. Valid values for 2.12 include all deployment types. type: string + finalBackupTags: + additionalProperties: + type: string + description: A map of tags to apply to the file system's final + backup. + type: object + x-kubernetes-map-type: granular importPath: description: S3 URI (with optional prefix) that you're using as the data repository for your FSx for Lustre file system. For @@ -832,6 +852,12 @@ spec: type: string type: array x-kubernetes-list-type: set + skipFinalBackup: + description: When enabled, will skip the default final backup + taken when the file system is deleted. This configuration must + be applied separately before attempting to delete the resource + to have the desired behavior. Defaults to true. + type: boolean storageCapacity: description: The storage capacity (GiB) of the file system. Minimum of 1200. See more details at Allowed values for Fsx storage @@ -1176,6 +1202,13 @@ spec: and PERSISTENT_1 deployment types. Valid values for 2.12 include all deployment types. type: string + finalBackupTags: + additionalProperties: + type: string + description: A map of tags to apply to the file system's final + backup. + type: object + x-kubernetes-map-type: granular id: description: Identifier of the file system, e.g., fs-12345678 type: string @@ -1300,6 +1333,12 @@ spec: type: string type: array x-kubernetes-list-type: set + skipFinalBackup: + description: When enabled, will skip the default final backup + taken when the file system is deleted. This configuration must + be applied separately before attempting to delete the resource + to have the desired behavior. Defaults to true. + type: boolean storageCapacity: description: The storage capacity (GiB) of the file system. Minimum of 1200. See more details at Allowed values for Fsx storage diff --git a/package/crds/fsx.aws.upbound.io_windowsfilesystems.yaml b/package/crds/fsx.aws.upbound.io_windowsfilesystems.yaml index d1b2c50927..e1d8da0f0e 100644 --- a/package/crds/fsx.aws.upbound.io_windowsfilesystems.yaml +++ b/package/crds/fsx.aws.upbound.io_windowsfilesystems.yaml @@ -233,6 +233,13 @@ spec: type: string type: object type: array + finalBackupTags: + additionalProperties: + type: string + description: A map of tags to apply to the file system's final + backup. + type: object + x-kubernetes-map-type: granular kmsKeyId: description: ARN for the KMS Key to encrypt the file system at rest. Defaults to an AWS managed KMS Key. @@ -764,6 +771,13 @@ spec: type: string type: object type: array + finalBackupTags: + additionalProperties: + type: string + description: A map of tags to apply to the file system's final + backup. + type: object + x-kubernetes-map-type: granular kmsKeyId: description: ARN for the KMS Key to encrypt the file system at rest. Defaults to an AWS managed KMS Key. @@ -1388,6 +1402,13 @@ spec: description: DNS name for the file system, e.g., fs-12345678.corp.example.com (domain name matching the Active Directory domain name) type: string + finalBackupTags: + additionalProperties: + type: string + description: A map of tags to apply to the file system's final + backup. + type: object + x-kubernetes-map-type: granular id: description: Identifier of the file system (e.g. fs-12345678). type: string diff --git a/package/crds/glue.aws.upbound.io_jobs.yaml b/package/crds/glue.aws.upbound.io_jobs.yaml index ca51e02f6b..fcd97a3027 100644 --- a/package/crds/glue.aws.upbound.io_jobs.yaml +++ b/package/crds/glue.aws.upbound.io_jobs.yaml @@ -139,6 +139,12 @@ spec: jobs should set this to 4.0 or greater. For information about available versions, see the AWS Glue Release Notes. type: string + jobRunQueuingEnabled: + description: Specifies whether job run queuing is enabled for + the job runs for this job. A value of true means job run queuing + is enabled for the job runs. If false or not populated, the + job runs will not be considered for queueing. + type: boolean maintenanceWindow: description: – Specifies the day of the week and hour for the maintenance window for streaming jobs. @@ -359,6 +365,12 @@ spec: jobs should set this to 4.0 or greater. For information about available versions, see the AWS Glue Release Notes. type: string + jobRunQueuingEnabled: + description: Specifies whether job run queuing is enabled for + the job runs for this job. A value of true means job run queuing + is enabled for the job runs. If false or not populated, the + job runs will not be considered for queueing. + type: boolean maintenanceWindow: description: – Specifies the day of the week and hour for the maintenance window for streaming jobs. @@ -743,6 +755,12 @@ spec: id: description: Job name type: string + jobRunQueuingEnabled: + description: Specifies whether job run queuing is enabled for + the job runs for this job. A value of true means job run queuing + is enabled for the job runs. If false or not populated, the + job runs will not be considered for queueing. + type: boolean maintenanceWindow: description: – Specifies the day of the week and hour for the maintenance window for streaming jobs. diff --git a/package/crds/imagebuilder.aws.upbound.io_imagepipelines.yaml b/package/crds/imagebuilder.aws.upbound.io_imagepipelines.yaml index 1c57088401..8562dc4cac 100644 --- a/package/crds/imagebuilder.aws.upbound.io_imagepipelines.yaml +++ b/package/crds/imagebuilder.aws.upbound.io_imagepipelines.yaml @@ -87,6 +87,10 @@ spec: description: Whether additional information about the image being created is collected. Defaults to true. type: boolean + executionRole: + description: Amazon Resource Name (ARN) of the service-linked + role to be used by Image Builder to execute workflows. + type: string imageRecipeArn: description: Amazon Resource Name (ARN) of the image recipe. type: string @@ -332,6 +336,37 @@ spec: description: Key-value map of resource tags. type: object x-kubernetes-map-type: granular + workflow: + description: Configuration block with the workflow configuration. + Detailed below. + items: + properties: + onFailure: + description: The action to take if the workflow fails. Must + be one of CONTINUE or ABORT. + type: string + parallelGroup: + description: The parallel group in which to run a test Workflow. + type: string + parameter: + description: Configuration block for the workflow parameters. + Detailed below. + items: + properties: + name: + description: The name of the Workflow parameter. + type: string + value: + description: The value of the Workflow parameter. + type: string + type: object + type: array + workflowArn: + description: Amazon Resource Name (ARN) of the Image Builder + Workflow. + type: string + type: object + type: array required: - region type: object @@ -362,6 +397,10 @@ spec: description: Whether additional information about the image being created is collected. Defaults to true. type: boolean + executionRole: + description: Amazon Resource Name (ARN) of the service-linked + role to be used by Image Builder to execute workflows. + type: string imageRecipeArn: description: Amazon Resource Name (ARN) of the image recipe. type: string @@ -603,6 +642,37 @@ spec: description: Key-value map of resource tags. type: object x-kubernetes-map-type: granular + workflow: + description: Configuration block with the workflow configuration. + Detailed below. + items: + properties: + onFailure: + description: The action to take if the workflow fails. Must + be one of CONTINUE or ABORT. + type: string + parallelGroup: + description: The parallel group in which to run a test Workflow. + type: string + parameter: + description: Configuration block for the workflow parameters. + Detailed below. + items: + properties: + name: + description: The name of the Workflow parameter. + type: string + value: + description: The value of the Workflow parameter. + type: string + type: object + type: array + workflowArn: + description: Amazon Resource Name (ARN) of the Image Builder + Workflow. + type: string + type: object + type: array type: object managementPolicies: default: @@ -810,6 +880,10 @@ spec: description: Whether additional information about the image being created is collected. Defaults to true. type: boolean + executionRole: + description: Amazon Resource Name (ARN) of the service-linked + role to be used by Image Builder to execute workflows. + type: string id: type: string imageRecipeArn: @@ -912,6 +986,37 @@ spec: block. type: object x-kubernetes-map-type: granular + workflow: + description: Configuration block with the workflow configuration. + Detailed below. + items: + properties: + onFailure: + description: The action to take if the workflow fails. Must + be one of CONTINUE or ABORT. + type: string + parallelGroup: + description: The parallel group in which to run a test Workflow. + type: string + parameter: + description: Configuration block for the workflow parameters. + Detailed below. + items: + properties: + name: + description: The name of the Workflow parameter. + type: string + value: + description: The value of the Workflow parameter. + type: string + type: object + type: array + workflowArn: + description: Amazon Resource Name (ARN) of the Image Builder + Workflow. + type: string + type: object + type: array type: object conditions: description: Conditions of the resource. diff --git a/package/crds/lambda.aws.upbound.io_eventsourcemappings.yaml b/package/crds/lambda.aws.upbound.io_eventsourcemappings.yaml index 92e118f53d..38382c9a68 100644 --- a/package/crds/lambda.aws.upbound.io_eventsourcemappings.yaml +++ b/package/crds/lambda.aws.upbound.io_eventsourcemappings.yaml @@ -260,6 +260,85 @@ spec: type: string type: array x-kubernetes-list-type: set + kmsKeyArn: + description: The ARN of the Key Management Service (KMS) customer + managed key that Lambda uses to encrypt your function's filter + criteria. + type: string + kmsKeyArnRef: + description: Reference to a Key in kms to populate kmsKeyArn. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + kmsKeyArnSelector: + description: Selector for a Key in kms to populate kmsKeyArn. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object maximumBatchingWindowInSeconds: description: The maximum amount of time to gather records before invoking the function, in seconds (between 0 and 300). Records @@ -595,6 +674,85 @@ spec: type: string type: array x-kubernetes-list-type: set + kmsKeyArn: + description: The ARN of the Key Management Service (KMS) customer + managed key that Lambda uses to encrypt your function's filter + criteria. + type: string + kmsKeyArnRef: + description: Reference to a Key in kms to populate kmsKeyArn. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + kmsKeyArnSelector: + description: Selector for a Key in kms to populate kmsKeyArn. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object maximumBatchingWindowInSeconds: description: The maximum amount of time to gather records before invoking the function, in seconds (between 0 and 300). Records @@ -1016,6 +1174,11 @@ spec: x-kubernetes-list-type: set id: type: string + kmsKeyArn: + description: The ARN of the Key Management Service (KMS) customer + managed key that Lambda uses to encrypt your function's filter + criteria. + type: string lastModified: description: The date this resource was last modified. type: string diff --git a/package/crds/opensearch.aws.upbound.io_domains.yaml b/package/crds/opensearch.aws.upbound.io_domains.yaml index 21fab9b121..60958d3730 100644 --- a/package/crds/opensearch.aws.upbound.io_domains.yaml +++ b/package/crds/opensearch.aws.upbound.io_domains.yaml @@ -1922,6 +1922,10 @@ spec: description: Domain-specific endpoint for Dashboard without https scheme. type: string + dashboardEndpointV2: + description: V2 domain endpoint for Dashboard that works with + both IPv4 and IPv6 addresses, without https scheme. + type: string domainEndpointOptions: description: Configuration block for domain endpoint HTTP(S) related options. Detailed below. @@ -1948,6 +1952,9 @@ spec: type: string type: object type: array + domainEndpointV2HostedZoneId: + description: Dual stack hosted zone ID for the domain. + type: string domainId: description: Unique identifier for the domain. type: string @@ -2006,6 +2013,11 @@ spec: description: Domain-specific endpoint used to submit index, search, and data upload requests. type: string + endpointV2: + description: V2 domain endpoint that works with both IPv4 and + IPv6 addresses, used to submit index, search, and data upload + requests. + type: string engineVersion: description: while Elasticsearch has elasticsearch_version type: string diff --git a/package/crds/rds.aws.upbound.io_clusters.yaml b/package/crds/rds.aws.upbound.io_clusters.yaml index b3c91cc33c..bf18b77aaf 100644 --- a/package/crds/rds.aws.upbound.io_clusters.yaml +++ b/package/crds/rds.aws.upbound.io_clusters.yaml @@ -431,6 +431,14 @@ spec: DB cluster. Valid Values: aurora-mysql, aurora-postgresql, mysql, postgres. (Note that mysql and postgres are Multi-AZ RDS clusters).' type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + is valid for cluster types Aurora DB clusters and Multi-AZ DB + clusters. Valid values are open-source-rds-extended-support, + open-source-rds-extended-support-disabled. Default value is + open-source-rds-extended-support. [Using Amazon RDS Extended + Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineMode: description: 'Database engine mode. Valid values: global (only valid for Aurora MySQL 1.21 and earlier), parallelquery, provisioned, @@ -660,6 +668,23 @@ spec: description: 'Network type of the cluster. Valid values: IPV4, DUAL.' type: string + performanceInsightsEnabled: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Enables + Performance Insights for the RDS Cluster + type: boolean + performanceInsightsKmsKeyId: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Specifies + the KMS Key ID to encrypt Performance Insights data. If not + specified, the default RDS KMS key will be used (aws/rds). + type: string + performanceInsightsRetentionPeriod: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Specifies + the amount of time to retain performance insights data for. + Defaults to 7 days if Performance Insights are enabled. Valid + values are 7, month * 31 (where month is a number of months + from 1-23), and 731. See here for more information on retention + periods. + type: number port: description: Port on which the DB accepts connections type: number @@ -774,6 +799,12 @@ spec: type: string type: object type: object + sourceClusterResourceId: + description: Cluster resource ID of the source database + cluster from which to restore. To be used for restoring + a deleted cluster in the same account which still has + a retained automatic backup available. + type: string useLatestRestorableTime: description: Set to true to restore the database cluster to the latest restorable backup time. Defaults to false. @@ -906,6 +937,12 @@ spec: 256. Valid Aurora PostgreSQL capacity values are (2, 4, 8, 16, 32, 64, 192, and 384). Defaults to 1. type: number + secondsBeforeTimeout: + description: Amount of time, in seconds, that Aurora Serverless + v1 tries to find a scaling point to perform seamless scaling + before enforcing the timeout action. Valid values are + 60 through 600. Defaults to 300. + type: number secondsUntilAutoPause: description: Time, in seconds, before an Aurora DB cluster in serverless mode is paused. Valid values are 300 through @@ -1438,6 +1475,14 @@ spec: DB cluster. Valid Values: aurora-mysql, aurora-postgresql, mysql, postgres. (Note that mysql and postgres are Multi-AZ RDS clusters).' type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + is valid for cluster types Aurora DB clusters and Multi-AZ DB + clusters. Valid values are open-source-rds-extended-support, + open-source-rds-extended-support-disabled. Default value is + open-source-rds-extended-support. [Using Amazon RDS Extended + Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineMode: description: 'Database engine mode. Valid values: global (only valid for Aurora MySQL 1.21 and earlier), parallelquery, provisioned, @@ -1667,6 +1712,23 @@ spec: description: 'Network type of the cluster. Valid values: IPV4, DUAL.' type: string + performanceInsightsEnabled: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Enables + Performance Insights for the RDS Cluster + type: boolean + performanceInsightsKmsKeyId: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Specifies + the KMS Key ID to encrypt Performance Insights data. If not + specified, the default RDS KMS key will be used (aws/rds). + type: string + performanceInsightsRetentionPeriod: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Specifies + the amount of time to retain performance insights data for. + Defaults to 7 days if Performance Insights are enabled. Valid + values are 7, month * 31 (where month is a number of months + from 1-23), and 731. See here for more information on retention + periods. + type: number port: description: Port on which the DB accepts connections type: number @@ -1777,6 +1839,12 @@ spec: type: string type: object type: object + sourceClusterResourceId: + description: Cluster resource ID of the source database + cluster from which to restore. To be used for restoring + a deleted cluster in the same account which still has + a retained automatic backup available. + type: string useLatestRestorableTime: description: Set to true to restore the database cluster to the latest restorable backup time. Defaults to false. @@ -1909,6 +1977,12 @@ spec: 256. Valid Aurora PostgreSQL capacity values are (2, 4, 8, 16, 32, 64, 192, and 384). Defaults to 1. type: number + secondsBeforeTimeout: + description: Amount of time, in seconds, that Aurora Serverless + v1 tries to find a scaling point to perform seamless scaling + before enforcing the timeout action. Valid values are + 60 through 600. Defaults to 300. + type: number secondsUntilAutoPause: description: Time, in seconds, before an Aurora DB cluster in serverless mode is paused. Valid values are 300 through @@ -2389,6 +2463,14 @@ spec: DB cluster. Valid Values: aurora-mysql, aurora-postgresql, mysql, postgres. (Note that mysql and postgres are Multi-AZ RDS clusters).' type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + is valid for cluster types Aurora DB clusters and Multi-AZ DB + clusters. Valid values are open-source-rds-extended-support, + open-source-rds-extended-support-disabled. Default value is + open-source-rds-extended-support. [Using Amazon RDS Extended + Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineMode: description: 'Database engine mode. Valid values: global (only valid for Aurora MySQL 1.21 and earlier), parallelquery, provisioned, @@ -2486,6 +2568,23 @@ spec: description: 'Network type of the cluster. Valid values: IPV4, DUAL.' type: string + performanceInsightsEnabled: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Enables + Performance Insights for the RDS Cluster + type: boolean + performanceInsightsKmsKeyId: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Specifies + the KMS Key ID to encrypt Performance Insights data. If not + specified, the default RDS KMS key will be used (aws/rds). + type: string + performanceInsightsRetentionPeriod: + description: Valid only for Non-Aurora Multi-AZ DB Clusters. Specifies + the amount of time to retain performance insights data for. + Defaults to 7 days if Performance Insights are enabled. Valid + values are 7, month * 31 (where month is a number of months + from 1-23), and 731. See here for more information on retention + periods. + type: number port: description: Port on which the DB accepts connections type: number @@ -2527,6 +2626,12 @@ spec: which to restore. When restoring from a cluster in another AWS account, the identifier is the ARN of that cluster. type: string + sourceClusterResourceId: + description: Cluster resource ID of the source database + cluster from which to restore. To be used for restoring + a deleted cluster in the same account which still has + a retained automatic backup available. + type: string useLatestRestorableTime: description: Set to true to restore the database cluster to the latest restorable backup time. Defaults to false. @@ -2585,6 +2690,12 @@ spec: 256. Valid Aurora PostgreSQL capacity values are (2, 4, 8, 16, 32, 64, 192, and 384). Defaults to 1. type: number + secondsBeforeTimeout: + description: Amount of time, in seconds, that Aurora Serverless + v1 tries to find a scaling point to perform seamless scaling + before enforcing the timeout action. Valid values are + 60 through 600. Defaults to 300. + type: number secondsUntilAutoPause: description: Time, in seconds, before an Aurora DB cluster in serverless mode is paused. Valid values are 300 through diff --git a/package/crds/rds.aws.upbound.io_instances.yaml b/package/crds/rds.aws.upbound.io_instances.yaml index 316c9dc836..a80534c890 100644 --- a/package/crds/rds.aws.upbound.io_instances.yaml +++ b/package/crds/rds.aws.upbound.io_instances.yaml @@ -117,6 +117,11 @@ spec: uses low-downtime updates, or will use RDS Blue/Green deployments. type: number + backupTarget: + description: Specifies where automated backups and manual snapshots + are stored. Possible values are region (default) and outposts. + See Working with Amazon RDS on AWS Outposts for more information. + type: string backupWindow: description: |- The daily time range (in UTC) during which automated backups are created if they are enabled. @@ -251,6 +256,11 @@ spec: type: string type: object type: object + dedicatedLogVolume: + description: Use a dedicated log volume (DLV) for the DB instance. + Requires Provisioned IOPS. See the AWS documentation for more + details. + type: boolean deleteAutomatedBackups: description: Specifies whether to remove automated backups immediately after the DB instance is deleted. Default is true. @@ -264,10 +274,34 @@ spec: description: The ID of the Directory Service Active Directory domain to create the instance in. type: string + domainAuthSecretArn: + description: The ARN for the Secrets Manager secret with the self + managed Active Directory credentials for the user joining the + domain. Conflicts with domain and domain_iam_role_name. + type: string + domainDnsIps: + description: The IPv4 DNS IP addresses of your primary and secondary + self managed Active Directory domain controllers. Two IP addresses + must be provided. If there isn't a secondary domain controller, + use the IP address of the primary domain controller for both + entries in the list. Conflicts with domain and domain_iam_role_name. + items: + type: string + type: array + x-kubernetes-list-type: set + domainFqdn: + description: The fully qualified domain name (FQDN) of the self + managed Active Directory domain. Conflicts with domain and domain_iam_role_name. + type: string domainIamRoleName: description: The name of the IAM role to be used when making API calls to the Directory Service. type: string + domainOu: + description: The self managed Active Directory organizational + unit for your DB instance to join. Conflicts with domain and + domain_iam_role_name. + type: string enabledCloudwatchLogsExports: description: 'Set of log types to enable for exporting to CloudWatch logs. If omitted, no logs will be exported. Valid values (depending @@ -287,6 +321,13 @@ spec: Aurora MySQL engines see Comparison in the [Amazon RDS Release Notes](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/Welcome.html). type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + applies only to RDS for MySQL and RDS for PostgreSQL. Valid + values are open-source-rds-extended-support, open-source-rds-extended-support-disabled. + Default value is open-source-rds-extended-support. [Using Amazon + RDS Extended Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineVersion: description: The engine version to use. If `autoMinorVersionUpgrade` is enabled, you can provide a prefix of the version such as @@ -309,6 +350,13 @@ spec: Specifies whether mappings of AWS Identity and Access Management (IAM) accounts to database accounts is enabled. type: boolean + identifier: + description: Required if restore_to_point_in_time is specified. + type: string + identifierPrefix: + description: Creates a unique identifier beginning with the specified + prefix. Conflicts with identifier. + type: string instanceClass: description: The instance type of the RDS instance. type: string @@ -619,6 +667,82 @@ spec: Name of the DB parameter group to associate. type: string + parameterGroupNameRef: + description: Reference to a ParameterGroup in rds to populate + parameterGroupName. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + parameterGroupNameSelector: + description: Selector for a ParameterGroup in rds to populate + parameterGroupName. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object passwordSecretRef: description: |- Password for the master DB user. Note that this may show up in @@ -873,6 +997,10 @@ spec: Guide for more information. type: string + upgradeStorageConfig: + description: Whether to upgrade the storage file system configuration + on the read replica. Can only be set with replicate_source_db. + type: boolean username: description: Username for the master DB user. Cannot be specified for a replica. @@ -1017,6 +1145,11 @@ spec: uses low-downtime updates, or will use RDS Blue/Green deployments. type: number + backupTarget: + description: Specifies where automated backups and manual snapshots + are stored. Possible values are region (default) and outposts. + See Working with Amazon RDS on AWS Outposts for more information. + type: string backupWindow: description: |- The daily time range (in UTC) during which automated backups are created if they are enabled. @@ -1151,6 +1284,11 @@ spec: type: string type: object type: object + dedicatedLogVolume: + description: Use a dedicated log volume (DLV) for the DB instance. + Requires Provisioned IOPS. See the AWS documentation for more + details. + type: boolean deleteAutomatedBackups: description: Specifies whether to remove automated backups immediately after the DB instance is deleted. Default is true. @@ -1164,10 +1302,34 @@ spec: description: The ID of the Directory Service Active Directory domain to create the instance in. type: string + domainAuthSecretArn: + description: The ARN for the Secrets Manager secret with the self + managed Active Directory credentials for the user joining the + domain. Conflicts with domain and domain_iam_role_name. + type: string + domainDnsIps: + description: The IPv4 DNS IP addresses of your primary and secondary + self managed Active Directory domain controllers. Two IP addresses + must be provided. If there isn't a secondary domain controller, + use the IP address of the primary domain controller for both + entries in the list. Conflicts with domain and domain_iam_role_name. + items: + type: string + type: array + x-kubernetes-list-type: set + domainFqdn: + description: The fully qualified domain name (FQDN) of the self + managed Active Directory domain. Conflicts with domain and domain_iam_role_name. + type: string domainIamRoleName: description: The name of the IAM role to be used when making API calls to the Directory Service. type: string + domainOu: + description: The self managed Active Directory organizational + unit for your DB instance to join. Conflicts with domain and + domain_iam_role_name. + type: string enabledCloudwatchLogsExports: description: 'Set of log types to enable for exporting to CloudWatch logs. If omitted, no logs will be exported. Valid values (depending @@ -1187,6 +1349,13 @@ spec: Aurora MySQL engines see Comparison in the [Amazon RDS Release Notes](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/Welcome.html). type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + applies only to RDS for MySQL and RDS for PostgreSQL. Valid + values are open-source-rds-extended-support, open-source-rds-extended-support-disabled. + Default value is open-source-rds-extended-support. [Using Amazon + RDS Extended Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineVersion: description: The engine version to use. If `autoMinorVersionUpgrade` is enabled, you can provide a prefix of the version such as @@ -1209,6 +1378,13 @@ spec: Specifies whether mappings of AWS Identity and Access Management (IAM) accounts to database accounts is enabled. type: boolean + identifier: + description: Required if restore_to_point_in_time is specified. + type: string + identifierPrefix: + description: Creates a unique identifier beginning with the specified + prefix. Conflicts with identifier. + type: string instanceClass: description: The instance type of the RDS instance. type: string @@ -1519,6 +1695,102 @@ spec: Name of the DB parameter group to associate. type: string + parameterGroupNameRef: + description: Reference to a ParameterGroup in rds to populate + parameterGroupName. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + parameterGroupNameSelector: + description: Selector for a ParameterGroup in rds to populate + parameterGroupName. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching labels + is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object + passwordSecretRef: + description: |- + Password for the master DB user. Note that this may show up in + logs, and it will be stored in the state file. Cannot be set if manage_master_user_password is set to true. + Password for the master DB user. If you set autoGeneratePassword to true, the Secret referenced here will be created or updated with generated password if it does not already contain one. + properties: + key: + description: The key to select. + type: string + name: + description: Name of the secret. + type: string + namespace: + description: Namespace of the secret. + type: string + required: + - key + - name + - namespace + type: object performanceInsightsEnabled: description: Specifies whether Performance Insights are enabled. Defaults to false. @@ -1749,6 +2021,10 @@ spec: Guide for more information. type: string + upgradeStorageConfig: + description: Whether to upgrade the storage file system configuration + on the read replica. Can only be set with replicate_source_db. + type: boolean username: description: Username for the master DB user. Cannot be specified for a replica. @@ -2062,6 +2338,11 @@ spec: uses low-downtime updates, or will use RDS Blue/Green deployments. type: number + backupTarget: + description: Specifies where automated backups and manual snapshots + are stored. Possible values are region (default) and outposts. + See Working with Amazon RDS on AWS Outposts for more information. + type: string backupWindow: description: |- The daily time range (in UTC) during which automated backups are created if they are enabled. @@ -2122,6 +2403,11 @@ spec: action CreateDBInstanceReadReplica for additional read replica contraints. type: string + dedicatedLogVolume: + description: Use a dedicated log volume (DLV) for the DB instance. + Requires Provisioned IOPS. See the AWS documentation for more + details. + type: boolean deleteAutomatedBackups: description: Specifies whether to remove automated backups immediately after the DB instance is deleted. Default is true. @@ -2135,10 +2421,34 @@ spec: description: The ID of the Directory Service Active Directory domain to create the instance in. type: string + domainAuthSecretArn: + description: The ARN for the Secrets Manager secret with the self + managed Active Directory credentials for the user joining the + domain. Conflicts with domain and domain_iam_role_name. + type: string + domainDnsIps: + description: The IPv4 DNS IP addresses of your primary and secondary + self managed Active Directory domain controllers. Two IP addresses + must be provided. If there isn't a secondary domain controller, + use the IP address of the primary domain controller for both + entries in the list. Conflicts with domain and domain_iam_role_name. + items: + type: string + type: array + x-kubernetes-list-type: set + domainFqdn: + description: The fully qualified domain name (FQDN) of the self + managed Active Directory domain. Conflicts with domain and domain_iam_role_name. + type: string domainIamRoleName: description: The name of the IAM role to be used when making API calls to the Directory Service. type: string + domainOu: + description: The self managed Active Directory organizational + unit for your DB instance to join. Conflicts with domain and + domain_iam_role_name. + type: string enabledCloudwatchLogsExports: description: 'Set of log types to enable for exporting to CloudWatch logs. If omitted, no logs will be exported. Valid values (depending @@ -2161,6 +2471,13 @@ spec: Aurora MySQL engines see Comparison in the [Amazon RDS Release Notes](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/Welcome.html). type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + applies only to RDS for MySQL and RDS for PostgreSQL. Valid + values are open-source-rds-extended-support, open-source-rds-extended-support-disabled. + Default value is open-source-rds-extended-support. [Using Amazon + RDS Extended Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineVersion: description: The engine version to use. If `autoMinorVersionUpgrade` is enabled, you can provide a prefix of the version such as @@ -2194,6 +2511,13 @@ spec: id: description: The RDS instance ID. type: string + identifier: + description: Required if restore_to_point_in_time is specified. + type: string + identifierPrefix: + description: Creates a unique identifier beginning with the specified + prefix. Conflicts with identifier. + type: string instanceClass: description: The instance type of the RDS instance. type: string @@ -2498,6 +2822,10 @@ spec: Guide for more information. type: string + upgradeStorageConfig: + description: Whether to upgrade the storage file system configuration + on the read replica. Can only be set with replicate_source_db. + type: boolean username: description: Username for the master DB user. Cannot be specified for a replica. @@ -2877,6 +3205,13 @@ spec: Aurora MySQL engines see Comparison in the [Amazon RDS Release Notes](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/Welcome.html). type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + applies only to RDS for MySQL and RDS for PostgreSQL. Valid + values are open-source-rds-extended-support, open-source-rds-extended-support-disabled. + Default value is open-source-rds-extended-support. [Using Amazon + RDS Extended Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineVersion: description: The engine version to use. If `autoMinorVersionUpgrade` is enabled, you can provide a prefix of the version such as @@ -3528,6 +3863,10 @@ spec: Guide for more information. type: string + upgradeStorageConfig: + description: Whether to upgrade the storage file system configuration + on the read replica. Can only be set with replicate_source_db. + type: boolean username: description: Username for the master DB user. Cannot be specified for a replica. @@ -3878,6 +4217,13 @@ spec: Aurora MySQL engines see Comparison in the [Amazon RDS Release Notes](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/Welcome.html). type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + applies only to RDS for MySQL and RDS for PostgreSQL. Valid + values are open-source-rds-extended-support, open-source-rds-extended-support-disabled. + Default value is open-source-rds-extended-support. [Using Amazon + RDS Extended Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineVersion: description: The engine version to use. If `autoMinorVersionUpgrade` is enabled, you can provide a prefix of the version such as @@ -4525,6 +4871,10 @@ spec: Guide for more information. type: string + upgradeStorageConfig: + description: Whether to upgrade the storage file system configuration + on the read replica. Can only be set with replicate_source_db. + type: boolean username: description: Username for the master DB user. Cannot be specified for a replica. @@ -4973,6 +5323,13 @@ spec: Aurora MySQL engines see Comparison in the [Amazon RDS Release Notes](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/Welcome.html). type: string + engineLifecycleSupport: + description: 'The life cycle type for this DB instance. This setting + applies only to RDS for MySQL and RDS for PostgreSQL. Valid + values are open-source-rds-extended-support, open-source-rds-extended-support-disabled. + Default value is open-source-rds-extended-support. [Using Amazon + RDS Extended Support]: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/extended-support.html' + type: string engineVersion: description: The engine version to use. If `autoMinorVersionUpgrade` is enabled, you can provide a prefix of the version such as @@ -5299,6 +5656,10 @@ spec: Guide for more information. type: string + upgradeStorageConfig: + description: Whether to upgrade the storage file system configuration + on the read replica. Can only be set with replicate_source_db. + type: boolean username: description: Username for the master DB user. Cannot be specified for a replica. diff --git a/package/crds/sagemaker.aws.upbound.io_domains.yaml b/package/crds/sagemaker.aws.upbound.io_domains.yaml index 3669567131..bafa9dc3c9 100644 --- a/package/crds/sagemaker.aws.upbound.io_domains.yaml +++ b/package/crds/sagemaker.aws.upbound.io_domains.yaml @@ -92,9 +92,124 @@ spec: Block below. items: properties: + customFileSystemConfig: + description: The settings for assigning a custom file system + to a user profile. Permitted users can access this file + system in Amazon SageMaker Studio. See custom_file_system_config + Block below. + items: + properties: + efsFileSystemConfig: + description: The default EBS storage settings for + a private space. See efs_file_system_config Block + below. + items: + properties: + fileSystemId: + description: The ID of your Amazon EFS file + system. + type: string + fileSystemPath: + description: The path to the file system directory + that is accessible in Amazon SageMaker Studio. + Permitted users can access only this directory + and below. + type: string + type: object + type: array + type: object + type: array + customPosixUserConfig: + description: Details about the POSIX identity that is used + for file system operations. See custom_posix_user_config + Block below. + items: + properties: + gid: + description: The POSIX group ID. + type: number + uid: + description: The POSIX user ID. + type: number + type: object + type: array executionRole: description: The execution role for the space. type: string + jupyterLabAppSettings: + description: The settings for the JupyterLab application. + See jupyter_lab_app_settings Block below. + items: + properties: + codeRepository: + description: A list of Git repositories that SageMaker + automatically displays to users for cloning in the + JupyterServer application. see code_repository Block + below. + items: + properties: + repositoryUrl: + description: The URL of the Git repository. + type: string + type: object + type: array + customImage: + description: A list of custom SageMaker images that + are configured to run as a KernelGateway app. see + custom_image Block below. + items: + properties: + appImageConfigName: + description: The name of the App Image Config. + type: string + imageName: + description: The name of the Custom Image. + type: string + imageVersionNumber: + description: The version number of the Custom + Image. + type: number + type: object + type: array + defaultResourceSpec: + description: The default instance type and the Amazon + Resource Name (ARN) of the SageMaker image created + on the instance. see default_resource_spec Block + below. + items: + properties: + instanceType: + description: The instance type that the image + version runs on.. For valid values see SageMaker + Instance Types. + type: string + lifecycleConfigArn: + description: The Amazon Resource Name (ARN) + of the Lifecycle Configuration attached to + the Resource. + type: string + sagemakerImageArn: + description: The ARN of the SageMaker image + that the image version belongs to. + type: string + sagemakerImageVersionAlias: + description: The SageMaker Image Version Alias. + type: string + sagemakerImageVersionArn: + description: The ARN of the image version created + on the instance. + type: string + type: object + type: array + lifecycleConfigArns: + description: The Amazon Resource Name (ARN) of the + Lifecycle Configurations. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array jupyterServerAppSettings: description: The Jupyter server's app settings. See jupyter_server_app_settings Block below. @@ -220,6 +335,29 @@ spec: type: string type: array x-kubernetes-list-type: set + spaceStorageSettings: + description: The storage settings for a private space. See + space_storage_settings Block below. + items: + properties: + defaultEbsStorageSettings: + description: The default EBS storage settings for + a private space. See default_ebs_storage_settings + Block below. + items: + properties: + defaultEbsVolumeSizeInGb: + description: The default size of the EBS storage + volume for a private space. + type: number + maximumEbsVolumeSizeInGb: + description: The maximum size of the EBS storage + volume for a private space. + type: number + type: object + type: array + type: object + type: array type: object type: array defaultUserSettings: @@ -1028,6 +1166,28 @@ spec: even if that is the default experience for the domain. Valid values are ENABLED and DISABLED. type: string + studioWebPortalSettings: + description: The Studio Web Portal settings. See studio_web_portal_settings + Block below. + items: + properties: + hiddenAppTypes: + description: The Applications supported in Studio + that are hidden from the Studio left navigation + pane. + items: + type: string + type: array + x-kubernetes-list-type: set + hiddenMlTools: + description: The machine learning tools that are hidden + from the Studio left navigation pane. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array tensorBoardAppSettings: description: The TensorBoard app settings. See tensor_board_app_settings Block below. @@ -1074,6 +1234,26 @@ spec: description: The domain settings. See domain_settings Block below. items: properties: + dockerSettings: + description: A collection of settings that configure the + domain’s Docker interaction. see docker_settings Block + below. + items: + properties: + enableDockerAccess: + description: Indicates whether the domain can access + Docker. Valid values are ENABLED and DISABLED. + type: string + vpcOnlyTrustedAccounts: + description: The list of Amazon Web Services accounts + that are trusted when the domain is created in VPC-only + mode. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array executionRoleIdentityConfig: description: The configuration for attaching a SageMaker user profile name to the execution role as a sts:SourceIdentity @@ -1438,9 +1618,124 @@ spec: Block below. items: properties: + customFileSystemConfig: + description: The settings for assigning a custom file system + to a user profile. Permitted users can access this file + system in Amazon SageMaker Studio. See custom_file_system_config + Block below. + items: + properties: + efsFileSystemConfig: + description: The default EBS storage settings for + a private space. See efs_file_system_config Block + below. + items: + properties: + fileSystemId: + description: The ID of your Amazon EFS file + system. + type: string + fileSystemPath: + description: The path to the file system directory + that is accessible in Amazon SageMaker Studio. + Permitted users can access only this directory + and below. + type: string + type: object + type: array + type: object + type: array + customPosixUserConfig: + description: Details about the POSIX identity that is used + for file system operations. See custom_posix_user_config + Block below. + items: + properties: + gid: + description: The POSIX group ID. + type: number + uid: + description: The POSIX user ID. + type: number + type: object + type: array executionRole: description: The execution role for the space. type: string + jupyterLabAppSettings: + description: The settings for the JupyterLab application. + See jupyter_lab_app_settings Block below. + items: + properties: + codeRepository: + description: A list of Git repositories that SageMaker + automatically displays to users for cloning in the + JupyterServer application. see code_repository Block + below. + items: + properties: + repositoryUrl: + description: The URL of the Git repository. + type: string + type: object + type: array + customImage: + description: A list of custom SageMaker images that + are configured to run as a KernelGateway app. see + custom_image Block below. + items: + properties: + appImageConfigName: + description: The name of the App Image Config. + type: string + imageName: + description: The name of the Custom Image. + type: string + imageVersionNumber: + description: The version number of the Custom + Image. + type: number + type: object + type: array + defaultResourceSpec: + description: The default instance type and the Amazon + Resource Name (ARN) of the SageMaker image created + on the instance. see default_resource_spec Block + below. + items: + properties: + instanceType: + description: The instance type that the image + version runs on.. For valid values see SageMaker + Instance Types. + type: string + lifecycleConfigArn: + description: The Amazon Resource Name (ARN) + of the Lifecycle Configuration attached to + the Resource. + type: string + sagemakerImageArn: + description: The ARN of the SageMaker image + that the image version belongs to. + type: string + sagemakerImageVersionAlias: + description: The SageMaker Image Version Alias. + type: string + sagemakerImageVersionArn: + description: The ARN of the image version created + on the instance. + type: string + type: object + type: array + lifecycleConfigArns: + description: The Amazon Resource Name (ARN) of the + Lifecycle Configurations. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array jupyterServerAppSettings: description: The Jupyter server's app settings. See jupyter_server_app_settings Block below. @@ -1566,6 +1861,29 @@ spec: type: string type: array x-kubernetes-list-type: set + spaceStorageSettings: + description: The storage settings for a private space. See + space_storage_settings Block below. + items: + properties: + defaultEbsStorageSettings: + description: The default EBS storage settings for + a private space. See default_ebs_storage_settings + Block below. + items: + properties: + defaultEbsVolumeSizeInGb: + description: The default size of the EBS storage + volume for a private space. + type: number + maximumEbsVolumeSizeInGb: + description: The maximum size of the EBS storage + volume for a private space. + type: number + type: object + type: array + type: object + type: array type: object type: array defaultUserSettings: @@ -2374,6 +2692,28 @@ spec: even if that is the default experience for the domain. Valid values are ENABLED and DISABLED. type: string + studioWebPortalSettings: + description: The Studio Web Portal settings. See studio_web_portal_settings + Block below. + items: + properties: + hiddenAppTypes: + description: The Applications supported in Studio + that are hidden from the Studio left navigation + pane. + items: + type: string + type: array + x-kubernetes-list-type: set + hiddenMlTools: + description: The machine learning tools that are hidden + from the Studio left navigation pane. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array tensorBoardAppSettings: description: The TensorBoard app settings. See tensor_board_app_settings Block below. @@ -2420,6 +2760,26 @@ spec: description: The domain settings. See domain_settings Block below. items: properties: + dockerSettings: + description: A collection of settings that configure the + domain’s Docker interaction. see docker_settings Block + below. + items: + properties: + enableDockerAccess: + description: Indicates whether the domain can access + Docker. Valid values are ENABLED and DISABLED. + type: string + vpcOnlyTrustedAccounts: + description: The list of Amazon Web Services accounts + that are trusted when the domain is created in VPC-only + mode. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array executionRoleIdentityConfig: description: The configuration for attaching a SageMaker user profile name to the execution role as a sts:SourceIdentity @@ -2954,9 +3314,124 @@ spec: Block below. items: properties: + customFileSystemConfig: + description: The settings for assigning a custom file system + to a user profile. Permitted users can access this file + system in Amazon SageMaker Studio. See custom_file_system_config + Block below. + items: + properties: + efsFileSystemConfig: + description: The default EBS storage settings for + a private space. See efs_file_system_config Block + below. + items: + properties: + fileSystemId: + description: The ID of your Amazon EFS file + system. + type: string + fileSystemPath: + description: The path to the file system directory + that is accessible in Amazon SageMaker Studio. + Permitted users can access only this directory + and below. + type: string + type: object + type: array + type: object + type: array + customPosixUserConfig: + description: Details about the POSIX identity that is used + for file system operations. See custom_posix_user_config + Block below. + items: + properties: + gid: + description: The POSIX group ID. + type: number + uid: + description: The POSIX user ID. + type: number + type: object + type: array executionRole: description: The execution role for the space. type: string + jupyterLabAppSettings: + description: The settings for the JupyterLab application. + See jupyter_lab_app_settings Block below. + items: + properties: + codeRepository: + description: A list of Git repositories that SageMaker + automatically displays to users for cloning in the + JupyterServer application. see code_repository Block + below. + items: + properties: + repositoryUrl: + description: The URL of the Git repository. + type: string + type: object + type: array + customImage: + description: A list of custom SageMaker images that + are configured to run as a KernelGateway app. see + custom_image Block below. + items: + properties: + appImageConfigName: + description: The name of the App Image Config. + type: string + imageName: + description: The name of the Custom Image. + type: string + imageVersionNumber: + description: The version number of the Custom + Image. + type: number + type: object + type: array + defaultResourceSpec: + description: The default instance type and the Amazon + Resource Name (ARN) of the SageMaker image created + on the instance. see default_resource_spec Block + below. + items: + properties: + instanceType: + description: The instance type that the image + version runs on.. For valid values see SageMaker + Instance Types. + type: string + lifecycleConfigArn: + description: The Amazon Resource Name (ARN) + of the Lifecycle Configuration attached to + the Resource. + type: string + sagemakerImageArn: + description: The ARN of the SageMaker image + that the image version belongs to. + type: string + sagemakerImageVersionAlias: + description: The SageMaker Image Version Alias. + type: string + sagemakerImageVersionArn: + description: The ARN of the image version created + on the instance. + type: string + type: object + type: array + lifecycleConfigArns: + description: The Amazon Resource Name (ARN) of the + Lifecycle Configurations. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array jupyterServerAppSettings: description: The Jupyter server's app settings. See jupyter_server_app_settings Block below. @@ -3082,6 +3557,29 @@ spec: type: string type: array x-kubernetes-list-type: set + spaceStorageSettings: + description: The storage settings for a private space. See + space_storage_settings Block below. + items: + properties: + defaultEbsStorageSettings: + description: The default EBS storage settings for + a private space. See default_ebs_storage_settings + Block below. + items: + properties: + defaultEbsVolumeSizeInGb: + description: The default size of the EBS storage + volume for a private space. + type: number + maximumEbsVolumeSizeInGb: + description: The maximum size of the EBS storage + volume for a private space. + type: number + type: object + type: array + type: object + type: array type: object type: array defaultUserSettings: @@ -3664,6 +4162,28 @@ spec: even if that is the default experience for the domain. Valid values are ENABLED and DISABLED. type: string + studioWebPortalSettings: + description: The Studio Web Portal settings. See studio_web_portal_settings + Block below. + items: + properties: + hiddenAppTypes: + description: The Applications supported in Studio + that are hidden from the Studio left navigation + pane. + items: + type: string + type: array + x-kubernetes-list-type: set + hiddenMlTools: + description: The machine learning tools that are hidden + from the Studio left navigation pane. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array tensorBoardAppSettings: description: The TensorBoard app settings. See tensor_board_app_settings Block below. @@ -3710,6 +4230,26 @@ spec: description: The domain settings. See domain_settings Block below. items: properties: + dockerSettings: + description: A collection of settings that configure the + domain’s Docker interaction. see docker_settings Block + below. + items: + properties: + enableDockerAccess: + description: Indicates whether the domain can access + Docker. Valid values are ENABLED and DISABLED. + type: string + vpcOnlyTrustedAccounts: + description: The list of Amazon Web Services accounts + that are trusted when the domain is created in VPC-only + mode. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array executionRoleIdentityConfig: description: The configuration for attaching a SageMaker user profile name to the execution role as a sts:SourceIdentity diff --git a/package/crds/sagemaker.aws.upbound.io_endpointconfigurations.yaml b/package/crds/sagemaker.aws.upbound.io_endpointconfigurations.yaml index 8fe1b864f7..c43b6606aa 100644 --- a/package/crds/sagemaker.aws.upbound.io_endpointconfigurations.yaml +++ b/package/crds/sagemaker.aws.upbound.io_endpointconfigurations.yaml @@ -335,6 +335,28 @@ spec: instanceType: description: The type of instance to start. type: string + managedInstanceScaling: + description: Settings that control the range in the number + of instances that the endpoint provisions as it scales + up or down to accommodate traffic. + items: + properties: + maxInstanceCount: + description: The maximum number of instances that + the endpoint can provision when it scales up to + accommodate an increase in traffic. + type: number + minInstanceCount: + description: The minimum number of instances that + the endpoint must retain when it scales down to + accommodate a decrease in traffic. + type: number + status: + description: Indicates whether managed instance scaling + is enabled. Valid values are ENABLED and DISABLED. + type: string + type: object + type: array modelDataDownloadTimeoutInSeconds: description: The timeout value, in seconds, to download and extract the model that you want to host from Amazon @@ -535,6 +557,28 @@ spec: instanceType: description: The type of instance to start. type: string + managedInstanceScaling: + description: Settings that control the range in the number + of instances that the endpoint provisions as it scales + up or down to accommodate traffic. + items: + properties: + maxInstanceCount: + description: The maximum number of instances that + the endpoint can provision when it scales up to + accommodate an increase in traffic. + type: number + minInstanceCount: + description: The minimum number of instances that + the endpoint must retain when it scales down to + accommodate a decrease in traffic. + type: number + status: + description: Indicates whether managed instance scaling + is enabled. Valid values are ENABLED and DISABLED. + type: string + type: object + type: array modelDataDownloadTimeoutInSeconds: description: The timeout value, in seconds, to download and extract the model that you want to host from Amazon @@ -876,6 +920,28 @@ spec: instanceType: description: The type of instance to start. type: string + managedInstanceScaling: + description: Settings that control the range in the number + of instances that the endpoint provisions as it scales + up or down to accommodate traffic. + items: + properties: + maxInstanceCount: + description: The maximum number of instances that + the endpoint can provision when it scales up to + accommodate an increase in traffic. + type: number + minInstanceCount: + description: The minimum number of instances that + the endpoint must retain when it scales down to + accommodate a decrease in traffic. + type: number + status: + description: Indicates whether managed instance scaling + is enabled. Valid values are ENABLED and DISABLED. + type: string + type: object + type: array modelDataDownloadTimeoutInSeconds: description: The timeout value, in seconds, to download and extract the model that you want to host from Amazon @@ -1072,6 +1138,28 @@ spec: instanceType: description: The type of instance to start. type: string + managedInstanceScaling: + description: Settings that control the range in the number + of instances that the endpoint provisions as it scales + up or down to accommodate traffic. + items: + properties: + maxInstanceCount: + description: The maximum number of instances that + the endpoint can provision when it scales up to + accommodate an increase in traffic. + type: number + minInstanceCount: + description: The minimum number of instances that + the endpoint must retain when it scales down to + accommodate a decrease in traffic. + type: number + status: + description: Indicates whether managed instance scaling + is enabled. Valid values are ENABLED and DISABLED. + type: string + type: object + type: array modelDataDownloadTimeoutInSeconds: description: The timeout value, in seconds, to download and extract the model that you want to host from Amazon @@ -1508,6 +1596,28 @@ spec: instanceType: description: The type of instance to start. type: string + managedInstanceScaling: + description: Settings that control the range in the number + of instances that the endpoint provisions as it scales + up or down to accommodate traffic. + items: + properties: + maxInstanceCount: + description: The maximum number of instances that + the endpoint can provision when it scales up to + accommodate an increase in traffic. + type: number + minInstanceCount: + description: The minimum number of instances that + the endpoint must retain when it scales down to + accommodate a decrease in traffic. + type: number + status: + description: Indicates whether managed instance scaling + is enabled. Valid values are ENABLED and DISABLED. + type: string + type: object + type: array modelDataDownloadTimeoutInSeconds: description: The timeout value, in seconds, to download and extract the model that you want to host from Amazon @@ -1628,6 +1738,28 @@ spec: instanceType: description: The type of instance to start. type: string + managedInstanceScaling: + description: Settings that control the range in the number + of instances that the endpoint provisions as it scales + up or down to accommodate traffic. + items: + properties: + maxInstanceCount: + description: The maximum number of instances that + the endpoint can provision when it scales up to + accommodate an increase in traffic. + type: number + minInstanceCount: + description: The minimum number of instances that + the endpoint must retain when it scales down to + accommodate a decrease in traffic. + type: number + status: + description: Indicates whether managed instance scaling + is enabled. Valid values are ENABLED and DISABLED. + type: string + type: object + type: array modelDataDownloadTimeoutInSeconds: description: The timeout value, in seconds, to download and extract the model that you want to host from Amazon diff --git a/package/crds/sagemaker.aws.upbound.io_models.yaml b/package/crds/sagemaker.aws.upbound.io_models.yaml index 1089de50f4..7e7a1f6b5a 100644 --- a/package/crds/sagemaker.aws.upbound.io_models.yaml +++ b/package/crds/sagemaker.aws.upbound.io_models.yaml @@ -131,6 +131,10 @@ spec: type: array type: object type: array + inferenceSpecificationName: + description: The inference specification name in the model + package version. + type: string mode: description: The container hosts value SingleModel/MultiModel. The default value is SingleModel. @@ -150,6 +154,28 @@ spec: description: 'How the model data is prepared. Allowed values are: None and Gzip.' type: string + modelAccessConfig: + description: Specifies the access configuration + file for the ML model. You can explicitly + accept the model end-user license agreement + (EULA) within the [model_access_config configuration + block]. see Model Access Config. + items: + properties: + acceptEula: + description: Specifies agreement to the + model end-user license agreement (EULA). + The AcceptEula value must be explicitly + defined as true in order to accept the + EULA that this model requires. You are + responsible for reviewing and complying + with any applicable license terms and + making sure they are acceptable for + your use case before downloading or + using a model. + type: boolean + type: object + type: array s3DataType: description: 'The type of model data to deploy. Allowed values are: S3Object and S3Prefix.' @@ -169,6 +195,25 @@ spec: description: The Amazon Resource Name (ARN) of the model package to use to create the model. type: string + multiModelConfig: + description: Specifies additional configuration for multi-model + endpoints. see Multi Model Config. + items: + properties: + modelCacheSetting: + description: 'Whether to cache models for a multi-model + endpoint. By default, multi-model endpoints cache + models so that a model does not have to be loaded + into memory each time it is invoked. Some use cases + do not benefit from model caching. For example, + if an endpoint hosts a large number of models that + are each invoked infrequently, the endpoint might + perform better if you disable model caching. To + disable model caching, set the value of this parameter + to Disabled. Allowed values are: Enabled and Disabled.' + type: string + type: object + type: array type: object type: array enableNetworkIsolation: @@ -323,6 +368,10 @@ spec: type: array type: object type: array + inferenceSpecificationName: + description: The inference specification name in the model + package version. + type: string mode: description: The container hosts value SingleModel/MultiModel. The default value is SingleModel. @@ -342,6 +391,28 @@ spec: description: 'How the model data is prepared. Allowed values are: None and Gzip.' type: string + modelAccessConfig: + description: Specifies the access configuration + file for the ML model. You can explicitly + accept the model end-user license agreement + (EULA) within the [model_access_config configuration + block]. see Model Access Config. + items: + properties: + acceptEula: + description: Specifies agreement to the + model end-user license agreement (EULA). + The AcceptEula value must be explicitly + defined as true in order to accept the + EULA that this model requires. You are + responsible for reviewing and complying + with any applicable license terms and + making sure they are acceptable for + your use case before downloading or + using a model. + type: boolean + type: object + type: array s3DataType: description: 'The type of model data to deploy. Allowed values are: S3Object and S3Prefix.' @@ -361,6 +432,25 @@ spec: description: The Amazon Resource Name (ARN) of the model package to use to create the model. type: string + multiModelConfig: + description: Specifies additional configuration for multi-model + endpoints. see Multi Model Config. + items: + properties: + modelCacheSetting: + description: 'Whether to cache models for a multi-model + endpoint. By default, multi-model endpoints cache + models so that a model does not have to be loaded + into memory each time it is invoked. Some use cases + do not benefit from model caching. For example, + if an endpoint hosts a large number of models that + are each invoked infrequently, the endpoint might + perform better if you disable model caching. To + disable model caching, set the value of this parameter + to Disabled. Allowed values are: Enabled and Disabled.' + type: string + type: object + type: array type: object type: array region: @@ -464,6 +554,10 @@ spec: type: array type: object type: array + inferenceSpecificationName: + description: The inference specification name in the model + package version. + type: string mode: description: The container hosts value SingleModel/MultiModel. The default value is SingleModel. @@ -483,6 +577,28 @@ spec: description: 'How the model data is prepared. Allowed values are: None and Gzip.' type: string + modelAccessConfig: + description: Specifies the access configuration + file for the ML model. You can explicitly + accept the model end-user license agreement + (EULA) within the [model_access_config configuration + block]. see Model Access Config. + items: + properties: + acceptEula: + description: Specifies agreement to the + model end-user license agreement (EULA). + The AcceptEula value must be explicitly + defined as true in order to accept the + EULA that this model requires. You are + responsible for reviewing and complying + with any applicable license terms and + making sure they are acceptable for + your use case before downloading or + using a model. + type: boolean + type: object + type: array s3DataType: description: 'The type of model data to deploy. Allowed values are: S3Object and S3Prefix.' @@ -502,6 +618,25 @@ spec: description: The Amazon Resource Name (ARN) of the model package to use to create the model. type: string + multiModelConfig: + description: Specifies additional configuration for multi-model + endpoints. see Multi Model Config. + items: + properties: + modelCacheSetting: + description: 'Whether to cache models for a multi-model + endpoint. By default, multi-model endpoints cache + models so that a model does not have to be loaded + into memory each time it is invoked. Some use cases + do not benefit from model caching. For example, + if an endpoint hosts a large number of models that + are each invoked infrequently, the endpoint might + perform better if you disable model caching. To + disable model caching, set the value of this parameter + to Disabled. Allowed values are: Enabled and Disabled.' + type: string + type: object + type: array type: object type: array enableNetworkIsolation: @@ -656,6 +791,10 @@ spec: type: array type: object type: array + inferenceSpecificationName: + description: The inference specification name in the model + package version. + type: string mode: description: The container hosts value SingleModel/MultiModel. The default value is SingleModel. @@ -675,6 +814,28 @@ spec: description: 'How the model data is prepared. Allowed values are: None and Gzip.' type: string + modelAccessConfig: + description: Specifies the access configuration + file for the ML model. You can explicitly + accept the model end-user license agreement + (EULA) within the [model_access_config configuration + block]. see Model Access Config. + items: + properties: + acceptEula: + description: Specifies agreement to the + model end-user license agreement (EULA). + The AcceptEula value must be explicitly + defined as true in order to accept the + EULA that this model requires. You are + responsible for reviewing and complying + with any applicable license terms and + making sure they are acceptable for + your use case before downloading or + using a model. + type: boolean + type: object + type: array s3DataType: description: 'The type of model data to deploy. Allowed values are: S3Object and S3Prefix.' @@ -694,6 +855,25 @@ spec: description: The Amazon Resource Name (ARN) of the model package to use to create the model. type: string + multiModelConfig: + description: Specifies additional configuration for multi-model + endpoints. see Multi Model Config. + items: + properties: + modelCacheSetting: + description: 'Whether to cache models for a multi-model + endpoint. By default, multi-model endpoints cache + models so that a model does not have to be loaded + into memory each time it is invoked. Some use cases + do not benefit from model caching. For example, + if an endpoint hosts a large number of models that + are each invoked infrequently, the endpoint might + perform better if you disable model caching. To + disable model caching, set the value of this parameter + to Disabled. Allowed values are: Enabled and Disabled.' + type: string + type: object + type: array type: object type: array tags: @@ -954,6 +1134,10 @@ spec: type: array type: object type: array + inferenceSpecificationName: + description: The inference specification name in the model + package version. + type: string mode: description: The container hosts value SingleModel/MultiModel. The default value is SingleModel. @@ -973,6 +1157,28 @@ spec: description: 'How the model data is prepared. Allowed values are: None and Gzip.' type: string + modelAccessConfig: + description: Specifies the access configuration + file for the ML model. You can explicitly + accept the model end-user license agreement + (EULA) within the [model_access_config configuration + block]. see Model Access Config. + items: + properties: + acceptEula: + description: Specifies agreement to the + model end-user license agreement (EULA). + The AcceptEula value must be explicitly + defined as true in order to accept the + EULA that this model requires. You are + responsible for reviewing and complying + with any applicable license terms and + making sure they are acceptable for + your use case before downloading or + using a model. + type: boolean + type: object + type: array s3DataType: description: 'The type of model data to deploy. Allowed values are: S3Object and S3Prefix.' @@ -992,6 +1198,25 @@ spec: description: The Amazon Resource Name (ARN) of the model package to use to create the model. type: string + multiModelConfig: + description: Specifies additional configuration for multi-model + endpoints. see Multi Model Config. + items: + properties: + modelCacheSetting: + description: 'Whether to cache models for a multi-model + endpoint. By default, multi-model endpoints cache + models so that a model does not have to be loaded + into memory each time it is invoked. Some use cases + do not benefit from model caching. For example, + if an endpoint hosts a large number of models that + are each invoked infrequently, the endpoint might + perform better if you disable model caching. To + disable model caching, set the value of this parameter + to Disabled. Allowed values are: Enabled and Disabled.' + type: string + type: object + type: array type: object type: array enableNetworkIsolation: @@ -1074,6 +1299,10 @@ spec: type: array type: object type: array + inferenceSpecificationName: + description: The inference specification name in the model + package version. + type: string mode: description: The container hosts value SingleModel/MultiModel. The default value is SingleModel. @@ -1093,6 +1322,28 @@ spec: description: 'How the model data is prepared. Allowed values are: None and Gzip.' type: string + modelAccessConfig: + description: Specifies the access configuration + file for the ML model. You can explicitly + accept the model end-user license agreement + (EULA) within the [model_access_config configuration + block]. see Model Access Config. + items: + properties: + acceptEula: + description: Specifies agreement to the + model end-user license agreement (EULA). + The AcceptEula value must be explicitly + defined as true in order to accept the + EULA that this model requires. You are + responsible for reviewing and complying + with any applicable license terms and + making sure they are acceptable for + your use case before downloading or + using a model. + type: boolean + type: object + type: array s3DataType: description: 'The type of model data to deploy. Allowed values are: S3Object and S3Prefix.' @@ -1112,6 +1363,25 @@ spec: description: The Amazon Resource Name (ARN) of the model package to use to create the model. type: string + multiModelConfig: + description: Specifies additional configuration for multi-model + endpoints. see Multi Model Config. + items: + properties: + modelCacheSetting: + description: 'Whether to cache models for a multi-model + endpoint. By default, multi-model endpoints cache + models so that a model does not have to be loaded + into memory each time it is invoked. Some use cases + do not benefit from model caching. For example, + if an endpoint hosts a large number of models that + are each invoked infrequently, the endpoint might + perform better if you disable model caching. To + disable model caching, set the value of this parameter + to Disabled. Allowed values are: Enabled and Disabled.' + type: string + type: object + type: array type: object type: array tags: diff --git a/package/crds/sagemaker.aws.upbound.io_userprofiles.yaml b/package/crds/sagemaker.aws.upbound.io_userprofiles.yaml index b2b88706ca..11fd66ec0f 100644 --- a/package/crds/sagemaker.aws.upbound.io_userprofiles.yaml +++ b/package/crds/sagemaker.aws.upbound.io_userprofiles.yaml @@ -750,6 +750,28 @@ spec: even if that is the default experience for the domain. Valid values are ENABLED and DISABLED. type: string + studioWebPortalSettings: + description: The Studio Web Portal settings. See studio_web_portal_settings + Block below. + items: + properties: + hiddenAppTypes: + description: The Applications supported in Studio + that are hidden from the Studio left navigation + pane. + items: + type: string + type: array + x-kubernetes-list-type: set + hiddenMlTools: + description: The machine learning tools that are hidden + from the Studio left navigation pane. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array tensorBoardAppSettings: description: The TensorBoard app settings. See TensorBoard App Settings below. @@ -1477,6 +1499,28 @@ spec: even if that is the default experience for the domain. Valid values are ENABLED and DISABLED. type: string + studioWebPortalSettings: + description: The Studio Web Portal settings. See studio_web_portal_settings + Block below. + items: + properties: + hiddenAppTypes: + description: The Applications supported in Studio + that are hidden from the Studio left navigation + pane. + items: + type: string + type: array + x-kubernetes-list-type: set + hiddenMlTools: + description: The machine learning tools that are hidden + from the Studio left navigation pane. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array tensorBoardAppSettings: description: The TensorBoard app settings. See TensorBoard App Settings below. @@ -2310,6 +2354,28 @@ spec: even if that is the default experience for the domain. Valid values are ENABLED and DISABLED. type: string + studioWebPortalSettings: + description: The Studio Web Portal settings. See studio_web_portal_settings + Block below. + items: + properties: + hiddenAppTypes: + description: The Applications supported in Studio + that are hidden from the Studio left navigation + pane. + items: + type: string + type: array + x-kubernetes-list-type: set + hiddenMlTools: + description: The machine learning tools that are hidden + from the Studio left navigation pane. + items: + type: string + type: array + x-kubernetes-list-type: set + type: object + type: array tensorBoardAppSettings: description: The TensorBoard app settings. See TensorBoard App Settings below. diff --git a/package/crds/sesv2.aws.upbound.io_configurationseteventdestinations.yaml b/package/crds/sesv2.aws.upbound.io_configurationseteventdestinations.yaml index f97429403a..afc528d45f 100644 --- a/package/crds/sesv2.aws.upbound.io_configurationseteventdestinations.yaml +++ b/package/crds/sesv2.aws.upbound.io_configurationseteventdestinations.yaml @@ -196,6 +196,16 @@ spec: specified event types are sent to the destinations. Default: false.' type: boolean + eventBridgeDestination: + items: + properties: + eventBusArn: + description: The Amazon Resource Name (ARN) of the + Amazon EventBridge bus to publish email events to. + Only the default bus is supported. + type: string + type: object + type: array kinesisFirehoseDestination: description: An object that defines an Amazon Kinesis Data Firehose destination for email events. See kinesis_firehose_destination @@ -696,6 +706,16 @@ spec: specified event types are sent to the destinations. Default: false.' type: boolean + eventBridgeDestination: + items: + properties: + eventBusArn: + description: The Amazon Resource Name (ARN) of the + Amazon EventBridge bus to publish email events to. + Only the default bus is supported. + type: string + type: object + type: array kinesisFirehoseDestination: description: An object that defines an Amazon Kinesis Data Firehose destination for email events. See kinesis_firehose_destination @@ -1283,6 +1303,16 @@ spec: specified event types are sent to the destinations. Default: false.' type: boolean + eventBridgeDestination: + items: + properties: + eventBusArn: + description: The Amazon Resource Name (ARN) of the + Amazon EventBridge bus to publish email events to. + Only the default bus is supported. + type: string + type: object + type: array kinesisFirehoseDestination: description: An object that defines an Amazon Kinesis Data Firehose destination for email events. See kinesis_firehose_destination diff --git a/package/crds/sfn.aws.upbound.io_statemachines.yaml b/package/crds/sfn.aws.upbound.io_statemachines.yaml index bd800e5c1d..4e11638424 100644 --- a/package/crds/sfn.aws.upbound.io_statemachines.yaml +++ b/package/crds/sfn.aws.upbound.io_statemachines.yaml @@ -77,6 +77,107 @@ spec: description: The Amazon States Language definition of the state machine. type: string + encryptionConfiguration: + description: Defines what encryption configuration is used to + encrypt data in the State Machine. For more information see + [TBD] in the AWS Step Functions User Guide. + items: + properties: + kmsDataKeyReusePeriodSeconds: + description: Maximum duration for which Step Functions will + reuse data keys. When the period expires, Step Functions + will call GenerateDataKey. This setting only applies to + customer managed KMS key and does not apply when type + is AWS_OWNED_KEY. + type: number + kmsKeyId: + description: The alias, alias ARN, key ID, or key ARN of + the symmetric encryption KMS key that encrypts the data + key. To specify a KMS key in a different AWS account, + the customer must use the key ARN or alias ARN. For more + information regarding kms_key_id, see KeyId in the KMS + documentation. + type: string + kmsKeyIdRef: + description: Reference to a Key in kms to populate kmsKeyId. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + kmsKeyIdSelector: + description: Selector for a Key in kms to populate kmsKeyId. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching + labels is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object + type: + description: 'The encryption option specified for the state + machine. Valid values: AWS_OWNED_KEY, CUSTOMER_MANAGED_KMS_KEY' + type: string + type: object + type: array loggingConfiguration: description: Defines what execution history events are logged and where they are logged. The logging_configuration parameter @@ -230,6 +331,107 @@ spec: description: The Amazon States Language definition of the state machine. type: string + encryptionConfiguration: + description: Defines what encryption configuration is used to + encrypt data in the State Machine. For more information see + [TBD] in the AWS Step Functions User Guide. + items: + properties: + kmsDataKeyReusePeriodSeconds: + description: Maximum duration for which Step Functions will + reuse data keys. When the period expires, Step Functions + will call GenerateDataKey. This setting only applies to + customer managed KMS key and does not apply when type + is AWS_OWNED_KEY. + type: number + kmsKeyId: + description: The alias, alias ARN, key ID, or key ARN of + the symmetric encryption KMS key that encrypts the data + key. To specify a KMS key in a different AWS account, + the customer must use the key ARN or alias ARN. For more + information regarding kms_key_id, see KeyId in the KMS + documentation. + type: string + kmsKeyIdRef: + description: Reference to a Key in kms to populate kmsKeyId. + properties: + name: + description: Name of the referenced object. + type: string + policy: + description: Policies for referencing. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + required: + - name + type: object + kmsKeyIdSelector: + description: Selector for a Key in kms to populate kmsKeyId. + properties: + matchControllerRef: + description: |- + MatchControllerRef ensures an object with the same controller reference + as the selecting object is selected. + type: boolean + matchLabels: + additionalProperties: + type: string + description: MatchLabels ensures an object with matching + labels is selected. + type: object + policy: + description: Policies for selection. + properties: + resolution: + default: Required + description: |- + Resolution specifies whether resolution of this reference is required. + The default is 'Required', which means the reconcile will fail if the + reference cannot be resolved. 'Optional' means this reference will be + a no-op if it cannot be resolved. + enum: + - Required + - Optional + type: string + resolve: + description: |- + Resolve specifies when this reference should be resolved. The default + is 'IfNotPresent', which will attempt to resolve the reference only when + the corresponding field is not present. Use 'Always' to resolve the + reference on every reconcile. + enum: + - Always + - IfNotPresent + type: string + type: object + type: object + type: + description: 'The encryption option specified for the state + machine. Valid values: AWS_OWNED_KEY, CUSTOMER_MANAGED_KMS_KEY' + type: string + type: object + type: array loggingConfiguration: description: Defines what execution history events are logged and where they are logged. The logging_configuration parameter @@ -549,6 +751,33 @@ spec: type: string description: type: string + encryptionConfiguration: + description: Defines what encryption configuration is used to + encrypt data in the State Machine. For more information see + [TBD] in the AWS Step Functions User Guide. + items: + properties: + kmsDataKeyReusePeriodSeconds: + description: Maximum duration for which Step Functions will + reuse data keys. When the period expires, Step Functions + will call GenerateDataKey. This setting only applies to + customer managed KMS key and does not apply when type + is AWS_OWNED_KEY. + type: number + kmsKeyId: + description: The alias, alias ARN, key ID, or key ARN of + the symmetric encryption KMS key that encrypts the data + key. To specify a KMS key in a different AWS account, + the customer must use the key ARN or alias ARN. For more + information regarding kms_key_id, see KeyId in the KMS + documentation. + type: string + type: + description: 'The encryption option specified for the state + machine. Valid values: AWS_OWNED_KEY, CUSTOMER_MANAGED_KMS_KEY' + type: string + type: object + type: array id: description: The ARN of the state machine. type: string