@pulumi/aiven 5.6.0 → 6.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (81) hide show
  1. package/README.md +2 -2
  2. package/accountTeamMember.d.ts +6 -6
  3. package/accountTeamProject.d.ts +3 -3
  4. package/azurePrivatelink.d.ts +3 -3
  5. package/cassandra.d.ts +2 -2
  6. package/cassandra.js +2 -2
  7. package/connectionPool.d.ts +3 -3
  8. package/flinkApplicationVersion.d.ts +40 -4
  9. package/flinkApplicationVersion.js +4 -6
  10. package/flinkApplicationVersion.js.map +1 -1
  11. package/getAccountTeamMember.d.ts +6 -6
  12. package/getAccountTeamProject.d.ts +1 -1
  13. package/getAzurePrivatelink.d.ts +1 -1
  14. package/getConnectionPool.d.ts +1 -1
  15. package/getFlinkApplicationVersion.d.ts +8 -0
  16. package/getFlinkApplicationVersion.js.map +1 -1
  17. package/getMirrorMakerReplicationFlow.d.ts +6 -6
  18. package/getOpenSearchAclRule.d.ts +9 -9
  19. package/getServiceIntegration.d.ts +3 -7
  20. package/getServiceIntegration.js.map +1 -1
  21. package/getTransitGatewayVpcAttachment.d.ts +6 -6
  22. package/grafana.d.ts +2 -2
  23. package/grafana.js +2 -2
  24. package/index.d.ts +0 -24
  25. package/index.js +4 -38
  26. package/index.js.map +1 -1
  27. package/influxDb.d.ts +1 -1
  28. package/influxDb.js +1 -1
  29. package/kafka.d.ts +7 -7
  30. package/kafka.js +7 -7
  31. package/kafkaConnect.d.ts +1 -1
  32. package/kafkaConnect.js +1 -1
  33. package/kafkaMirrorMaker.d.ts +3 -3
  34. package/kafkaMirrorMaker.js +3 -3
  35. package/kafkaUser.d.ts +14 -0
  36. package/kafkaUser.js +14 -0
  37. package/kafkaUser.js.map +1 -1
  38. package/mirrorMakerReplicationFlow.d.ts +6 -6
  39. package/mySql.d.ts +2 -2
  40. package/mySql.js +2 -2
  41. package/openSearch.d.ts +4 -4
  42. package/openSearch.js +4 -4
  43. package/openSearchAclConfig.d.ts +1 -1
  44. package/openSearchAclConfig.js +1 -1
  45. package/openSearchAclRule.d.ts +11 -11
  46. package/openSearchAclRule.js +2 -2
  47. package/package.json +2 -2
  48. package/package.json.dev +2 -2
  49. package/redis.d.ts +1 -1
  50. package/redis.js +1 -1
  51. package/serviceIntegration.d.ts +3 -15
  52. package/serviceIntegration.js +0 -2
  53. package/serviceIntegration.js.map +1 -1
  54. package/staticIp.d.ts +2 -2
  55. package/transitGatewayVpcAttachment.d.ts +6 -6
  56. package/types/input.d.ts +646 -626
  57. package/types/output.d.ts +1153 -1061
  58. package/database.d.ts +0 -133
  59. package/database.js +0 -90
  60. package/database.js.map +0 -1
  61. package/flinkJob.d.ts +0 -136
  62. package/flinkJob.js +0 -100
  63. package/flinkJob.js.map +0 -1
  64. package/flinkJobTable.d.ts +0 -268
  65. package/flinkJobTable.js +0 -126
  66. package/flinkJobTable.js.map +0 -1
  67. package/getDatabase.d.ts +0 -103
  68. package/getDatabase.js +0 -57
  69. package/getDatabase.js.map +0 -1
  70. package/getServiceUser.d.ts +0 -130
  71. package/getServiceUser.js +0 -53
  72. package/getServiceUser.js.map +0 -1
  73. package/getVpcPeeringConnection.d.ts +0 -120
  74. package/getVpcPeeringConnection.js +0 -53
  75. package/getVpcPeeringConnection.js.map +0 -1
  76. package/serviceUser.d.ts +0 -204
  77. package/serviceUser.js +0 -108
  78. package/serviceUser.js.map +0 -1
  79. package/vpcPeeringConnection.d.ts +0 -155
  80. package/vpcPeeringConnection.js +0 -83
  81. package/vpcPeeringConnection.js.map +0 -1
@@ -1,268 +0,0 @@
1
- import * as pulumi from "@pulumi/pulumi";
2
- import * as inputs from "./types/input";
3
- import * as outputs from "./types/output";
4
- /**
5
- * The Flink Table resource allows the creation and management of Aiven Tables.
6
- *
7
- * ## Example Usage
8
- *
9
- * ```typescript
10
- * import * as pulumi from "@pulumi/pulumi";
11
- * import * as aiven from "@pulumi/aiven";
12
- *
13
- * const table = new aiven.FlinkJobTable("table", {
14
- * project: data.aiven_project.pr1.project,
15
- * serviceName: aiven_flink.flink.service_name,
16
- * tableName: "<TABLE_NAME>",
17
- * integrationId: aiven_service_integration.flink_kafka.service_id,
18
- * jdbcTable: "<JDBC_TABLE_NAME>",
19
- * kafkaTopic: aiven_kafka_topic.table_topic.topic_name,
20
- * schemaSql: ` \`+"\`cpu\`"+\` INT,
21
- * \`+"\`node\`"+\` INT,
22
- * \`+"\`occurred_at\`"+\` TIMESTAMP(3) METADATA FROM 'timestamp',
23
- * WATERMARK FOR \`+"\`occurred_at\`"+\` AS \`+"\`occurred_at\`"+\` - INTERVAL '5' SECOND
24
- * `,
25
- * });
26
- * ```
27
- *
28
- * ## Import
29
- *
30
- * ```sh
31
- * $ pulumi import aiven:index/flinkJobTable:FlinkJobTable table project/service_name/table_id
32
- * ```
33
- */
34
- export declare class FlinkJobTable extends pulumi.CustomResource {
35
- /**
36
- * Get an existing FlinkJobTable resource's state with the given name, ID, and optional extra
37
- * properties used to qualify the lookup.
38
- *
39
- * @param name The _unique_ name of the resulting resource.
40
- * @param id The _unique_ provider ID of the resource to lookup.
41
- * @param state Any extra arguments used during the lookup.
42
- * @param opts Optional settings to control the behavior of the CustomResource.
43
- */
44
- static get(name: string, id: pulumi.Input<pulumi.ID>, state?: FlinkJobTableState, opts?: pulumi.CustomResourceOptions): FlinkJobTable;
45
- /**
46
- * Returns true if the given object is an instance of FlinkJobTable. This is designed to work even
47
- * when multiple copies of the Pulumi SDK have been loaded into the same process.
48
- */
49
- static isInstance(obj: any): obj is FlinkJobTable;
50
- /**
51
- * The id of the service integration that is used with this table. It must have the service integration type `flink`. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
52
- */
53
- readonly integrationId: pulumi.Output<string>;
54
- /**
55
- * Name of the jdbc table that is to be connected to this table. Valid if the service integration id refers to a mysql or postgres service. This property cannot be changed, doing so forces recreation of the resource.
56
- */
57
- readonly jdbcTable: pulumi.Output<string | undefined>;
58
- /**
59
- * When used as a source, upsert Kafka connectors update values that use an existing key and delete values that are null. For sinks, the connector correspondingly writes update or delete messages in a compacted topic. If no matching key is found, the values are added as new entries. For more information, see the Apache Flink documentation The possible values are `kafka` and `upsert-kafka`. This property cannot be changed, doing so forces recreation of the resource.
60
- */
61
- readonly kafkaConnectorType: pulumi.Output<string | undefined>;
62
- /**
63
- * Defines an explicit list of physical columns from the table schema that configure the data type for the key format. This property cannot be changed, doing so forces recreation of the resource.
64
- */
65
- readonly kafkaKeyFields: pulumi.Output<string[] | undefined>;
66
- /**
67
- * Kafka Key Format The possible values are `avro`, `avro-confluent`, `debezium-avro-confluent`, `debezium-json` and `json`. This property cannot be changed, doing so forces recreation of the resource.
68
- */
69
- readonly kafkaKeyFormat: pulumi.Output<string | undefined>;
70
- /**
71
- * Startup mode The possible values are `earliest-offset`, `latest-offset`, `group-offsets` and `timestamp`. This property cannot be changed, doing so forces recreation of the resource.
72
- */
73
- readonly kafkaStartupMode: pulumi.Output<string | undefined>;
74
- /**
75
- * Name of the kafka topic that is to be connected to this table. Valid if the service integration id refers to a kafka service. This property cannot be changed, doing so forces recreation of the resource.
76
- */
77
- readonly kafkaTopic: pulumi.Output<string | undefined>;
78
- /**
79
- * Controls how key columns are handled in the message value. Select ALL to include the physical columns of the table schema in the message value. Select EXCEPT_KEY to exclude the physical columns of the table schema from the message value. This is the default for upsert Kafka connectors. The possible values are `[ALL EXCEPT_KEY]`. This property cannot be changed, doing so forces recreation of the resource.
80
- */
81
- readonly kafkaValueFieldsInclude: pulumi.Output<string | undefined>;
82
- /**
83
- * Kafka Value Format The possible values are `avro`, `avro-confluent`, `debezium-avro-confluent`, `debezium-json` and `json`. This property cannot be changed, doing so forces recreation of the resource.
84
- */
85
- readonly kafkaValueFormat: pulumi.Output<string | undefined>;
86
- /**
87
- * [LIKE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#like) statement for table creation. This property cannot be changed, doing so forces recreation of the resource.
88
- */
89
- readonly likeOptions: pulumi.Output<string | undefined>;
90
- /**
91
- * For an OpenSearch table, the OpenSearch index the table outputs to. This property cannot be changed, doing so forces recreation of the resource.
92
- */
93
- readonly opensearchIndex: pulumi.Output<string | undefined>;
94
- /**
95
- * Identifies the project this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
96
- */
97
- readonly project: pulumi.Output<string>;
98
- /**
99
- * The SQL statement to create the table. This property cannot be changed, doing so forces recreation of the resource.
100
- */
101
- readonly schemaSql: pulumi.Output<string>;
102
- /**
103
- * Specifies the name of the service that this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
104
- */
105
- readonly serviceName: pulumi.Output<string>;
106
- /**
107
- * The Table ID of the flink table in the flink service.
108
- */
109
- readonly tableId: pulumi.Output<string>;
110
- /**
111
- * Specifies the name of the table. This property cannot be changed, doing so forces recreation of the resource.
112
- */
113
- readonly tableName: pulumi.Output<string>;
114
- /**
115
- * Kafka upsert connector configuration.
116
- */
117
- readonly upsertKafka: pulumi.Output<outputs.FlinkJobTableUpsertKafka | undefined>;
118
- /**
119
- * Create a FlinkJobTable resource with the given unique name, arguments, and options.
120
- *
121
- * @param name The _unique_ name of the resource.
122
- * @param args The arguments to use to populate this resource's properties.
123
- * @param opts A bag of options that control this resource's behavior.
124
- */
125
- constructor(name: string, args: FlinkJobTableArgs, opts?: pulumi.CustomResourceOptions);
126
- }
127
- /**
128
- * Input properties used for looking up and filtering FlinkJobTable resources.
129
- */
130
- export interface FlinkJobTableState {
131
- /**
132
- * The id of the service integration that is used with this table. It must have the service integration type `flink`. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
133
- */
134
- integrationId?: pulumi.Input<string>;
135
- /**
136
- * Name of the jdbc table that is to be connected to this table. Valid if the service integration id refers to a mysql or postgres service. This property cannot be changed, doing so forces recreation of the resource.
137
- */
138
- jdbcTable?: pulumi.Input<string>;
139
- /**
140
- * When used as a source, upsert Kafka connectors update values that use an existing key and delete values that are null. For sinks, the connector correspondingly writes update or delete messages in a compacted topic. If no matching key is found, the values are added as new entries. For more information, see the Apache Flink documentation The possible values are `kafka` and `upsert-kafka`. This property cannot be changed, doing so forces recreation of the resource.
141
- */
142
- kafkaConnectorType?: pulumi.Input<string>;
143
- /**
144
- * Defines an explicit list of physical columns from the table schema that configure the data type for the key format. This property cannot be changed, doing so forces recreation of the resource.
145
- */
146
- kafkaKeyFields?: pulumi.Input<pulumi.Input<string>[]>;
147
- /**
148
- * Kafka Key Format The possible values are `avro`, `avro-confluent`, `debezium-avro-confluent`, `debezium-json` and `json`. This property cannot be changed, doing so forces recreation of the resource.
149
- */
150
- kafkaKeyFormat?: pulumi.Input<string>;
151
- /**
152
- * Startup mode The possible values are `earliest-offset`, `latest-offset`, `group-offsets` and `timestamp`. This property cannot be changed, doing so forces recreation of the resource.
153
- */
154
- kafkaStartupMode?: pulumi.Input<string>;
155
- /**
156
- * Name of the kafka topic that is to be connected to this table. Valid if the service integration id refers to a kafka service. This property cannot be changed, doing so forces recreation of the resource.
157
- */
158
- kafkaTopic?: pulumi.Input<string>;
159
- /**
160
- * Controls how key columns are handled in the message value. Select ALL to include the physical columns of the table schema in the message value. Select EXCEPT_KEY to exclude the physical columns of the table schema from the message value. This is the default for upsert Kafka connectors. The possible values are `[ALL EXCEPT_KEY]`. This property cannot be changed, doing so forces recreation of the resource.
161
- */
162
- kafkaValueFieldsInclude?: pulumi.Input<string>;
163
- /**
164
- * Kafka Value Format The possible values are `avro`, `avro-confluent`, `debezium-avro-confluent`, `debezium-json` and `json`. This property cannot be changed, doing so forces recreation of the resource.
165
- */
166
- kafkaValueFormat?: pulumi.Input<string>;
167
- /**
168
- * [LIKE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#like) statement for table creation. This property cannot be changed, doing so forces recreation of the resource.
169
- */
170
- likeOptions?: pulumi.Input<string>;
171
- /**
172
- * For an OpenSearch table, the OpenSearch index the table outputs to. This property cannot be changed, doing so forces recreation of the resource.
173
- */
174
- opensearchIndex?: pulumi.Input<string>;
175
- /**
176
- * Identifies the project this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
177
- */
178
- project?: pulumi.Input<string>;
179
- /**
180
- * The SQL statement to create the table. This property cannot be changed, doing so forces recreation of the resource.
181
- */
182
- schemaSql?: pulumi.Input<string>;
183
- /**
184
- * Specifies the name of the service that this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
185
- */
186
- serviceName?: pulumi.Input<string>;
187
- /**
188
- * The Table ID of the flink table in the flink service.
189
- */
190
- tableId?: pulumi.Input<string>;
191
- /**
192
- * Specifies the name of the table. This property cannot be changed, doing so forces recreation of the resource.
193
- */
194
- tableName?: pulumi.Input<string>;
195
- /**
196
- * Kafka upsert connector configuration.
197
- */
198
- upsertKafka?: pulumi.Input<inputs.FlinkJobTableUpsertKafka>;
199
- }
200
- /**
201
- * The set of arguments for constructing a FlinkJobTable resource.
202
- */
203
- export interface FlinkJobTableArgs {
204
- /**
205
- * The id of the service integration that is used with this table. It must have the service integration type `flink`. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
206
- */
207
- integrationId: pulumi.Input<string>;
208
- /**
209
- * Name of the jdbc table that is to be connected to this table. Valid if the service integration id refers to a mysql or postgres service. This property cannot be changed, doing so forces recreation of the resource.
210
- */
211
- jdbcTable?: pulumi.Input<string>;
212
- /**
213
- * When used as a source, upsert Kafka connectors update values that use an existing key and delete values that are null. For sinks, the connector correspondingly writes update or delete messages in a compacted topic. If no matching key is found, the values are added as new entries. For more information, see the Apache Flink documentation The possible values are `kafka` and `upsert-kafka`. This property cannot be changed, doing so forces recreation of the resource.
214
- */
215
- kafkaConnectorType?: pulumi.Input<string>;
216
- /**
217
- * Defines an explicit list of physical columns from the table schema that configure the data type for the key format. This property cannot be changed, doing so forces recreation of the resource.
218
- */
219
- kafkaKeyFields?: pulumi.Input<pulumi.Input<string>[]>;
220
- /**
221
- * Kafka Key Format The possible values are `avro`, `avro-confluent`, `debezium-avro-confluent`, `debezium-json` and `json`. This property cannot be changed, doing so forces recreation of the resource.
222
- */
223
- kafkaKeyFormat?: pulumi.Input<string>;
224
- /**
225
- * Startup mode The possible values are `earliest-offset`, `latest-offset`, `group-offsets` and `timestamp`. This property cannot be changed, doing so forces recreation of the resource.
226
- */
227
- kafkaStartupMode?: pulumi.Input<string>;
228
- /**
229
- * Name of the kafka topic that is to be connected to this table. Valid if the service integration id refers to a kafka service. This property cannot be changed, doing so forces recreation of the resource.
230
- */
231
- kafkaTopic?: pulumi.Input<string>;
232
- /**
233
- * Controls how key columns are handled in the message value. Select ALL to include the physical columns of the table schema in the message value. Select EXCEPT_KEY to exclude the physical columns of the table schema from the message value. This is the default for upsert Kafka connectors. The possible values are `[ALL EXCEPT_KEY]`. This property cannot be changed, doing so forces recreation of the resource.
234
- */
235
- kafkaValueFieldsInclude?: pulumi.Input<string>;
236
- /**
237
- * Kafka Value Format The possible values are `avro`, `avro-confluent`, `debezium-avro-confluent`, `debezium-json` and `json`. This property cannot be changed, doing so forces recreation of the resource.
238
- */
239
- kafkaValueFormat?: pulumi.Input<string>;
240
- /**
241
- * [LIKE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#like) statement for table creation. This property cannot be changed, doing so forces recreation of the resource.
242
- */
243
- likeOptions?: pulumi.Input<string>;
244
- /**
245
- * For an OpenSearch table, the OpenSearch index the table outputs to. This property cannot be changed, doing so forces recreation of the resource.
246
- */
247
- opensearchIndex?: pulumi.Input<string>;
248
- /**
249
- * Identifies the project this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
250
- */
251
- project: pulumi.Input<string>;
252
- /**
253
- * The SQL statement to create the table. This property cannot be changed, doing so forces recreation of the resource.
254
- */
255
- schemaSql: pulumi.Input<string>;
256
- /**
257
- * Specifies the name of the service that this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
258
- */
259
- serviceName: pulumi.Input<string>;
260
- /**
261
- * Specifies the name of the table. This property cannot be changed, doing so forces recreation of the resource.
262
- */
263
- tableName: pulumi.Input<string>;
264
- /**
265
- * Kafka upsert connector configuration.
266
- */
267
- upsertKafka?: pulumi.Input<inputs.FlinkJobTableUpsertKafka>;
268
- }
package/flinkJobTable.js DELETED
@@ -1,126 +0,0 @@
1
- "use strict";
2
- // *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. ***
3
- // *** Do not edit by hand unless you're certain you know what you are doing! ***
4
- Object.defineProperty(exports, "__esModule", { value: true });
5
- exports.FlinkJobTable = void 0;
6
- const pulumi = require("@pulumi/pulumi");
7
- const utilities = require("./utilities");
8
- /**
9
- * The Flink Table resource allows the creation and management of Aiven Tables.
10
- *
11
- * ## Example Usage
12
- *
13
- * ```typescript
14
- * import * as pulumi from "@pulumi/pulumi";
15
- * import * as aiven from "@pulumi/aiven";
16
- *
17
- * const table = new aiven.FlinkJobTable("table", {
18
- * project: data.aiven_project.pr1.project,
19
- * serviceName: aiven_flink.flink.service_name,
20
- * tableName: "<TABLE_NAME>",
21
- * integrationId: aiven_service_integration.flink_kafka.service_id,
22
- * jdbcTable: "<JDBC_TABLE_NAME>",
23
- * kafkaTopic: aiven_kafka_topic.table_topic.topic_name,
24
- * schemaSql: ` \`+"\`cpu\`"+\` INT,
25
- * \`+"\`node\`"+\` INT,
26
- * \`+"\`occurred_at\`"+\` TIMESTAMP(3) METADATA FROM 'timestamp',
27
- * WATERMARK FOR \`+"\`occurred_at\`"+\` AS \`+"\`occurred_at\`"+\` - INTERVAL '5' SECOND
28
- * `,
29
- * });
30
- * ```
31
- *
32
- * ## Import
33
- *
34
- * ```sh
35
- * $ pulumi import aiven:index/flinkJobTable:FlinkJobTable table project/service_name/table_id
36
- * ```
37
- */
38
- class FlinkJobTable extends pulumi.CustomResource {
39
- /**
40
- * Get an existing FlinkJobTable resource's state with the given name, ID, and optional extra
41
- * properties used to qualify the lookup.
42
- *
43
- * @param name The _unique_ name of the resulting resource.
44
- * @param id The _unique_ provider ID of the resource to lookup.
45
- * @param state Any extra arguments used during the lookup.
46
- * @param opts Optional settings to control the behavior of the CustomResource.
47
- */
48
- static get(name, id, state, opts) {
49
- return new FlinkJobTable(name, state, Object.assign(Object.assign({}, opts), { id: id }));
50
- }
51
- /**
52
- * Returns true if the given object is an instance of FlinkJobTable. This is designed to work even
53
- * when multiple copies of the Pulumi SDK have been loaded into the same process.
54
- */
55
- static isInstance(obj) {
56
- if (obj === undefined || obj === null) {
57
- return false;
58
- }
59
- return obj['__pulumiType'] === FlinkJobTable.__pulumiType;
60
- }
61
- constructor(name, argsOrState, opts) {
62
- let resourceInputs = {};
63
- opts = opts || {};
64
- if (opts.id) {
65
- const state = argsOrState;
66
- resourceInputs["integrationId"] = state ? state.integrationId : undefined;
67
- resourceInputs["jdbcTable"] = state ? state.jdbcTable : undefined;
68
- resourceInputs["kafkaConnectorType"] = state ? state.kafkaConnectorType : undefined;
69
- resourceInputs["kafkaKeyFields"] = state ? state.kafkaKeyFields : undefined;
70
- resourceInputs["kafkaKeyFormat"] = state ? state.kafkaKeyFormat : undefined;
71
- resourceInputs["kafkaStartupMode"] = state ? state.kafkaStartupMode : undefined;
72
- resourceInputs["kafkaTopic"] = state ? state.kafkaTopic : undefined;
73
- resourceInputs["kafkaValueFieldsInclude"] = state ? state.kafkaValueFieldsInclude : undefined;
74
- resourceInputs["kafkaValueFormat"] = state ? state.kafkaValueFormat : undefined;
75
- resourceInputs["likeOptions"] = state ? state.likeOptions : undefined;
76
- resourceInputs["opensearchIndex"] = state ? state.opensearchIndex : undefined;
77
- resourceInputs["project"] = state ? state.project : undefined;
78
- resourceInputs["schemaSql"] = state ? state.schemaSql : undefined;
79
- resourceInputs["serviceName"] = state ? state.serviceName : undefined;
80
- resourceInputs["tableId"] = state ? state.tableId : undefined;
81
- resourceInputs["tableName"] = state ? state.tableName : undefined;
82
- resourceInputs["upsertKafka"] = state ? state.upsertKafka : undefined;
83
- }
84
- else {
85
- const args = argsOrState;
86
- if ((!args || args.integrationId === undefined) && !opts.urn) {
87
- throw new Error("Missing required property 'integrationId'");
88
- }
89
- if ((!args || args.project === undefined) && !opts.urn) {
90
- throw new Error("Missing required property 'project'");
91
- }
92
- if ((!args || args.schemaSql === undefined) && !opts.urn) {
93
- throw new Error("Missing required property 'schemaSql'");
94
- }
95
- if ((!args || args.serviceName === undefined) && !opts.urn) {
96
- throw new Error("Missing required property 'serviceName'");
97
- }
98
- if ((!args || args.tableName === undefined) && !opts.urn) {
99
- throw new Error("Missing required property 'tableName'");
100
- }
101
- resourceInputs["integrationId"] = args ? args.integrationId : undefined;
102
- resourceInputs["jdbcTable"] = args ? args.jdbcTable : undefined;
103
- resourceInputs["kafkaConnectorType"] = args ? args.kafkaConnectorType : undefined;
104
- resourceInputs["kafkaKeyFields"] = args ? args.kafkaKeyFields : undefined;
105
- resourceInputs["kafkaKeyFormat"] = args ? args.kafkaKeyFormat : undefined;
106
- resourceInputs["kafkaStartupMode"] = args ? args.kafkaStartupMode : undefined;
107
- resourceInputs["kafkaTopic"] = args ? args.kafkaTopic : undefined;
108
- resourceInputs["kafkaValueFieldsInclude"] = args ? args.kafkaValueFieldsInclude : undefined;
109
- resourceInputs["kafkaValueFormat"] = args ? args.kafkaValueFormat : undefined;
110
- resourceInputs["likeOptions"] = args ? args.likeOptions : undefined;
111
- resourceInputs["opensearchIndex"] = args ? args.opensearchIndex : undefined;
112
- resourceInputs["project"] = args ? args.project : undefined;
113
- resourceInputs["schemaSql"] = args ? args.schemaSql : undefined;
114
- resourceInputs["serviceName"] = args ? args.serviceName : undefined;
115
- resourceInputs["tableName"] = args ? args.tableName : undefined;
116
- resourceInputs["upsertKafka"] = args ? args.upsertKafka : undefined;
117
- resourceInputs["tableId"] = undefined /*out*/;
118
- }
119
- opts = pulumi.mergeOptions(utilities.resourceOptsDefaults(), opts);
120
- super(FlinkJobTable.__pulumiType, name, resourceInputs, opts);
121
- }
122
- }
123
- exports.FlinkJobTable = FlinkJobTable;
124
- /** @internal */
125
- FlinkJobTable.__pulumiType = 'aiven:index/flinkJobTable:FlinkJobTable';
126
- //# sourceMappingURL=flinkJobTable.js.map
@@ -1 +0,0 @@
1
- {"version":3,"file":"flinkJobTable.js","sourceRoot":"","sources":["../flinkJobTable.ts"],"names":[],"mappings":";AAAA,wFAAwF;AACxF,iFAAiF;;;AAEjF,yCAAyC;AAGzC,yCAAyC;AAEzC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA6BG;AACH,MAAa,aAAc,SAAQ,MAAM,CAAC,cAAc;IACpD;;;;;;;;OAQG;IACI,MAAM,CAAC,GAAG,CAAC,IAAY,EAAE,EAA2B,EAAE,KAA0B,EAAE,IAAmC;QACxH,OAAO,IAAI,aAAa,CAAC,IAAI,EAAO,KAAK,kCAAO,IAAI,KAAE,EAAE,EAAE,EAAE,IAAG,CAAC;IACpE,CAAC;IAKD;;;OAGG;IACI,MAAM,CAAC,UAAU,CAAC,GAAQ;QAC7B,IAAI,GAAG,KAAK,SAAS,IAAI,GAAG,KAAK,IAAI,EAAE;YACnC,OAAO,KAAK,CAAC;SAChB;QACD,OAAO,GAAG,CAAC,cAAc,CAAC,KAAK,aAAa,CAAC,YAAY,CAAC;IAC9D,CAAC;IA+ED,YAAY,IAAY,EAAE,WAAoD,EAAE,IAAmC;QAC/G,IAAI,cAAc,GAAkB,EAAE,CAAC;QACvC,IAAI,GAAG,IAAI,IAAI,EAAE,CAAC;QAClB,IAAI,IAAI,CAAC,EAAE,EAAE;YACT,MAAM,KAAK,GAAG,WAA6C,CAAC;YAC5D,cAAc,CAAC,eAAe,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,aAAa,CAAC,CAAC,CAAC,SAAS,CAAC;YAC1E,cAAc,CAAC,WAAW,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,SAAS,CAAC,CAAC,CAAC,SAAS,CAAC;YAClE,cAAc,CAAC,oBAAoB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,kBAAkB,CAAC,CAAC,CAAC,SAAS,CAAC;YACpF,cAAc,CAAC,gBAAgB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,cAAc,CAAC,CAAC,CAAC,SAAS,CAAC;YAC5E,cAAc,CAAC,gBAAgB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,cAAc,CAAC,CAAC,CAAC,SAAS,CAAC;YAC5E,cAAc,CAAC,kBAAkB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,CAAC,CAAC,SAAS,CAAC;YAChF,cAAc,CAAC,YAAY,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,UAAU,CAAC,CAAC,CAAC,SAAS,CAAC;YACpE,cAAc,CAAC,yBAAyB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,uBAAuB,CAAC,CAAC,CAAC,SAAS,CAAC;YAC9F,cAAc,CAAC,kBAAkB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,gBAAgB,CAAC,CAAC,CAAC,SAAS,CAAC;YAChF,cAAc,CAAC,aAAa,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,CAAC,CAAC,SAAS,CAAC;YACtE,cAAc,CAAC,iBAAiB,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,eAAe,CAAC,CAAC,CAAC,SAAS,CAAC;YAC9E,cAAc,CAAC,SAAS,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,CAAC,SAAS,CAAC;YAC9D,cAAc,CAAC,WAAW,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,SAAS,CAAC,CAAC,CAAC,SAAS,CAAC;YAClE,cAAc,CAAC,aAAa,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,CAAC,CAAC,SAAS,CAAC;YACtE,cAAc,CAAC,SAAS,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,CAAC,SAAS,CAAC;YAC9D,cAAc,CAAC,WAAW,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,SAAS,CAAC,CAAC,CAAC,SAAS,CAAC;YAClE,cAAc,CAAC,aAAa,CAAC,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,WAAW,CAAC,CAAC,CAAC,SAAS,CAAC;SACzE;aAAM;YACH,MAAM,IAAI,GAAG,WAA4C,CAAC;YAC1D,IAAI,CAAC,CAAC,IAAI,IAAI,IAAI,CAAC,aAAa,KAAK,SAAS,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,EAAE;gBAC1D,MAAM,IAAI,KAAK,CAAC,2CAA2C,CAAC,CAAC;aAChE;YACD,IAAI,CAAC,CAAC,IAAI,IAAI,IAAI,CAAC,OAAO,KAAK,SAAS,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,EAAE;gBACpD,MAAM,IAAI,KAAK,CAAC,qCAAqC,CAAC,CAAC;aAC1D;YACD,IAAI,CAAC,CAAC,IAAI,IAAI,IAAI,CAAC,SAAS,KAAK,SAAS,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,EAAE;gBACtD,MAAM,IAAI,KAAK,CAAC,uCAAuC,CAAC,CAAC;aAC5D;YACD,IAAI,CAAC,CAAC,IAAI,IAAI,IAAI,CAAC,WAAW,KAAK,SAAS,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,EAAE;gBACxD,MAAM,IAAI,KAAK,CAAC,yCAAyC,CAAC,CAAC;aAC9D;YACD,IAAI,CAAC,CAAC,IAAI,IAAI,IAAI,CAAC,SAAS,KAAK,SAAS,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,EAAE;gBACtD,MAAM,IAAI,KAAK,CAAC,uCAAuC,CAAC,CAAC;aAC5D;YACD,cAAc,CAAC,eAAe,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,aAAa,CAAC,CAAC,CAAC,SAAS,CAAC;YACxE,cAAc,CAAC,WAAW,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,SAAS,CAAC,CAAC,CAAC,SAAS,CAAC;YAChE,cAAc,CAAC,oBAAoB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,kBAAkB,CAAC,CAAC,CAAC,SAAS,CAAC;YAClF,cAAc,CAAC,gBAAgB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,cAAc,CAAC,CAAC,CAAC,SAAS,CAAC;YAC1E,cAAc,CAAC,gBAAgB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,cAAc,CAAC,CAAC,CAAC,SAAS,CAAC;YAC1E,cAAc,CAAC,kBAAkB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,gBAAgB,CAAC,CAAC,CAAC,SAAS,CAAC;YAC9E,cAAc,CAAC,YAAY,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,UAAU,CAAC,CAAC,CAAC,SAAS,CAAC;YAClE,cAAc,CAAC,yBAAyB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,uBAAuB,CAAC,CAAC,CAAC,SAAS,CAAC;YAC5F,cAAc,CAAC,kBAAkB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,gBAAgB,CAAC,CAAC,CAAC,SAAS,CAAC;YAC9E,cAAc,CAAC,aAAa,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,WAAW,CAAC,CAAC,CAAC,SAAS,CAAC;YACpE,cAAc,CAAC,iBAAiB,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,eAAe,CAAC,CAAC,CAAC,SAAS,CAAC;YAC5E,cAAc,CAAC,SAAS,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,OAAO,CAAC,CAAC,CAAC,SAAS,CAAC;YAC5D,cAAc,CAAC,WAAW,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,SAAS,CAAC,CAAC,CAAC,SAAS,CAAC;YAChE,cAAc,CAAC,aAAa,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,WAAW,CAAC,CAAC,CAAC,SAAS,CAAC;YACpE,cAAc,CAAC,WAAW,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,SAAS,CAAC,CAAC,CAAC,SAAS,CAAC;YAChE,cAAc,CAAC,aAAa,CAAC,GAAG,IAAI,CAAC,CAAC,CAAC,IAAI,CAAC,WAAW,CAAC,CAAC,CAAC,SAAS,CAAC;YACpE,cAAc,CAAC,SAAS,CAAC,GAAG,SAAS,CAAC,OAAO,CAAC;SACjD;QACD,IAAI,GAAG,MAAM,CAAC,YAAY,CAAC,SAAS,CAAC,oBAAoB,EAAE,EAAE,IAAI,CAAC,CAAC;QACnE,KAAK,CAAC,aAAa,CAAC,YAAY,EAAE,IAAI,EAAE,cAAc,EAAE,IAAI,CAAC,CAAC;IAClE,CAAC;;AApKL,sCAqKC;AAvJG,gBAAgB;AACO,0BAAY,GAAG,yCAAyC,CAAC"}
package/getDatabase.d.ts DELETED
@@ -1,103 +0,0 @@
1
- import * as pulumi from "@pulumi/pulumi";
2
- /**
3
- * The Database data source provides information about the existing Aiven Database.
4
- *
5
- * ~>**Deprecated** The Database data source is deprecated, please use service-specific data sources instead, for example: `aiven.PgDatabase`, `aiven.MysqlDatabase` etc.
6
- *
7
- * ## Example Usage
8
- *
9
- * ```typescript
10
- * import * as pulumi from "@pulumi/pulumi";
11
- * import * as aiven from "@pulumi/aiven";
12
- *
13
- * const mydatabase = aiven.getDatabase({
14
- * project: aiven_project.myproject.project,
15
- * serviceName: aiven_pg.mypg.service_name,
16
- * databaseName: "<DATABASE_NAME>",
17
- * });
18
- * ```
19
- */
20
- export declare function getDatabase(args: GetDatabaseArgs, opts?: pulumi.InvokeOptions): Promise<GetDatabaseResult>;
21
- /**
22
- * A collection of arguments for invoking getDatabase.
23
- */
24
- export interface GetDatabaseArgs {
25
- /**
26
- * The name of the service database. This property cannot be changed, doing so forces recreation of the resource.
27
- */
28
- databaseName: string;
29
- /**
30
- * Identifies the project this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
31
- */
32
- project: string;
33
- /**
34
- * Specifies the name of the service that this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
35
- */
36
- serviceName: string;
37
- }
38
- /**
39
- * A collection of values returned by getDatabase.
40
- */
41
- export interface GetDatabaseResult {
42
- /**
43
- * The name of the service database. This property cannot be changed, doing so forces recreation of the resource.
44
- */
45
- readonly databaseName: string;
46
- /**
47
- * The provider-assigned unique ID for this managed resource.
48
- */
49
- readonly id: string;
50
- /**
51
- * Default string sort order (`LC_COLLATE`) of the database. The default value is `en_US.UTF-8`. This property cannot be changed, doing so forces recreation of the resource.
52
- */
53
- readonly lcCollate: string;
54
- /**
55
- * Default character classification (`LC_CTYPE`) of the database. The default value is `en_US.UTF-8`. This property cannot be changed, doing so forces recreation of the resource.
56
- */
57
- readonly lcCtype: string;
58
- /**
59
- * Identifies the project this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
60
- */
61
- readonly project: string;
62
- /**
63
- * Specifies the name of the service that this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
64
- */
65
- readonly serviceName: string;
66
- readonly terminationProtection: boolean;
67
- }
68
- /**
69
- * The Database data source provides information about the existing Aiven Database.
70
- *
71
- * ~>**Deprecated** The Database data source is deprecated, please use service-specific data sources instead, for example: `aiven.PgDatabase`, `aiven.MysqlDatabase` etc.
72
- *
73
- * ## Example Usage
74
- *
75
- * ```typescript
76
- * import * as pulumi from "@pulumi/pulumi";
77
- * import * as aiven from "@pulumi/aiven";
78
- *
79
- * const mydatabase = aiven.getDatabase({
80
- * project: aiven_project.myproject.project,
81
- * serviceName: aiven_pg.mypg.service_name,
82
- * databaseName: "<DATABASE_NAME>",
83
- * });
84
- * ```
85
- */
86
- export declare function getDatabaseOutput(args: GetDatabaseOutputArgs, opts?: pulumi.InvokeOptions): pulumi.Output<GetDatabaseResult>;
87
- /**
88
- * A collection of arguments for invoking getDatabase.
89
- */
90
- export interface GetDatabaseOutputArgs {
91
- /**
92
- * The name of the service database. This property cannot be changed, doing so forces recreation of the resource.
93
- */
94
- databaseName: pulumi.Input<string>;
95
- /**
96
- * Identifies the project this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
97
- */
98
- project: pulumi.Input<string>;
99
- /**
100
- * Specifies the name of the service that this resource belongs to. To set up proper dependencies please refer to this variable as a reference. This property cannot be changed, doing so forces recreation of the resource.
101
- */
102
- serviceName: pulumi.Input<string>;
103
- }
package/getDatabase.js DELETED
@@ -1,57 +0,0 @@
1
- "use strict";
2
- // *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. ***
3
- // *** Do not edit by hand unless you're certain you know what you are doing! ***
4
- Object.defineProperty(exports, "__esModule", { value: true });
5
- exports.getDatabaseOutput = exports.getDatabase = void 0;
6
- const pulumi = require("@pulumi/pulumi");
7
- const utilities = require("./utilities");
8
- /**
9
- * The Database data source provides information about the existing Aiven Database.
10
- *
11
- * ~>**Deprecated** The Database data source is deprecated, please use service-specific data sources instead, for example: `aiven.PgDatabase`, `aiven.MysqlDatabase` etc.
12
- *
13
- * ## Example Usage
14
- *
15
- * ```typescript
16
- * import * as pulumi from "@pulumi/pulumi";
17
- * import * as aiven from "@pulumi/aiven";
18
- *
19
- * const mydatabase = aiven.getDatabase({
20
- * project: aiven_project.myproject.project,
21
- * serviceName: aiven_pg.mypg.service_name,
22
- * databaseName: "<DATABASE_NAME>",
23
- * });
24
- * ```
25
- */
26
- function getDatabase(args, opts) {
27
- opts = pulumi.mergeOptions(utilities.resourceOptsDefaults(), opts || {});
28
- return pulumi.runtime.invoke("aiven:index/getDatabase:getDatabase", {
29
- "databaseName": args.databaseName,
30
- "project": args.project,
31
- "serviceName": args.serviceName,
32
- }, opts);
33
- }
34
- exports.getDatabase = getDatabase;
35
- /**
36
- * The Database data source provides information about the existing Aiven Database.
37
- *
38
- * ~>**Deprecated** The Database data source is deprecated, please use service-specific data sources instead, for example: `aiven.PgDatabase`, `aiven.MysqlDatabase` etc.
39
- *
40
- * ## Example Usage
41
- *
42
- * ```typescript
43
- * import * as pulumi from "@pulumi/pulumi";
44
- * import * as aiven from "@pulumi/aiven";
45
- *
46
- * const mydatabase = aiven.getDatabase({
47
- * project: aiven_project.myproject.project,
48
- * serviceName: aiven_pg.mypg.service_name,
49
- * databaseName: "<DATABASE_NAME>",
50
- * });
51
- * ```
52
- */
53
- function getDatabaseOutput(args, opts) {
54
- return pulumi.output(args).apply((a) => getDatabase(a, opts));
55
- }
56
- exports.getDatabaseOutput = getDatabaseOutput;
57
- //# sourceMappingURL=getDatabase.js.map
@@ -1 +0,0 @@
1
- {"version":3,"file":"getDatabase.js","sourceRoot":"","sources":["../getDatabase.ts"],"names":[],"mappings":";AAAA,wFAAwF;AACxF,iFAAiF;;;AAEjF,yCAAyC;AACzC,yCAAyC;AAEzC;;;;;;;;;;;;;;;;;GAiBG;AACH,SAAgB,WAAW,CAAC,IAAqB,EAAE,IAA2B;IAE1E,IAAI,GAAG,MAAM,CAAC,YAAY,CAAC,SAAS,CAAC,oBAAoB,EAAE,EAAE,IAAI,IAAI,EAAE,CAAC,CAAC;IACzE,OAAO,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,qCAAqC,EAAE;QAChE,cAAc,EAAE,IAAI,CAAC,YAAY;QACjC,SAAS,EAAE,IAAI,CAAC,OAAO;QACvB,aAAa,EAAE,IAAI,CAAC,WAAW;KAClC,EAAE,IAAI,CAAC,CAAC;AACb,CAAC;AARD,kCAQC;AAkDD;;;;;;;;;;;;;;;;;GAiBG;AACH,SAAgB,iBAAiB,CAAC,IAA2B,EAAE,IAA2B;IACtF,OAAO,MAAM,CAAC,MAAM,CAAC,IAAI,CAAC,CAAC,KAAK,CAAC,CAAC,CAAM,EAAE,EAAE,CAAC,WAAW,CAAC,CAAC,EAAE,IAAI,CAAC,CAAC,CAAA;AACtE,CAAC;AAFD,8CAEC"}