@pulumi/mongodbatlas 3.37.0-alpha.1762410787 → 3.37.0-alpha.1763072353

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (80) hide show
  1. package/dataLakePipeline.d.ts +54 -0
  2. package/dataLakePipeline.js +54 -0
  3. package/dataLakePipeline.js.map +1 -1
  4. package/eventTrigger.d.ts +2 -2
  5. package/eventTrigger.js +2 -2
  6. package/flexCluster.d.ts +26 -0
  7. package/flexCluster.js +26 -0
  8. package/flexCluster.js.map +1 -1
  9. package/get509AuthenticationDatabaseUser.d.ts +58 -2
  10. package/get509AuthenticationDatabaseUser.js +58 -2
  11. package/get509AuthenticationDatabaseUser.js.map +1 -1
  12. package/getAccessListApiKey.d.ts +30 -0
  13. package/getAccessListApiKey.js +30 -0
  14. package/getAccessListApiKey.js.map +1 -1
  15. package/getAccessListApiKeys.d.ts +30 -0
  16. package/getAccessListApiKeys.js +30 -0
  17. package/getAccessListApiKeys.js.map +1 -1
  18. package/getCloudBackupSchedule.d.ts +108 -0
  19. package/getCloudBackupSchedule.js +108 -0
  20. package/getCloudBackupSchedule.js.map +1 -1
  21. package/getCloudBackupSnapshotRestoreJob.d.ts +54 -0
  22. package/getCloudBackupSnapshotRestoreJob.js +54 -0
  23. package/getCloudBackupSnapshotRestoreJob.js.map +1 -1
  24. package/getCustomDbRoles.d.ts +50 -0
  25. package/getCustomDbRoles.js +50 -0
  26. package/getCustomDbRoles.js.map +1 -1
  27. package/getDataLakePipeline.d.ts +116 -0
  28. package/getDataLakePipeline.js +116 -0
  29. package/getDataLakePipeline.js.map +1 -1
  30. package/getDataLakePipelineRuns.d.ts +74 -0
  31. package/getDataLakePipelineRuns.js +74 -0
  32. package/getDataLakePipelineRuns.js.map +1 -1
  33. package/getFederatedSettingsOrgRoleMappings.d.ts +68 -0
  34. package/getFederatedSettingsOrgRoleMappings.js +68 -0
  35. package/getFederatedSettingsOrgRoleMappings.js.map +1 -1
  36. package/getFlexCluster.d.ts +50 -0
  37. package/getFlexCluster.js +50 -0
  38. package/getFlexCluster.js.map +1 -1
  39. package/getFlexClusters.d.ts +50 -0
  40. package/getFlexClusters.js +50 -0
  41. package/getFlexClusters.js.map +1 -1
  42. package/getSearchDeployment.d.ts +78 -0
  43. package/getSearchDeployment.js +78 -0
  44. package/getSearchDeployment.js.map +1 -1
  45. package/getStreamPrivatelinkEndpoint.d.ts +140 -0
  46. package/getStreamPrivatelinkEndpoint.js +140 -0
  47. package/getStreamPrivatelinkEndpoint.js.map +1 -1
  48. package/getStreamPrivatelinkEndpoints.d.ts +140 -0
  49. package/getStreamPrivatelinkEndpoints.js +140 -0
  50. package/getStreamPrivatelinkEndpoints.js.map +1 -1
  51. package/getStreamProcessor.d.ts +264 -0
  52. package/getStreamProcessor.js +264 -0
  53. package/getStreamProcessor.js.map +1 -1
  54. package/getStreamProcessors.d.ts +264 -0
  55. package/getStreamProcessors.js +264 -0
  56. package/getStreamProcessors.js.map +1 -1
  57. package/getX509AuthenticationDatabaseUser.d.ts +58 -2
  58. package/getX509AuthenticationDatabaseUser.js +58 -2
  59. package/getX509AuthenticationDatabaseUser.js.map +1 -1
  60. package/package.json +2 -2
  61. package/searchDeployment.d.ts +40 -0
  62. package/searchDeployment.js +40 -0
  63. package/searchDeployment.js.map +1 -1
  64. package/searchIndex.d.ts +39 -39
  65. package/searchIndex.js +39 -39
  66. package/streamConnection.d.ts +40 -0
  67. package/streamConnection.js +40 -0
  68. package/streamConnection.js.map +1 -1
  69. package/streamInstance.d.ts +14 -0
  70. package/streamInstance.js +14 -0
  71. package/streamInstance.js.map +1 -1
  72. package/streamPrivatelinkEndpoint.d.ts +70 -0
  73. package/streamPrivatelinkEndpoint.js +70 -0
  74. package/streamPrivatelinkEndpoint.js.map +1 -1
  75. package/streamProcessor.d.ts +133 -0
  76. package/streamProcessor.js +133 -0
  77. package/streamProcessor.js.map +1 -1
  78. package/x509authenticationDatabaseUser.d.ts +63 -0
  79. package/x509authenticationDatabaseUser.js +63 -0
  80. package/x509authenticationDatabaseUser.js.map +1 -1
@@ -9,6 +9,76 @@ import * as outputs from "./types/output";
9
9
  *
10
10
  * ### S
11
11
  *
12
+ * ### AWS Confluent Privatelink
13
+ * ```typescript
14
+ * import * as pulumi from "@pulumi/pulumi";
15
+ * import * as confluent from "@pulumi/confluent";
16
+ * import * as mongodbatlas from "@pulumi/mongodbatlas";
17
+ * import * as std from "@pulumi/std";
18
+ *
19
+ * const staging = new confluent.index.Environment("staging", {displayName: "Staging"});
20
+ * const privateLink = new confluent.index.Network("private_link", {
21
+ * displayName: "terraform-test-private-link-network-manual",
22
+ * cloud: "AWS",
23
+ * region: awsRegion,
24
+ * connectionTypes: ["PRIVATELINK"],
25
+ * zones: std.index.keys({
26
+ * input: subnetsToPrivatelink,
27
+ * }).result,
28
+ * environment: [{
29
+ * id: staging.id,
30
+ * }],
31
+ * dnsConfig: [{
32
+ * resolution: "PRIVATE",
33
+ * }],
34
+ * });
35
+ * const aws = new confluent.index.PrivateLinkAccess("aws", {
36
+ * displayName: "example-private-link-access",
37
+ * aws: [{
38
+ * account: awsAccountId,
39
+ * }],
40
+ * environment: [{
41
+ * id: staging.id,
42
+ * }],
43
+ * network: [{
44
+ * id: privateLink.id,
45
+ * }],
46
+ * });
47
+ * const dedicated = new confluent.index.KafkaCluster("dedicated", {
48
+ * displayName: "example-dedicated-cluster",
49
+ * availability: "MULTI_ZONE",
50
+ * cloud: privateLink.cloud,
51
+ * region: privateLink.region,
52
+ * dedicated: [{
53
+ * cku: 2,
54
+ * }],
55
+ * environment: [{
56
+ * id: staging.id,
57
+ * }],
58
+ * network: [{
59
+ * id: privateLink.id,
60
+ * }],
61
+ * });
62
+ * const test = new mongodbatlas.StreamPrivatelinkEndpoint("test", {
63
+ * projectId: projectId,
64
+ * dnsDomain: privateLink.dnsDomain,
65
+ * providerName: "AWS",
66
+ * region: awsRegion,
67
+ * vendor: "CONFLUENT",
68
+ * serviceEndpointId: privateLink.aws[0].privateLinkEndpointService,
69
+ * dnsSubDomains: privateLink.zonalSubdomains,
70
+ * });
71
+ * const singularDatasource = test.id.apply(id => mongodbatlas.getStreamPrivatelinkEndpointOutput({
72
+ * projectId: projectId,
73
+ * id: id,
74
+ * }));
75
+ * const pluralDatasource = mongodbatlas.getStreamPrivatelinkEndpoints({
76
+ * projectId: projectId,
77
+ * });
78
+ * export const interfaceEndpointId = singularDatasource.apply(singularDatasource => singularDatasource.interfaceEndpointId);
79
+ * export const interfaceEndpointIds = pluralDatasource.then(pluralDatasource => pluralDatasource.results.map(__item => __item.interfaceEndpointId));
80
+ * ```
81
+ *
12
82
  * ### AWS S3 Privatelink
13
83
  * ```typescript
14
84
  * import * as pulumi from "@pulumi/pulumi";
@@ -72,6 +142,76 @@ export interface GetStreamPrivatelinkEndpointsResult {
72
142
  *
73
143
  * ### S
74
144
  *
145
+ * ### AWS Confluent Privatelink
146
+ * ```typescript
147
+ * import * as pulumi from "@pulumi/pulumi";
148
+ * import * as confluent from "@pulumi/confluent";
149
+ * import * as mongodbatlas from "@pulumi/mongodbatlas";
150
+ * import * as std from "@pulumi/std";
151
+ *
152
+ * const staging = new confluent.index.Environment("staging", {displayName: "Staging"});
153
+ * const privateLink = new confluent.index.Network("private_link", {
154
+ * displayName: "terraform-test-private-link-network-manual",
155
+ * cloud: "AWS",
156
+ * region: awsRegion,
157
+ * connectionTypes: ["PRIVATELINK"],
158
+ * zones: std.index.keys({
159
+ * input: subnetsToPrivatelink,
160
+ * }).result,
161
+ * environment: [{
162
+ * id: staging.id,
163
+ * }],
164
+ * dnsConfig: [{
165
+ * resolution: "PRIVATE",
166
+ * }],
167
+ * });
168
+ * const aws = new confluent.index.PrivateLinkAccess("aws", {
169
+ * displayName: "example-private-link-access",
170
+ * aws: [{
171
+ * account: awsAccountId,
172
+ * }],
173
+ * environment: [{
174
+ * id: staging.id,
175
+ * }],
176
+ * network: [{
177
+ * id: privateLink.id,
178
+ * }],
179
+ * });
180
+ * const dedicated = new confluent.index.KafkaCluster("dedicated", {
181
+ * displayName: "example-dedicated-cluster",
182
+ * availability: "MULTI_ZONE",
183
+ * cloud: privateLink.cloud,
184
+ * region: privateLink.region,
185
+ * dedicated: [{
186
+ * cku: 2,
187
+ * }],
188
+ * environment: [{
189
+ * id: staging.id,
190
+ * }],
191
+ * network: [{
192
+ * id: privateLink.id,
193
+ * }],
194
+ * });
195
+ * const test = new mongodbatlas.StreamPrivatelinkEndpoint("test", {
196
+ * projectId: projectId,
197
+ * dnsDomain: privateLink.dnsDomain,
198
+ * providerName: "AWS",
199
+ * region: awsRegion,
200
+ * vendor: "CONFLUENT",
201
+ * serviceEndpointId: privateLink.aws[0].privateLinkEndpointService,
202
+ * dnsSubDomains: privateLink.zonalSubdomains,
203
+ * });
204
+ * const singularDatasource = test.id.apply(id => mongodbatlas.getStreamPrivatelinkEndpointOutput({
205
+ * projectId: projectId,
206
+ * id: id,
207
+ * }));
208
+ * const pluralDatasource = mongodbatlas.getStreamPrivatelinkEndpoints({
209
+ * projectId: projectId,
210
+ * });
211
+ * export const interfaceEndpointId = singularDatasource.apply(singularDatasource => singularDatasource.interfaceEndpointId);
212
+ * export const interfaceEndpointIds = pluralDatasource.then(pluralDatasource => pluralDatasource.results.map(__item => __item.interfaceEndpointId));
213
+ * ```
214
+ *
75
215
  * ### AWS S3 Privatelink
76
216
  * ```typescript
77
217
  * import * as pulumi from "@pulumi/pulumi";
@@ -14,6 +14,76 @@ const utilities = require("./utilities");
14
14
  *
15
15
  * ### S
16
16
  *
17
+ * ### AWS Confluent Privatelink
18
+ * ```typescript
19
+ * import * as pulumi from "@pulumi/pulumi";
20
+ * import * as confluent from "@pulumi/confluent";
21
+ * import * as mongodbatlas from "@pulumi/mongodbatlas";
22
+ * import * as std from "@pulumi/std";
23
+ *
24
+ * const staging = new confluent.index.Environment("staging", {displayName: "Staging"});
25
+ * const privateLink = new confluent.index.Network("private_link", {
26
+ * displayName: "terraform-test-private-link-network-manual",
27
+ * cloud: "AWS",
28
+ * region: awsRegion,
29
+ * connectionTypes: ["PRIVATELINK"],
30
+ * zones: std.index.keys({
31
+ * input: subnetsToPrivatelink,
32
+ * }).result,
33
+ * environment: [{
34
+ * id: staging.id,
35
+ * }],
36
+ * dnsConfig: [{
37
+ * resolution: "PRIVATE",
38
+ * }],
39
+ * });
40
+ * const aws = new confluent.index.PrivateLinkAccess("aws", {
41
+ * displayName: "example-private-link-access",
42
+ * aws: [{
43
+ * account: awsAccountId,
44
+ * }],
45
+ * environment: [{
46
+ * id: staging.id,
47
+ * }],
48
+ * network: [{
49
+ * id: privateLink.id,
50
+ * }],
51
+ * });
52
+ * const dedicated = new confluent.index.KafkaCluster("dedicated", {
53
+ * displayName: "example-dedicated-cluster",
54
+ * availability: "MULTI_ZONE",
55
+ * cloud: privateLink.cloud,
56
+ * region: privateLink.region,
57
+ * dedicated: [{
58
+ * cku: 2,
59
+ * }],
60
+ * environment: [{
61
+ * id: staging.id,
62
+ * }],
63
+ * network: [{
64
+ * id: privateLink.id,
65
+ * }],
66
+ * });
67
+ * const test = new mongodbatlas.StreamPrivatelinkEndpoint("test", {
68
+ * projectId: projectId,
69
+ * dnsDomain: privateLink.dnsDomain,
70
+ * providerName: "AWS",
71
+ * region: awsRegion,
72
+ * vendor: "CONFLUENT",
73
+ * serviceEndpointId: privateLink.aws[0].privateLinkEndpointService,
74
+ * dnsSubDomains: privateLink.zonalSubdomains,
75
+ * });
76
+ * const singularDatasource = test.id.apply(id => mongodbatlas.getStreamPrivatelinkEndpointOutput({
77
+ * projectId: projectId,
78
+ * id: id,
79
+ * }));
80
+ * const pluralDatasource = mongodbatlas.getStreamPrivatelinkEndpoints({
81
+ * projectId: projectId,
82
+ * });
83
+ * export const interfaceEndpointId = singularDatasource.apply(singularDatasource => singularDatasource.interfaceEndpointId);
84
+ * export const interfaceEndpointIds = pluralDatasource.then(pluralDatasource => pluralDatasource.results.map(__item => __item.interfaceEndpointId));
85
+ * ```
86
+ *
17
87
  * ### AWS S3 Privatelink
18
88
  * ```typescript
19
89
  * import * as pulumi from "@pulumi/pulumi";
@@ -66,6 +136,76 @@ exports.getStreamPrivatelinkEndpoints = getStreamPrivatelinkEndpoints;
66
136
  *
67
137
  * ### S
68
138
  *
139
+ * ### AWS Confluent Privatelink
140
+ * ```typescript
141
+ * import * as pulumi from "@pulumi/pulumi";
142
+ * import * as confluent from "@pulumi/confluent";
143
+ * import * as mongodbatlas from "@pulumi/mongodbatlas";
144
+ * import * as std from "@pulumi/std";
145
+ *
146
+ * const staging = new confluent.index.Environment("staging", {displayName: "Staging"});
147
+ * const privateLink = new confluent.index.Network("private_link", {
148
+ * displayName: "terraform-test-private-link-network-manual",
149
+ * cloud: "AWS",
150
+ * region: awsRegion,
151
+ * connectionTypes: ["PRIVATELINK"],
152
+ * zones: std.index.keys({
153
+ * input: subnetsToPrivatelink,
154
+ * }).result,
155
+ * environment: [{
156
+ * id: staging.id,
157
+ * }],
158
+ * dnsConfig: [{
159
+ * resolution: "PRIVATE",
160
+ * }],
161
+ * });
162
+ * const aws = new confluent.index.PrivateLinkAccess("aws", {
163
+ * displayName: "example-private-link-access",
164
+ * aws: [{
165
+ * account: awsAccountId,
166
+ * }],
167
+ * environment: [{
168
+ * id: staging.id,
169
+ * }],
170
+ * network: [{
171
+ * id: privateLink.id,
172
+ * }],
173
+ * });
174
+ * const dedicated = new confluent.index.KafkaCluster("dedicated", {
175
+ * displayName: "example-dedicated-cluster",
176
+ * availability: "MULTI_ZONE",
177
+ * cloud: privateLink.cloud,
178
+ * region: privateLink.region,
179
+ * dedicated: [{
180
+ * cku: 2,
181
+ * }],
182
+ * environment: [{
183
+ * id: staging.id,
184
+ * }],
185
+ * network: [{
186
+ * id: privateLink.id,
187
+ * }],
188
+ * });
189
+ * const test = new mongodbatlas.StreamPrivatelinkEndpoint("test", {
190
+ * projectId: projectId,
191
+ * dnsDomain: privateLink.dnsDomain,
192
+ * providerName: "AWS",
193
+ * region: awsRegion,
194
+ * vendor: "CONFLUENT",
195
+ * serviceEndpointId: privateLink.aws[0].privateLinkEndpointService,
196
+ * dnsSubDomains: privateLink.zonalSubdomains,
197
+ * });
198
+ * const singularDatasource = test.id.apply(id => mongodbatlas.getStreamPrivatelinkEndpointOutput({
199
+ * projectId: projectId,
200
+ * id: id,
201
+ * }));
202
+ * const pluralDatasource = mongodbatlas.getStreamPrivatelinkEndpoints({
203
+ * projectId: projectId,
204
+ * });
205
+ * export const interfaceEndpointId = singularDatasource.apply(singularDatasource => singularDatasource.interfaceEndpointId);
206
+ * export const interfaceEndpointIds = pluralDatasource.then(pluralDatasource => pluralDatasource.results.map(__item => __item.interfaceEndpointId));
207
+ * ```
208
+ *
69
209
  * ### AWS S3 Privatelink
70
210
  * ```typescript
71
211
  * import * as pulumi from "@pulumi/pulumi";
@@ -1 +1 @@
1
- {"version":3,"file":"getStreamPrivatelinkEndpoints.js","sourceRoot":"","sources":["../getStreamPrivatelinkEndpoints.ts"],"names":[],"mappings":";AAAA,sEAAsE;AACtE,iFAAiF;;;AAEjF,yCAAyC;AAGzC,yCAAyC;AAEzC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA4CG;AACH,SAAgB,6BAA6B,CAAC,IAAuC,EAAE,IAA2B;IAC9G,IAAI,GAAG,MAAM,CAAC,YAAY,CAAC,SAAS,CAAC,oBAAoB,EAAE,EAAE,IAAI,IAAI,EAAE,CAAC,CAAC;IACzE,OAAO,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,gFAAgF,EAAE;QAC3G,WAAW,EAAE,IAAI,CAAC,SAAS;KAC9B,EAAE,IAAI,CAAC,CAAC;AACb,CAAC;AALD,sEAKC;AAoBD;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GA4CG;AACH,SAAgB,mCAAmC,CAAC,IAA6C,EAAE,IAAiC;IAChI,IAAI,GAAG,MAAM,CAAC,YAAY,CAAC,SAAS,CAAC,oBAAoB,EAAE,EAAE,IAAI,IAAI,EAAE,CAAC,CAAC;IACzE,OAAO,MAAM,CAAC,OAAO,CAAC,YAAY,CAAC,gFAAgF,EAAE;QACjH,WAAW,EAAE,IAAI,CAAC,SAAS;KAC9B,EAAE,IAAI,CAAC,CAAC;AACb,CAAC;AALD,kFAKC"}
1
+ {"version":3,"file":"getStreamPrivatelinkEndpoints.js","sourceRoot":"","sources":["../getStreamPrivatelinkEndpoints.ts"],"names":[],"mappings":";AAAA,sEAAsE;AACtE,iFAAiF;;;AAEjF,yCAAyC;AAGzC,yCAAyC;AAEzC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAkHG;AACH,SAAgB,6BAA6B,CAAC,IAAuC,EAAE,IAA2B;IAC9G,IAAI,GAAG,MAAM,CAAC,YAAY,CAAC,SAAS,CAAC,oBAAoB,EAAE,EAAE,IAAI,IAAI,EAAE,CAAC,CAAC;IACzE,OAAO,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,gFAAgF,EAAE;QAC3G,WAAW,EAAE,IAAI,CAAC,SAAS;KAC9B,EAAE,IAAI,CAAC,CAAC;AACb,CAAC;AALD,sEAKC;AAoBD;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAkHG;AACH,SAAgB,mCAAmC,CAAC,IAA6C,EAAE,IAAiC;IAChI,IAAI,GAAG,MAAM,CAAC,YAAY,CAAC,SAAS,CAAC,oBAAoB,EAAE,EAAE,IAAI,IAAI,EAAE,CAAC,CAAC;IACzE,OAAO,MAAM,CAAC,OAAO,CAAC,YAAY,CAAC,gFAAgF,EAAE;QACjH,WAAW,EAAE,IAAI,CAAC,SAAS;KAC9B,EAAE,IAAI,CAAC,CAAC;AACb,CAAC;AALD,kFAKC"}
@@ -8,6 +8,138 @@ import * as outputs from "./types/output";
8
8
  * ## Example Usage
9
9
  *
10
10
  * ### S
11
+ * ```typescript
12
+ * import * as pulumi from "@pulumi/pulumi";
13
+ * import * as mongodbatlas from "@pulumi/mongodbatlas";
14
+ *
15
+ * const example = new mongodbatlas.StreamInstance("example", {
16
+ * projectId: projectId,
17
+ * instanceName: "InstanceName",
18
+ * dataProcessRegion: {
19
+ * region: "VIRGINIA_USA",
20
+ * cloudProvider: "AWS",
21
+ * },
22
+ * });
23
+ * const example_sample = new mongodbatlas.StreamConnection("example-sample", {
24
+ * projectId: projectId,
25
+ * instanceName: example.instanceName,
26
+ * connectionName: "sample_stream_solar",
27
+ * type: "Sample",
28
+ * });
29
+ * const example_cluster = new mongodbatlas.StreamConnection("example-cluster", {
30
+ * projectId: projectId,
31
+ * instanceName: example.instanceName,
32
+ * connectionName: "ClusterConnection",
33
+ * type: "Cluster",
34
+ * clusterName: clusterName,
35
+ * dbRoleToExecute: {
36
+ * role: "atlasAdmin",
37
+ * type: "BUILT_IN",
38
+ * },
39
+ * });
40
+ * const example_kafka = new mongodbatlas.StreamConnection("example-kafka", {
41
+ * projectId: projectId,
42
+ * instanceName: example.instanceName,
43
+ * connectionName: "KafkaPlaintextConnection",
44
+ * type: "Kafka",
45
+ * authentication: {
46
+ * mechanism: "PLAIN",
47
+ * username: kafkaUsername,
48
+ * password: kafkaPassword,
49
+ * },
50
+ * bootstrapServers: "localhost:9092,localhost:9092",
51
+ * config: {
52
+ * "auto.offset.reset": "earliest",
53
+ * },
54
+ * security: {
55
+ * protocol: "SASL_PLAINTEXT",
56
+ * },
57
+ * });
58
+ * const stream_processor_sample_example = new mongodbatlas.StreamProcessor("stream-processor-sample-example", {
59
+ * projectId: projectId,
60
+ * instanceName: example.instanceName,
61
+ * processorName: "sampleProcessorName",
62
+ * pipeline: JSON.stringify([
63
+ * {
64
+ * $source: {
65
+ * connectionName: mongodbatlasStreamConnection["example-sample"].connectionName,
66
+ * },
67
+ * },
68
+ * {
69
+ * $emit: {
70
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
71
+ * db: "sample",
72
+ * coll: "solar",
73
+ * timeseries: {
74
+ * timeField: "_ts",
75
+ * },
76
+ * },
77
+ * },
78
+ * ]),
79
+ * state: "STARTED",
80
+ * });
81
+ * const stream_processor_cluster_to_kafka_example = new mongodbatlas.StreamProcessor("stream-processor-cluster-to-kafka-example", {
82
+ * projectId: projectId,
83
+ * instanceName: example.instanceName,
84
+ * processorName: "clusterProcessorName",
85
+ * pipeline: JSON.stringify([
86
+ * {
87
+ * $source: {
88
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
89
+ * },
90
+ * },
91
+ * {
92
+ * $emit: {
93
+ * connectionName: mongodbatlasStreamConnection["example-kafka"].connectionName,
94
+ * topic: "topic_from_cluster",
95
+ * },
96
+ * },
97
+ * ]),
98
+ * state: "CREATED",
99
+ * });
100
+ * const stream_processor_kafka_to_cluster_example = new mongodbatlas.StreamProcessor("stream-processor-kafka-to-cluster-example", {
101
+ * projectId: projectId,
102
+ * instanceName: example.instanceName,
103
+ * processorName: "kafkaProcessorName",
104
+ * pipeline: JSON.stringify([
105
+ * {
106
+ * $source: {
107
+ * connectionName: mongodbatlasStreamConnection["example-kafka"].connectionName,
108
+ * topic: "topic_source",
109
+ * },
110
+ * },
111
+ * {
112
+ * $emit: {
113
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
114
+ * db: "kafka",
115
+ * coll: "topic_source",
116
+ * timeseries: {
117
+ * timeField: "ts",
118
+ * },
119
+ * },
120
+ * },
121
+ * ]),
122
+ * state: "CREATED",
123
+ * options: {
124
+ * dlq: {
125
+ * coll: "exampleColumn",
126
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
127
+ * db: "exampleDb",
128
+ * },
129
+ * },
130
+ * });
131
+ * const example_stream_processors = example.instanceName.apply(instanceName => mongodbatlas.getStreamProcessorsOutput({
132
+ * projectId: projectId,
133
+ * instanceName: instanceName,
134
+ * }));
135
+ * const example_stream_processor = pulumi.all([example.instanceName, stream_processor_sample_example.processorName]).apply(([instanceName, processorName]) => mongodbatlas.getStreamProcessorOutput({
136
+ * projectId: projectId,
137
+ * instanceName: instanceName,
138
+ * processorName: processorName,
139
+ * }));
140
+ * export const streamProcessorsState = example_stream_processor.apply(example_stream_processor => example_stream_processor.state);
141
+ * export const streamProcessorsResults = example_stream_processors.apply(example_stream_processors => example_stream_processors.results);
142
+ * ```
11
143
  */
12
144
  export declare function getStreamProcessor(args: GetStreamProcessorArgs, opts?: pulumi.InvokeOptions): Promise<GetStreamProcessorResult>;
13
145
  /**
@@ -57,6 +189,138 @@ export interface GetStreamProcessorResult {
57
189
  * ## Example Usage
58
190
  *
59
191
  * ### S
192
+ * ```typescript
193
+ * import * as pulumi from "@pulumi/pulumi";
194
+ * import * as mongodbatlas from "@pulumi/mongodbatlas";
195
+ *
196
+ * const example = new mongodbatlas.StreamInstance("example", {
197
+ * projectId: projectId,
198
+ * instanceName: "InstanceName",
199
+ * dataProcessRegion: {
200
+ * region: "VIRGINIA_USA",
201
+ * cloudProvider: "AWS",
202
+ * },
203
+ * });
204
+ * const example_sample = new mongodbatlas.StreamConnection("example-sample", {
205
+ * projectId: projectId,
206
+ * instanceName: example.instanceName,
207
+ * connectionName: "sample_stream_solar",
208
+ * type: "Sample",
209
+ * });
210
+ * const example_cluster = new mongodbatlas.StreamConnection("example-cluster", {
211
+ * projectId: projectId,
212
+ * instanceName: example.instanceName,
213
+ * connectionName: "ClusterConnection",
214
+ * type: "Cluster",
215
+ * clusterName: clusterName,
216
+ * dbRoleToExecute: {
217
+ * role: "atlasAdmin",
218
+ * type: "BUILT_IN",
219
+ * },
220
+ * });
221
+ * const example_kafka = new mongodbatlas.StreamConnection("example-kafka", {
222
+ * projectId: projectId,
223
+ * instanceName: example.instanceName,
224
+ * connectionName: "KafkaPlaintextConnection",
225
+ * type: "Kafka",
226
+ * authentication: {
227
+ * mechanism: "PLAIN",
228
+ * username: kafkaUsername,
229
+ * password: kafkaPassword,
230
+ * },
231
+ * bootstrapServers: "localhost:9092,localhost:9092",
232
+ * config: {
233
+ * "auto.offset.reset": "earliest",
234
+ * },
235
+ * security: {
236
+ * protocol: "SASL_PLAINTEXT",
237
+ * },
238
+ * });
239
+ * const stream_processor_sample_example = new mongodbatlas.StreamProcessor("stream-processor-sample-example", {
240
+ * projectId: projectId,
241
+ * instanceName: example.instanceName,
242
+ * processorName: "sampleProcessorName",
243
+ * pipeline: JSON.stringify([
244
+ * {
245
+ * $source: {
246
+ * connectionName: mongodbatlasStreamConnection["example-sample"].connectionName,
247
+ * },
248
+ * },
249
+ * {
250
+ * $emit: {
251
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
252
+ * db: "sample",
253
+ * coll: "solar",
254
+ * timeseries: {
255
+ * timeField: "_ts",
256
+ * },
257
+ * },
258
+ * },
259
+ * ]),
260
+ * state: "STARTED",
261
+ * });
262
+ * const stream_processor_cluster_to_kafka_example = new mongodbatlas.StreamProcessor("stream-processor-cluster-to-kafka-example", {
263
+ * projectId: projectId,
264
+ * instanceName: example.instanceName,
265
+ * processorName: "clusterProcessorName",
266
+ * pipeline: JSON.stringify([
267
+ * {
268
+ * $source: {
269
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
270
+ * },
271
+ * },
272
+ * {
273
+ * $emit: {
274
+ * connectionName: mongodbatlasStreamConnection["example-kafka"].connectionName,
275
+ * topic: "topic_from_cluster",
276
+ * },
277
+ * },
278
+ * ]),
279
+ * state: "CREATED",
280
+ * });
281
+ * const stream_processor_kafka_to_cluster_example = new mongodbatlas.StreamProcessor("stream-processor-kafka-to-cluster-example", {
282
+ * projectId: projectId,
283
+ * instanceName: example.instanceName,
284
+ * processorName: "kafkaProcessorName",
285
+ * pipeline: JSON.stringify([
286
+ * {
287
+ * $source: {
288
+ * connectionName: mongodbatlasStreamConnection["example-kafka"].connectionName,
289
+ * topic: "topic_source",
290
+ * },
291
+ * },
292
+ * {
293
+ * $emit: {
294
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
295
+ * db: "kafka",
296
+ * coll: "topic_source",
297
+ * timeseries: {
298
+ * timeField: "ts",
299
+ * },
300
+ * },
301
+ * },
302
+ * ]),
303
+ * state: "CREATED",
304
+ * options: {
305
+ * dlq: {
306
+ * coll: "exampleColumn",
307
+ * connectionName: mongodbatlasStreamConnection["example-cluster"].connectionName,
308
+ * db: "exampleDb",
309
+ * },
310
+ * },
311
+ * });
312
+ * const example_stream_processors = example.instanceName.apply(instanceName => mongodbatlas.getStreamProcessorsOutput({
313
+ * projectId: projectId,
314
+ * instanceName: instanceName,
315
+ * }));
316
+ * const example_stream_processor = pulumi.all([example.instanceName, stream_processor_sample_example.processorName]).apply(([instanceName, processorName]) => mongodbatlas.getStreamProcessorOutput({
317
+ * projectId: projectId,
318
+ * instanceName: instanceName,
319
+ * processorName: processorName,
320
+ * }));
321
+ * export const streamProcessorsState = example_stream_processor.apply(example_stream_processor => example_stream_processor.state);
322
+ * export const streamProcessorsResults = example_stream_processors.apply(example_stream_processors => example_stream_processors.results);
323
+ * ```
60
324
  */
61
325
  export declare function getStreamProcessorOutput(args: GetStreamProcessorOutputArgs, opts?: pulumi.InvokeOutputOptions): pulumi.Output<GetStreamProcessorResult>;
62
326
  /**