@aws-sdk/client-gameliftstreams 3.953.0 → 3.955.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist-cjs/index.js CHANGED
@@ -311,11 +311,13 @@ const _LTFRRi = "ListTagsForResourceResponse";
311
311
  const _LUA = "LastUpdatedAt";
312
312
  const _Lo = "Location";
313
313
  const _M = "Message";
314
+ const _MC = "MaximumCapacity";
314
315
  const _MR = "MaxResults";
315
316
  const _NT = "NextToken";
316
317
  const _ODC = "OnDemandCapacity";
317
318
  const _OU = "OutputUri";
318
319
  const _P = "Protocol";
320
+ const _PSC = "PerformanceStatsConfiguration";
319
321
  const _RA = "ResourceArn";
320
322
  const _RC = "RequestedCapacity";
321
323
  const _RE = "RuntimeEnvironment";
@@ -340,8 +342,10 @@ const _SSSI = "StartStreamSessionInput";
340
342
  const _SSSL = "StreamSessionSummaryList";
341
343
  const _SSSO = "StartStreamSessionOutput";
342
344
  const _SSSt = "StartStreamSession";
345
+ const _SWC = "SharedWithClient";
343
346
  const _T = "Tags";
344
347
  const _TE = "ThrottlingException";
348
+ const _TIC = "TargetIdleCapacity";
345
349
  const _TK = "TagKeys";
346
350
  const _TR = "TagResource";
347
351
  const _TRR = "TagResourceRequest";
@@ -506,7 +510,7 @@ var GetStreamSessionOutput$ = [
506
510
  n0,
507
511
  _GSSO,
508
512
  0,
509
- [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
513
+ [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _PSC, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
510
514
  [
511
515
  0,
512
516
  0,
@@ -522,6 +526,7 @@ var GetStreamSessionOutput$ = [
522
526
  1,
523
527
  64 | 0,
524
528
  128 | 0,
529
+ () => PerformanceStatsConfiguration$,
525
530
  0,
526
531
  0,
527
532
  4,
@@ -615,15 +620,23 @@ var ListStreamSessionsOutput$ = [
615
620
  ];
616
621
  var ListTagsForResourceRequest$ = [3, n0, _LTFRR, 0, [_RA], [[0, 1]]];
617
622
  var ListTagsForResourceResponse$ = [3, n0, _LTFRRi, 0, [_T], [128 | 0]];
618
- var LocationConfiguration$ = [3, n0, _LCo, 0, [_LN, _AOC, _ODC], [0, 1, 1]];
623
+ var LocationConfiguration$ = [
624
+ 3,
625
+ n0,
626
+ _LCo,
627
+ 0,
628
+ [_LN, _AOC, _ODC, _TIC, _MC],
629
+ [0, 1, 1, 1, 1],
630
+ ];
619
631
  var LocationState$ = [
620
632
  3,
621
633
  n0,
622
634
  _LSo,
623
635
  0,
624
- [_LN, _S, _AOC, _ODC, _RC, _AC, _IC],
625
- [0, 0, 1, 1, 1, 1, 1],
636
+ [_LN, _S, _AOC, _ODC, _TIC, _MC, _RC, _AC, _IC],
637
+ [0, 0, 1, 1, 1, 1, 1, 1, 1],
626
638
  ];
639
+ var PerformanceStatsConfiguration$ = [3, n0, _PSC, 0, [_SWC], [2]];
627
640
  var RemoveStreamGroupLocationsInput$ = [
628
641
  3,
629
642
  n0,
@@ -646,15 +659,29 @@ var StartStreamSessionInput$ = [
646
659
  n0,
647
660
  _SSSI,
648
661
  0,
649
- [_CT, _D, _I, _P, _SR, _AIp, _UI, _L, _CTS, _SLS, _ALA, _AEV],
650
- [[0, 4], 0, [0, 1], 0, [() => SignalRequest, 0], 0, 0, 64 | 0, 1, 1, 64 | 0, 128 | 0],
662
+ [_CT, _D, _I, _P, _SR, _AIp, _UI, _L, _CTS, _SLS, _ALA, _AEV, _PSC],
663
+ [
664
+ [0, 4],
665
+ 0,
666
+ [0, 1],
667
+ 0,
668
+ [() => SignalRequest, 0],
669
+ 0,
670
+ 0,
671
+ 64 | 0,
672
+ 1,
673
+ 1,
674
+ 64 | 0,
675
+ 128 | 0,
676
+ () => PerformanceStatsConfiguration$,
677
+ ],
651
678
  ];
652
679
  var StartStreamSessionOutput$ = [
653
680
  3,
654
681
  n0,
655
682
  _SSSO,
656
683
  0,
657
- [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
684
+ [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _PSC, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
658
685
  [
659
686
  0,
660
687
  0,
@@ -670,6 +697,7 @@ var StartStreamSessionOutput$ = [
670
697
  1,
671
698
  64 | 0,
672
699
  128 | 0,
700
+ () => PerformanceStatsConfiguration$,
673
701
  0,
674
702
  0,
675
703
  4,
@@ -691,8 +719,8 @@ var StreamSessionSummary$ = [
691
719
  n0,
692
720
  _SSS,
693
721
  0,
694
- [_A, _UI, _S, _P, _LUA, _CA, _AAp, _EFM, _Lo],
695
- [0, 0, 0, 0, 4, 4, 0, () => ExportFilesMetadata$, 0],
722
+ [_A, _UI, _S, _SRt, _P, _LUA, _CA, _AAp, _EFM, _Lo],
723
+ [0, 0, 0, 0, 0, 4, 4, 0, () => ExportFilesMetadata$, 0],
696
724
  ];
697
725
  var TagResourceRequest$ = [3, n0, _TRR, 0, [_RA, _T], [[0, 1], 128 | 0]];
698
726
  var TagResourceResponse$ = [3, n0, _TRRa, 0, [], []];
@@ -1506,6 +1534,13 @@ const StreamClass = {
1506
1534
  gen5n_high: "gen5n_high",
1507
1535
  gen5n_ultra: "gen5n_ultra",
1508
1536
  gen5n_win2022: "gen5n_win2022",
1537
+ gen6n_high: "gen6n_high",
1538
+ gen6n_medium: "gen6n_medium",
1539
+ gen6n_pro: "gen6n_pro",
1540
+ gen6n_pro_win2022: "gen6n_pro_win2022",
1541
+ gen6n_small: "gen6n_small",
1542
+ gen6n_ultra: "gen6n_ultra",
1543
+ gen6n_ultra_win2022: "gen6n_ultra_win2022",
1509
1544
  };
1510
1545
  const StreamGroupStatus = {
1511
1546
  ACTIVATING: "ACTIVATING",
@@ -1643,6 +1678,7 @@ exports.ListTagsForResourceRequest$ = ListTagsForResourceRequest$;
1643
1678
  exports.ListTagsForResourceResponse$ = ListTagsForResourceResponse$;
1644
1679
  exports.LocationConfiguration$ = LocationConfiguration$;
1645
1680
  exports.LocationState$ = LocationState$;
1681
+ exports.PerformanceStatsConfiguration$ = PerformanceStatsConfiguration$;
1646
1682
  exports.Protocol = Protocol;
1647
1683
  exports.RemoveStreamGroupLocations$ = RemoveStreamGroupLocations$;
1648
1684
  exports.RemoveStreamGroupLocationsCommand = RemoveStreamGroupLocationsCommand;
package/dist-es/index.js CHANGED
@@ -6,4 +6,5 @@ export * from "./pagination";
6
6
  export * from "./waiters";
7
7
  export * from "./models/enums";
8
8
  export * from "./models/errors";
9
+ export * from "./models/models_0";
9
10
  export { GameLiftStreamsServiceException } from "./models/GameLiftStreamsServiceException";
@@ -31,6 +31,13 @@ export const StreamClass = {
31
31
  gen5n_high: "gen5n_high",
32
32
  gen5n_ultra: "gen5n_ultra",
33
33
  gen5n_win2022: "gen5n_win2022",
34
+ gen6n_high: "gen6n_high",
35
+ gen6n_medium: "gen6n_medium",
36
+ gen6n_pro: "gen6n_pro",
37
+ gen6n_pro_win2022: "gen6n_pro_win2022",
38
+ gen6n_small: "gen6n_small",
39
+ gen6n_ultra: "gen6n_ultra",
40
+ gen6n_ultra_win2022: "gen6n_ultra_win2022",
34
41
  };
35
42
  export const StreamGroupStatus = {
36
43
  ACTIVATING: "ACTIVATING",
@@ -90,11 +90,13 @@ const _LTFRRi = "ListTagsForResourceResponse";
90
90
  const _LUA = "LastUpdatedAt";
91
91
  const _Lo = "Location";
92
92
  const _M = "Message";
93
+ const _MC = "MaximumCapacity";
93
94
  const _MR = "MaxResults";
94
95
  const _NT = "NextToken";
95
96
  const _ODC = "OnDemandCapacity";
96
97
  const _OU = "OutputUri";
97
98
  const _P = "Protocol";
99
+ const _PSC = "PerformanceStatsConfiguration";
98
100
  const _RA = "ResourceArn";
99
101
  const _RC = "RequestedCapacity";
100
102
  const _RE = "RuntimeEnvironment";
@@ -119,8 +121,10 @@ const _SSSI = "StartStreamSessionInput";
119
121
  const _SSSL = "StreamSessionSummaryList";
120
122
  const _SSSO = "StartStreamSessionOutput";
121
123
  const _SSSt = "StartStreamSession";
124
+ const _SWC = "SharedWithClient";
122
125
  const _T = "Tags";
123
126
  const _TE = "ThrottlingException";
127
+ const _TIC = "TargetIdleCapacity";
124
128
  const _TK = "TagKeys";
125
129
  const _TR = "TagResource";
126
130
  const _TRR = "TagResourceRequest";
@@ -288,7 +292,7 @@ export var GetStreamSessionOutput$ = [
288
292
  n0,
289
293
  _GSSO,
290
294
  0,
291
- [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
295
+ [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _PSC, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
292
296
  [
293
297
  0,
294
298
  0,
@@ -304,6 +308,7 @@ export var GetStreamSessionOutput$ = [
304
308
  1,
305
309
  64 | 0,
306
310
  128 | 0,
311
+ () => PerformanceStatsConfiguration$,
307
312
  0,
308
313
  0,
309
314
  4,
@@ -397,15 +402,23 @@ export var ListStreamSessionsOutput$ = [
397
402
  ];
398
403
  export var ListTagsForResourceRequest$ = [3, n0, _LTFRR, 0, [_RA], [[0, 1]]];
399
404
  export var ListTagsForResourceResponse$ = [3, n0, _LTFRRi, 0, [_T], [128 | 0]];
400
- export var LocationConfiguration$ = [3, n0, _LCo, 0, [_LN, _AOC, _ODC], [0, 1, 1]];
405
+ export var LocationConfiguration$ = [
406
+ 3,
407
+ n0,
408
+ _LCo,
409
+ 0,
410
+ [_LN, _AOC, _ODC, _TIC, _MC],
411
+ [0, 1, 1, 1, 1],
412
+ ];
401
413
  export var LocationState$ = [
402
414
  3,
403
415
  n0,
404
416
  _LSo,
405
417
  0,
406
- [_LN, _S, _AOC, _ODC, _RC, _AC, _IC],
407
- [0, 0, 1, 1, 1, 1, 1],
418
+ [_LN, _S, _AOC, _ODC, _TIC, _MC, _RC, _AC, _IC],
419
+ [0, 0, 1, 1, 1, 1, 1, 1, 1],
408
420
  ];
421
+ export var PerformanceStatsConfiguration$ = [3, n0, _PSC, 0, [_SWC], [2]];
409
422
  export var RemoveStreamGroupLocationsInput$ = [
410
423
  3,
411
424
  n0,
@@ -428,15 +441,29 @@ export var StartStreamSessionInput$ = [
428
441
  n0,
429
442
  _SSSI,
430
443
  0,
431
- [_CT, _D, _I, _P, _SR, _AIp, _UI, _L, _CTS, _SLS, _ALA, _AEV],
432
- [[0, 4], 0, [0, 1], 0, [() => SignalRequest, 0], 0, 0, 64 | 0, 1, 1, 64 | 0, 128 | 0],
444
+ [_CT, _D, _I, _P, _SR, _AIp, _UI, _L, _CTS, _SLS, _ALA, _AEV, _PSC],
445
+ [
446
+ [0, 4],
447
+ 0,
448
+ [0, 1],
449
+ 0,
450
+ [() => SignalRequest, 0],
451
+ 0,
452
+ 0,
453
+ 64 | 0,
454
+ 1,
455
+ 1,
456
+ 64 | 0,
457
+ 128 | 0,
458
+ () => PerformanceStatsConfiguration$,
459
+ ],
433
460
  ];
434
461
  export var StartStreamSessionOutput$ = [
435
462
  3,
436
463
  n0,
437
464
  _SSSO,
438
465
  0,
439
- [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
466
+ [_A, _D, _SGI, _UI, _S, _SRt, _P, _Lo, _SR, _SRi, _CTS, _SLS, _ALA, _AEV, _PSC, _LFLU, _WSPU, _LUA, _CA, _AAp, _EFM],
440
467
  [
441
468
  0,
442
469
  0,
@@ -452,6 +479,7 @@ export var StartStreamSessionOutput$ = [
452
479
  1,
453
480
  64 | 0,
454
481
  128 | 0,
482
+ () => PerformanceStatsConfiguration$,
455
483
  0,
456
484
  0,
457
485
  4,
@@ -473,8 +501,8 @@ export var StreamSessionSummary$ = [
473
501
  n0,
474
502
  _SSS,
475
503
  0,
476
- [_A, _UI, _S, _P, _LUA, _CA, _AAp, _EFM, _Lo],
477
- [0, 0, 0, 0, 4, 4, 0, () => ExportFilesMetadata$, 0],
504
+ [_A, _UI, _S, _SRt, _P, _LUA, _CA, _AAp, _EFM, _Lo],
505
+ [0, 0, 0, 0, 0, 4, 4, 0, () => ExportFilesMetadata$, 0],
478
506
  ];
479
507
  export var TagResourceRequest$ = [3, n0, _TRR, 0, [_RA, _T], [[0, 1], 128 | 0]];
480
508
  export var TagResourceResponse$ = [3, n0, _TRRa, 0, [], []];
@@ -43,6 +43,8 @@ declare const AddStreamGroupLocationsCommand_base: {
43
43
  * LocationName: "STRING_VALUE", // required
44
44
  * AlwaysOnCapacity: Number("int"),
45
45
  * OnDemandCapacity: Number("int"),
46
+ * TargetIdleCapacity: Number("int"),
47
+ * MaximumCapacity: Number("int"),
46
48
  * },
47
49
  * ],
48
50
  * };
@@ -56,6 +58,8 @@ declare const AddStreamGroupLocationsCommand_base: {
56
58
  * // Status: "ACTIVATING" || "ACTIVE" || "ERROR" || "REMOVING",
57
59
  * // AlwaysOnCapacity: Number("int"),
58
60
  * // OnDemandCapacity: Number("int"),
61
+ * // TargetIdleCapacity: Number("int"),
62
+ * // MaximumCapacity: Number("int"),
59
63
  * // RequestedCapacity: Number("int"),
60
64
  * // AllocatedCapacity: Number("int"),
61
65
  * // IdleCapacity: Number("int"),
@@ -27,7 +27,7 @@ declare const CreateStreamGroupCommand_base: {
27
27
  getEndpointParameterInstructions(): import("@smithy/middleware-endpoint").EndpointParameterInstructions;
28
28
  };
29
29
  /**
30
- * <p> Stream groups manage how Amazon GameLift Streams allocates resources and handles concurrent streams, allowing you to effectively manage capacity and costs. Within a stream group, you specify an application to stream, streaming locations and their capacity, and the stream class you want to use when streaming applications to your end-users. A stream class defines the hardware configuration of the compute resources that Amazon GameLift Streams will use when streaming, such as the CPU, GPU, and memory. </p> <p> Stream capacity represents the number of concurrent streams that can be active at a time. You set stream capacity per location, per stream group. There are two types of capacity, always-on and on-demand: </p> <ul> <li> <p> <b>Always-on</b>: The streaming capacity that is allocated and ready to handle stream requests without delay. You pay for this capacity whether it's in use or not. Best for quickest time from streaming request to streaming session. Default is 1 (2 for high stream classes) when creating a stream group or adding a location. </p> </li> <li> <p> <b>On-demand</b>: The streaming capacity that Amazon GameLift Streams can allocate in response to stream requests, and then de-allocate when the session has terminated. This offers a cost control measure at the expense of a greater startup time (typically under 5 minutes). Default is 0 when creating a stream group or adding a location. </p> </li> </ul> <p>Values for capacity must be whole number multiples of the tenancy value of the stream group's stream class.</p> <p> To adjust the capacity of any <code>ACTIVE</code> stream group, call <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_UpdateStreamGroup.html">UpdateStreamGroup</a>. </p> <p> If the <code>CreateStreamGroup</code> request is successful, Amazon GameLift Streams assigns a unique ID to the stream group resource and sets the status to <code>ACTIVATING</code>. It can take a few minutes for Amazon GameLift Streams to finish creating the stream group while it searches for unallocated compute resources and provisions them. When complete, the stream group status will be <code>ACTIVE</code> and you can start stream sessions by using <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_StartStreamSession.html">StartStreamSession</a>. To check the stream group's status, call <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_GetStreamGroup.html">GetStreamGroup</a>. </p> <p>Stream groups should be recreated every 3-4 weeks to pick up important service updates and fixes. Stream groups that are older than 180 days can no longer be updated with new application associations. Stream groups expire when they are 365 days old, at which point they can no longer stream sessions. The exact expiration date is indicated by the date value in the <code>ExpiresAt</code> field.</p>
30
+ * <p> Stream groups manage how Amazon GameLift Streams allocates resources and handles concurrent streams, allowing you to effectively manage capacity and costs. Within a stream group, you specify an application to stream, streaming locations and their capacity, and the stream class you want to use when streaming applications to your end-users. A stream class defines the hardware configuration of the compute resources that Amazon GameLift Streams will use when streaming, such as the CPU, GPU, and memory. </p> <p> Stream capacity represents the number of concurrent streams that can be active at a time. You set stream capacity per location, per stream group. The following capacity settings are available: </p> <ul> <li> <p> <b>Always-on capacity</b>: This setting, if non-zero, indicates minimum streaming capacity which is allocated to you and is never released back to the service. You pay for this base level of capacity at all times, whether used or idle. </p> </li> <li> <p> <b>Maximum capacity</b>: This indicates the maximum capacity that the service can allocate for you. Newly created streams may take a few minutes to start. Capacity is released back to the service when idle. You pay for capacity that is allocated to you until it is released. </p> </li> <li> <p> <b>Target-idle capacity</b>: This indicates idle capacity which the service pre-allocates and holds for you in anticipation of future activity. This helps to insulate your users from capacity-allocation delays. You pay for capacity which is held in this intentional idle state. </p> </li> </ul> <p>Values for capacity must be whole number multiples of the tenancy value of the stream group's stream class.</p> <p> To adjust the capacity of any <code>ACTIVE</code> stream group, call <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_UpdateStreamGroup.html">UpdateStreamGroup</a>. </p> <p> If the <code>CreateStreamGroup</code> request is successful, Amazon GameLift Streams assigns a unique ID to the stream group resource and sets the status to <code>ACTIVATING</code>. It can take a few minutes for Amazon GameLift Streams to finish creating the stream group while it searches for unallocated compute resources and provisions them. When complete, the stream group status will be <code>ACTIVE</code> and you can start stream sessions by using <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_StartStreamSession.html">StartStreamSession</a>. To check the stream group's status, call <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_GetStreamGroup.html">GetStreamGroup</a>. </p> <p>Stream groups should be recreated every 3-4 weeks to pick up important service updates and fixes. Stream groups that are older than 180 days can no longer be updated with new application associations. Stream groups expire when they are 365 days old, at which point they can no longer stream sessions. The exact expiration date is indicated by the date value in the <code>ExpiresAt</code> field.</p>
31
31
  * @example
32
32
  * Use a bare-bones client and the command you need to make an API call.
33
33
  * ```javascript
@@ -38,13 +38,15 @@ declare const CreateStreamGroupCommand_base: {
38
38
  * const client = new GameLiftStreamsClient(config);
39
39
  * const input = { // CreateStreamGroupInput
40
40
  * Description: "STRING_VALUE", // required
41
- * StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022", // required
41
+ * StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022" || "gen6n_small" || "gen6n_medium" || "gen6n_high" || "gen6n_ultra" || "gen6n_ultra_win2022" || "gen6n_pro" || "gen6n_pro_win2022", // required
42
42
  * DefaultApplicationIdentifier: "STRING_VALUE",
43
43
  * LocationConfigurations: [ // LocationConfigurations
44
44
  * { // LocationConfiguration
45
45
  * LocationName: "STRING_VALUE", // required
46
46
  * AlwaysOnCapacity: Number("int"),
47
47
  * OnDemandCapacity: Number("int"),
48
+ * TargetIdleCapacity: Number("int"),
49
+ * MaximumCapacity: Number("int"),
48
50
  * },
49
51
  * ],
50
52
  * Tags: { // Tags
@@ -67,12 +69,14 @@ declare const CreateStreamGroupCommand_base: {
67
69
  * // Status: "ACTIVATING" || "ACTIVE" || "ERROR" || "REMOVING",
68
70
  * // AlwaysOnCapacity: Number("int"),
69
71
  * // OnDemandCapacity: Number("int"),
72
+ * // TargetIdleCapacity: Number("int"),
73
+ * // MaximumCapacity: Number("int"),
70
74
  * // RequestedCapacity: Number("int"),
71
75
  * // AllocatedCapacity: Number("int"),
72
76
  * // IdleCapacity: Number("int"),
73
77
  * // },
74
78
  * // ],
75
- * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022",
79
+ * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022" || "gen6n_small" || "gen6n_medium" || "gen6n_high" || "gen6n_ultra" || "gen6n_ultra_win2022" || "gen6n_pro" || "gen6n_pro_win2022",
76
80
  * // Id: "STRING_VALUE",
77
81
  * // Status: "ACTIVATING" || "UPDATING_LOCATIONS" || "ACTIVE" || "ACTIVE_WITH_ERRORS" || "ERROR" || "DELETING" || "EXPIRED",
78
82
  * // StatusReason: "internalError" || "noAvailableInstances",
@@ -54,12 +54,14 @@ declare const GetStreamGroupCommand_base: {
54
54
  * // Status: "ACTIVATING" || "ACTIVE" || "ERROR" || "REMOVING",
55
55
  * // AlwaysOnCapacity: Number("int"),
56
56
  * // OnDemandCapacity: Number("int"),
57
+ * // TargetIdleCapacity: Number("int"),
58
+ * // MaximumCapacity: Number("int"),
57
59
  * // RequestedCapacity: Number("int"),
58
60
  * // AllocatedCapacity: Number("int"),
59
61
  * // IdleCapacity: Number("int"),
60
62
  * // },
61
63
  * // ],
62
- * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022",
64
+ * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022" || "gen6n_small" || "gen6n_medium" || "gen6n_high" || "gen6n_ultra" || "gen6n_ultra_win2022" || "gen6n_pro" || "gen6n_pro_win2022",
63
65
  * // Id: "STRING_VALUE",
64
66
  * // Status: "ACTIVATING" || "UPDATING_LOCATIONS" || "ACTIVE" || "ACTIVE_WITH_ERRORS" || "ERROR" || "DELETING" || "EXPIRED",
65
67
  * // StatusReason: "internalError" || "noAvailableInstances",
@@ -61,6 +61,9 @@ declare const GetStreamSessionCommand_base: {
61
61
  * // AdditionalEnvironmentVariables: { // EnvironmentVariables
62
62
  * // "<keys>": "STRING_VALUE",
63
63
  * // },
64
+ * // PerformanceStatsConfiguration: { // PerformanceStatsConfiguration
65
+ * // SharedWithClient: true || false,
66
+ * // },
64
67
  * // LogFileLocationUri: "STRING_VALUE",
65
68
  * // WebSdkProtocolUrl: "STRING_VALUE",
66
69
  * // LastUpdatedAt: new Date("TIMESTAMP"),
@@ -52,7 +52,7 @@ declare const ListStreamGroupsCommand_base: {
52
52
  * // Id: "STRING_VALUE",
53
53
  * // Arn: "STRING_VALUE",
54
54
  * // },
55
- * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022",
55
+ * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022" || "gen6n_small" || "gen6n_medium" || "gen6n_high" || "gen6n_ultra" || "gen6n_ultra_win2022" || "gen6n_pro" || "gen6n_pro_win2022",
56
56
  * // Status: "ACTIVATING" || "UPDATING_LOCATIONS" || "ACTIVE" || "ACTIVE_WITH_ERRORS" || "ERROR" || "DELETING" || "EXPIRED",
57
57
  * // CreatedAt: new Date("TIMESTAMP"),
58
58
  * // LastUpdatedAt: new Date("TIMESTAMP"),
@@ -50,6 +50,7 @@ declare const ListStreamSessionsByAccountCommand_base: {
50
50
  * // Arn: "STRING_VALUE",
51
51
  * // UserId: "STRING_VALUE",
52
52
  * // Status: "ACTIVATING" || "ACTIVE" || "CONNECTED" || "PENDING_CLIENT_RECONNECTION" || "RECONNECTING" || "TERMINATING" || "TERMINATED" || "ERROR",
53
+ * // StatusReason: "internalError" || "invalidSignalRequest" || "placementTimeout" || "applicationLogS3DestinationError" || "applicationExit" || "connectionTimeout" || "reconnectionTimeout" || "maxSessionLengthTimeout" || "idleTimeout" || "apiTerminated",
53
54
  * // Protocol: "WebRTC",
54
55
  * // LastUpdatedAt: new Date("TIMESTAMP"),
55
56
  * // CreatedAt: new Date("TIMESTAMP"),
@@ -51,6 +51,7 @@ declare const ListStreamSessionsCommand_base: {
51
51
  * // Arn: "STRING_VALUE",
52
52
  * // UserId: "STRING_VALUE",
53
53
  * // Status: "ACTIVATING" || "ACTIVE" || "CONNECTED" || "PENDING_CLIENT_RECONNECTION" || "RECONNECTING" || "TERMINATING" || "TERMINATED" || "ERROR",
54
+ * // StatusReason: "internalError" || "invalidSignalRequest" || "placementTimeout" || "applicationLogS3DestinationError" || "applicationExit" || "connectionTimeout" || "reconnectionTimeout" || "maxSessionLengthTimeout" || "idleTimeout" || "apiTerminated",
54
55
  * // Protocol: "WebRTC",
55
56
  * // LastUpdatedAt: new Date("TIMESTAMP"),
56
57
  * // CreatedAt: new Date("TIMESTAMP"),
@@ -55,6 +55,9 @@ declare const StartStreamSessionCommand_base: {
55
55
  * AdditionalEnvironmentVariables: { // EnvironmentVariables
56
56
  * "<keys>": "STRING_VALUE",
57
57
  * },
58
+ * PerformanceStatsConfiguration: { // PerformanceStatsConfiguration
59
+ * SharedWithClient: true || false,
60
+ * },
58
61
  * };
59
62
  * const command = new StartStreamSessionCommand(input);
60
63
  * const response = await client.send(command);
@@ -77,6 +80,9 @@ declare const StartStreamSessionCommand_base: {
77
80
  * // AdditionalEnvironmentVariables: { // EnvironmentVariables
78
81
  * // "<keys>": "STRING_VALUE",
79
82
  * // },
83
+ * // PerformanceStatsConfiguration: { // PerformanceStatsConfiguration
84
+ * // SharedWithClient: true || false,
85
+ * // },
80
86
  * // LogFileLocationUri: "STRING_VALUE",
81
87
  * // WebSdkProtocolUrl: "STRING_VALUE",
82
88
  * // LastUpdatedAt: new Date("TIMESTAMP"),
@@ -27,7 +27,7 @@ declare const UpdateStreamGroupCommand_base: {
27
27
  getEndpointParameterInstructions(): import("@smithy/middleware-endpoint").EndpointParameterInstructions;
28
28
  };
29
29
  /**
30
- * <p> Updates the configuration settings for an Amazon GameLift Streams stream group resource. To update a stream group, it must be in <code>ACTIVE</code> status. You can change the description, the set of locations, and the requested capacity of a stream group per location. If you want to change the stream class, create a new stream group. </p> <p> Stream capacity represents the number of concurrent streams that can be active at a time. You set stream capacity per location, per stream group. There are two types of capacity, always-on and on-demand: </p> <ul> <li> <p> <b>Always-on</b>: The streaming capacity that is allocated and ready to handle stream requests without delay. You pay for this capacity whether it's in use or not. Best for quickest time from streaming request to streaming session. Default is 1 (2 for high stream classes) when creating a stream group or adding a location. </p> </li> <li> <p> <b>On-demand</b>: The streaming capacity that Amazon GameLift Streams can allocate in response to stream requests, and then de-allocate when the session has terminated. This offers a cost control measure at the expense of a greater startup time (typically under 5 minutes). Default is 0 when creating a stream group or adding a location. </p> </li> </ul> <p>Values for capacity must be whole number multiples of the tenancy value of the stream group's stream class.</p> <p>To update a stream group, specify the stream group's Amazon Resource Name (ARN) and provide the new values. If the request is successful, Amazon GameLift Streams returns the complete updated metadata for the stream group. Expired stream groups cannot be updated.</p>
30
+ * <p> Updates the configuration settings for an Amazon GameLift Streams stream group resource. To update a stream group, it must be in <code>ACTIVE</code> status. You can change the description, the set of locations, and the requested capacity of a stream group per location. If you want to change the stream class, create a new stream group. </p> <p> Stream capacity represents the number of concurrent streams that can be active at a time. You set stream capacity per location, per stream group. The following capacity settings are available: </p> <ul> <li> <p> <b>Always-on capacity</b>: This setting, if non-zero, indicates minimum streaming capacity which is allocated to you and is never released back to the service. You pay for this base level of capacity at all times, whether used or idle. </p> </li> <li> <p> <b>Maximum capacity</b>: This indicates the maximum capacity that the service can allocate for you. Newly created streams may take a few minutes to start. Capacity is released back to the service when idle. You pay for capacity that is allocated to you until it is released. </p> </li> <li> <p> <b>Target-idle capacity</b>: This indicates idle capacity which the service pre-allocates and holds for you in anticipation of future activity. This helps to insulate your users from capacity-allocation delays. You pay for capacity which is held in this intentional idle state. </p> </li> </ul> <p>Values for capacity must be whole number multiples of the tenancy value of the stream group's stream class.</p> <p>To update a stream group, specify the stream group's Amazon Resource Name (ARN) and provide the new values. If the request is successful, Amazon GameLift Streams returns the complete updated metadata for the stream group. Expired stream groups cannot be updated.</p>
31
31
  * @example
32
32
  * Use a bare-bones client and the command you need to make an API call.
33
33
  * ```javascript
@@ -43,6 +43,8 @@ declare const UpdateStreamGroupCommand_base: {
43
43
  * LocationName: "STRING_VALUE", // required
44
44
  * AlwaysOnCapacity: Number("int"),
45
45
  * OnDemandCapacity: Number("int"),
46
+ * TargetIdleCapacity: Number("int"),
47
+ * MaximumCapacity: Number("int"),
46
48
  * },
47
49
  * ],
48
50
  * Description: "STRING_VALUE",
@@ -63,12 +65,14 @@ declare const UpdateStreamGroupCommand_base: {
63
65
  * // Status: "ACTIVATING" || "ACTIVE" || "ERROR" || "REMOVING",
64
66
  * // AlwaysOnCapacity: Number("int"),
65
67
  * // OnDemandCapacity: Number("int"),
68
+ * // TargetIdleCapacity: Number("int"),
69
+ * // MaximumCapacity: Number("int"),
66
70
  * // RequestedCapacity: Number("int"),
67
71
  * // AllocatedCapacity: Number("int"),
68
72
  * // IdleCapacity: Number("int"),
69
73
  * // },
70
74
  * // ],
71
- * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022",
75
+ * // StreamClass: "gen4n_high" || "gen4n_ultra" || "gen4n_win2022" || "gen5n_high" || "gen5n_ultra" || "gen5n_win2022" || "gen6n_small" || "gen6n_medium" || "gen6n_high" || "gen6n_ultra" || "gen6n_ultra_win2022" || "gen6n_pro" || "gen6n_pro_win2022",
72
76
  * // Id: "STRING_VALUE",
73
77
  * // Status: "ACTIVATING" || "UPDATING_LOCATIONS" || "ACTIVE" || "ACTIVE_WITH_ERRORS" || "ERROR" || "DELETING" || "EXPIRED",
74
78
  * // StatusReason: "internalError" || "noAvailableInstances",
@@ -14,5 +14,5 @@ export * from "./pagination";
14
14
  export * from "./waiters";
15
15
  export * from "./models/enums";
16
16
  export * from "./models/errors";
17
- export type * from "./models/models_0";
17
+ export * from "./models/models_0";
18
18
  export { GameLiftStreamsServiceException } from "./models/GameLiftStreamsServiceException";
@@ -75,6 +75,13 @@ export declare const StreamClass: {
75
75
  readonly gen5n_high: "gen5n_high";
76
76
  readonly gen5n_ultra: "gen5n_ultra";
77
77
  readonly gen5n_win2022: "gen5n_win2022";
78
+ readonly gen6n_high: "gen6n_high";
79
+ readonly gen6n_medium: "gen6n_medium";
80
+ readonly gen6n_pro: "gen6n_pro";
81
+ readonly gen6n_pro_win2022: "gen6n_pro_win2022";
82
+ readonly gen6n_small: "gen6n_small";
83
+ readonly gen6n_ultra: "gen6n_ultra";
84
+ readonly gen6n_ultra_win2022: "gen6n_ultra_win2022";
78
85
  };
79
86
  /**
80
87
  * @public
@@ -10,15 +10,27 @@ export interface LocationConfiguration {
10
10
  */
11
11
  LocationName: string | undefined;
12
12
  /**
13
- * <p>The streaming capacity that is allocated and ready to handle stream requests without delay. You pay for this capacity whether it's in use or not. Best for quickest time from streaming request to streaming session. Default is 1 (2 for high stream classes) when creating a stream group or adding a location.</p>
13
+ * <p>This setting, if non-zero, indicates minimum streaming capacity which is allocated to you and is never released back to the service. You pay for this base level of capacity at all times, whether used or idle.</p>
14
14
  * @public
15
15
  */
16
16
  AlwaysOnCapacity?: number | undefined;
17
17
  /**
18
- * <p>The streaming capacity that Amazon GameLift Streams can allocate in response to stream requests, and then de-allocate when the session has terminated. This offers a cost control measure at the expense of a greater startup time (typically under 5 minutes). Default is 0 when creating a stream group or adding a location.</p>
18
+ * <p>This field is deprecated. Use <code>MaximumCapacity</code> instead. This parameter cannot be used with <code>MaximumCapacity</code> or <code>TargetIdleCapacity</code> in the same location configuration.</p> <p>The streaming capacity that Amazon GameLift Streams can allocate in response to stream requests, and then de-allocate when the session has terminated. This offers a cost control measure at the expense of a greater startup time (typically under 5 minutes). Default is 0 when creating a stream group or adding a location.</p>
19
+ *
20
+ * @deprecated This input field is deprecated in favor of explicit MaximumCapacity values.
19
21
  * @public
20
22
  */
21
23
  OnDemandCapacity?: number | undefined;
24
+ /**
25
+ * <p>This indicates idle capacity which the service pre-allocates and holds for you in anticipation of future activity. This helps to insulate your users from capacity-allocation delays. You pay for capacity which is held in this intentional idle state.</p>
26
+ * @public
27
+ */
28
+ TargetIdleCapacity?: number | undefined;
29
+ /**
30
+ * <p>This indicates the maximum capacity that the service can allocate for you. Newly created streams may take a few minutes to start. Capacity is released back to the service when idle. You pay for capacity that is allocated to you until it is released.</p>
31
+ * @public
32
+ */
33
+ MaximumCapacity?: number | undefined;
22
34
  }
23
35
  /**
24
36
  * @public
@@ -51,7 +63,7 @@ export interface LocationState {
51
63
  */
52
64
  Status?: StreamGroupLocationStatus | undefined;
53
65
  /**
54
- * <p>The streaming capacity that is allocated and ready to handle stream requests without delay. You pay for this capacity whether it's in use or not. Best for quickest time from streaming request to streaming session. Default is 1 (2 for high stream classes) when creating a stream group or adding a location.</p>
66
+ * <p>This setting, if non-zero, indicates minimum streaming capacity which is allocated to you and is never released back to the service. You pay for this base level of capacity at all times, whether used or idle.</p>
55
67
  * @public
56
68
  */
57
69
  AlwaysOnCapacity?: number | undefined;
@@ -60,13 +72,23 @@ export interface LocationState {
60
72
  * @public
61
73
  */
62
74
  OnDemandCapacity?: number | undefined;
75
+ /**
76
+ * <p>This indicates idle capacity which the service pre-allocates and holds for you in anticipation of future activity. This helps to insulate your users from capacity-allocation delays. You pay for capacity which is held in this intentional idle state.</p>
77
+ * @public
78
+ */
79
+ TargetIdleCapacity?: number | undefined;
80
+ /**
81
+ * <p>This indicates the maximum capacity that the service can allocate for you. Newly created streams may take a few minutes to start. Capacity is released back to the service when idle. You pay for capacity that is allocated to you until it is released.</p>
82
+ * @public
83
+ */
84
+ MaximumCapacity?: number | undefined;
63
85
  /**
64
86
  * <p>This value is the always-on capacity that you most recently requested for a stream group. You request capacity separately for each location in a stream group. In response to an increase in requested capacity, Amazon GameLift Streams attempts to provision compute resources to make the stream group's allocated capacity meet requested capacity. When always-on capacity is decreased, it can take a few minutes to deprovision allocated capacity to match the requested capacity.</p>
65
87
  * @public
66
88
  */
67
89
  RequestedCapacity?: number | undefined;
68
90
  /**
69
- * <p>This value is the stream capacity that Amazon GameLift Streams has provisioned in a stream group that can respond immediately to stream requests. It includes resources that are currently streaming and resources that are idle and ready to respond to stream requests. You pay for this capacity whether it's in use or not. After making changes to capacity, it can take a few minutes for the allocated capacity count to reflect the change while compute resources are allocated or deallocated. Similarly, when allocated on-demand capacity is no longer needed, it can take a few minutes for Amazon GameLift Streams to spin down the allocated capacity.</p>
91
+ * <p>This value is the stream capacity that Amazon GameLift Streams has provisioned in a stream group that can respond immediately to stream requests. It includes resources that are currently streaming and resources that are idle and ready to respond to stream requests. When target-idle capacity is configured, the idle resources include the capacity buffer maintained beyond ongoing sessions. You pay for this capacity whether it's in use or not. After making changes to capacity, it can take a few minutes for the allocated capacity count to reflect the change while compute resources are allocated or deallocated. Similarly, when allocated on-demand capacity is no longer needed, it can take a few minutes for Amazon GameLift Streams to spin down the allocated capacity.</p>
70
92
  * @public
71
93
  */
72
94
  AllocatedCapacity?: number | undefined;
@@ -549,7 +571,7 @@ export interface CreateStreamGroupInput {
549
571
  */
550
572
  Description: string | undefined;
551
573
  /**
552
- * <p>The target stream quality for sessions that are hosted in this stream group. Set a stream class that is appropriate to the type of content that you're streaming. Stream class determines the type of computing resources Amazon GameLift Streams uses and impacts the cost of streaming. The following options are available: </p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
574
+ * <p>The target stream quality for sessions that are hosted in this stream group. Set a stream class that is appropriate to the type of content that you're streaming. Stream class determines the type of computing resources Amazon GameLift Streams uses and impacts the cost of streaming. The following options are available: </p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen6n_pro_win2022</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_pro</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra_win2022</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_medium</code> (NVIDIA, medium)</b> Supports applications with moderate 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 2 vCPUs, 8 GB RAM, 6 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 4 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_small</code> (NVIDIA, small)</b> Supports applications with lightweight 3D scene complexity and low CPU usage. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 1 vCPUs, 4 GB RAM, 2 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 12 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
553
575
  * @public
554
576
  */
555
577
  StreamClass: StreamClass | undefined;
@@ -615,7 +637,7 @@ export interface CreateStreamGroupOutput {
615
637
  */
616
638
  LocationStates?: LocationState[] | undefined;
617
639
  /**
618
- * <p>The target stream quality for the stream group.</p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
640
+ * <p>The target stream quality for the stream group.</p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen6n_pro_win2022</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_pro</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra_win2022</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_medium</code> (NVIDIA, medium)</b> Supports applications with moderate 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 2 vCPUs, 8 GB RAM, 6 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 4 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_small</code> (NVIDIA, small)</b> Supports applications with lightweight 3D scene complexity and low CPU usage. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 1 vCPUs, 4 GB RAM, 2 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 12 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
619
641
  * @public
620
642
  */
621
643
  StreamClass?: StreamClass | undefined;
@@ -630,7 +652,7 @@ export interface CreateStreamGroupOutput {
630
652
  */
631
653
  Status?: StreamGroupStatus | undefined;
632
654
  /**
633
- * <p> A short description of the reason that the stream group is in <code>ERROR</code> status. The possible reasons can be one of the following: </p> <ul> <li> <p> <code>internalError</code>: The request can't process right now because of an issue with the server. Try again later.</p> </li> <li> <p> <code>noAvailableInstances</code>: Amazon GameLift Streams does not currently have enough available on-demand capacity to fulfill your request. Wait a few minutes and retry the request as capacity can shift frequently. You can also try to make the request using a different stream class or in another region.</p> </li> </ul>
655
+ * <p> A short description of the reason that the stream group is in <code>ERROR</code> status. The possible reasons can be one of the following: </p> <ul> <li> <p> <code>internalError</code>: The request can't process right now because of an issue with the server. Try again later.</p> </li> <li> <p> <code>noAvailableInstances</code>: Amazon GameLift Streams does not currently have enough available capacity to fulfill your request. Wait a few minutes and retry the request as capacity can shift frequently. You can also try to make the request using a different stream class or in another region.</p> </li> </ul>
634
656
  * @public
635
657
  */
636
658
  StatusReason?: StreamGroupStatusReason | undefined;
@@ -791,6 +813,17 @@ export interface GetStreamSessionInput {
791
813
  */
792
814
  StreamSessionIdentifier: string | undefined;
793
815
  }
816
+ /**
817
+ * <p>Configuration settings for sharing the stream session's performance stats with the client</p>
818
+ * @public
819
+ */
820
+ export interface PerformanceStatsConfiguration {
821
+ /**
822
+ * <p>Performance stats for the session are streamed to the client when set to <code>true</code>. Defaults to <code>false</code>.</p>
823
+ * @public
824
+ */
825
+ SharedWithClient?: boolean | undefined;
826
+ }
794
827
  /**
795
828
  * @public
796
829
  */
@@ -865,6 +898,11 @@ export interface GetStreamSessionOutput {
865
898
  * @public
866
899
  */
867
900
  AdditionalEnvironmentVariables?: Record<string, string> | undefined;
901
+ /**
902
+ * <p>The performance stats configuration for the stream session</p>
903
+ * @public
904
+ */
905
+ PerformanceStatsConfiguration?: PerformanceStatsConfiguration | undefined;
868
906
  /**
869
907
  * <p>Access location for log files that your content generates during a stream session. These log files are uploaded to cloud storage location at the end of a stream session. The Amazon GameLift Streams application resource defines which log files to upload.</p>
870
908
  * @public
@@ -946,6 +984,11 @@ export interface StreamSessionSummary {
946
984
  * @public
947
985
  */
948
986
  Status?: StreamSessionStatus | undefined;
987
+ /**
988
+ * <p>A short description of the reason the stream session is in <code>ERROR</code> status or <code>TERMINATED</code> status.</p> <p> <code>ERROR</code> status reasons:</p> <ul> <li> <p> <code>applicationLogS3DestinationError</code>: Could not write the application log to the Amazon S3 bucket that is configured for the streaming application. Make sure the bucket still exists.</p> </li> <li> <p> <code>internalError</code>: An internal service error occurred. Start a new stream session to continue streaming.</p> </li> <li> <p> <code>invalidSignalRequest</code>: The WebRTC signal request that was sent is not valid. When starting or reconnecting to a stream session, use <code>generateSignalRequest</code> in the Amazon GameLift Streams Web SDK to generate a new signal request.</p> </li> <li> <p> <code>placementTimeout</code>: Amazon GameLift Streams could not find available stream capacity to start a stream session. Increase the stream capacity in the stream group or wait until capacity becomes available.</p> </li> </ul> <p> <code>TERMINATED</code> status reasons:</p> <ul> <li> <p> <code>apiTerminated</code>: The stream session was terminated by an API call to <a href="https://docs.aws.amazon.com/gameliftstreams/latest/apireference/API_TerminateStreamSession.html">TerminateStreamSession</a>.</p> </li> <li> <p> <code>applicationExit</code>: The streaming application exited or crashed. The stream session was terminated because the application is no longer running.</p> </li> <li> <p> <code>connectionTimeout</code>: The stream session was terminated because the client failed to connect within the connection timeout period specified by <code>ConnectionTimeoutSeconds</code>.</p> </li> <li> <p> <code>idleTimeout</code>: The stream session was terminated because it exceeded the idle timeout period of 60 minutes with no user input activity.</p> </li> <li> <p> <code>maxSessionLengthTimeout</code>: The stream session was terminated because it exceeded the maximum session length timeout period specified by <code>SessionLengthSeconds</code>.</p> </li> <li> <p> <code>reconnectionTimeout</code>: The stream session was terminated because the client failed to reconnect within the reconnection timeout period specified by <code>ConnectionTimeoutSeconds</code> after losing connection.</p> </li> </ul>
989
+ * @public
990
+ */
991
+ StatusReason?: StreamSessionStatusReason | undefined;
949
992
  /**
950
993
  * <p>The data transfer protocol in use with the stream session.</p>
951
994
  * @public
@@ -1131,6 +1174,11 @@ export interface StartStreamSessionInput {
1131
1174
  * @public
1132
1175
  */
1133
1176
  AdditionalEnvironmentVariables?: Record<string, string> | undefined;
1177
+ /**
1178
+ * <p>Configuration settings for sharing the stream session's performance stats with the client</p>
1179
+ * @public
1180
+ */
1181
+ PerformanceStatsConfiguration?: PerformanceStatsConfiguration | undefined;
1134
1182
  }
1135
1183
  /**
1136
1184
  * @public
@@ -1206,6 +1254,11 @@ export interface StartStreamSessionOutput {
1206
1254
  * @public
1207
1255
  */
1208
1256
  AdditionalEnvironmentVariables?: Record<string, string> | undefined;
1257
+ /**
1258
+ * <p>The performance stats configuration for the stream session</p>
1259
+ * @public
1260
+ */
1261
+ PerformanceStatsConfiguration?: PerformanceStatsConfiguration | undefined;
1209
1262
  /**
1210
1263
  * <p>Access location for log files that your content generates during a stream session. These log files are uploaded to cloud storage location at the end of a stream session. The Amazon GameLift Streams application resource defines which log files to upload.</p>
1211
1264
  * @public
@@ -1272,7 +1325,7 @@ export interface GetStreamGroupOutput {
1272
1325
  */
1273
1326
  LocationStates?: LocationState[] | undefined;
1274
1327
  /**
1275
- * <p>The target stream quality for the stream group.</p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
1328
+ * <p>The target stream quality for the stream group.</p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen6n_pro_win2022</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_pro</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra_win2022</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_medium</code> (NVIDIA, medium)</b> Supports applications with moderate 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 2 vCPUs, 8 GB RAM, 6 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 4 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_small</code> (NVIDIA, small)</b> Supports applications with lightweight 3D scene complexity and low CPU usage. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 1 vCPUs, 4 GB RAM, 2 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 12 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
1276
1329
  * @public
1277
1330
  */
1278
1331
  StreamClass?: StreamClass | undefined;
@@ -1287,7 +1340,7 @@ export interface GetStreamGroupOutput {
1287
1340
  */
1288
1341
  Status?: StreamGroupStatus | undefined;
1289
1342
  /**
1290
- * <p> A short description of the reason that the stream group is in <code>ERROR</code> status. The possible reasons can be one of the following: </p> <ul> <li> <p> <code>internalError</code>: The request can't process right now because of an issue with the server. Try again later.</p> </li> <li> <p> <code>noAvailableInstances</code>: Amazon GameLift Streams does not currently have enough available on-demand capacity to fulfill your request. Wait a few minutes and retry the request as capacity can shift frequently. You can also try to make the request using a different stream class or in another region.</p> </li> </ul>
1343
+ * <p> A short description of the reason that the stream group is in <code>ERROR</code> status. The possible reasons can be one of the following: </p> <ul> <li> <p> <code>internalError</code>: The request can't process right now because of an issue with the server. Try again later.</p> </li> <li> <p> <code>noAvailableInstances</code>: Amazon GameLift Streams does not currently have enough available capacity to fulfill your request. Wait a few minutes and retry the request as capacity can shift frequently. You can also try to make the request using a different stream class or in another region.</p> </li> </ul>
1291
1344
  * @public
1292
1345
  */
1293
1346
  StatusReason?: StreamGroupStatusReason | undefined;
@@ -1353,7 +1406,7 @@ export interface StreamGroupSummary {
1353
1406
  */
1354
1407
  DefaultApplication?: DefaultApplication | undefined;
1355
1408
  /**
1356
- * <p>The target stream quality for the stream group. </p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
1409
+ * <p>The target stream quality for the stream group. </p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen6n_pro_win2022</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_pro</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra_win2022</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_medium</code> (NVIDIA, medium)</b> Supports applications with moderate 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 2 vCPUs, 8 GB RAM, 6 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 4 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_small</code> (NVIDIA, small)</b> Supports applications with lightweight 3D scene complexity and low CPU usage. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 1 vCPUs, 4 GB RAM, 2 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 12 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
1357
1410
  * @public
1358
1411
  */
1359
1412
  StreamClass?: StreamClass | undefined;
@@ -1443,7 +1496,7 @@ export interface UpdateStreamGroupOutput {
1443
1496
  */
1444
1497
  LocationStates?: LocationState[] | undefined;
1445
1498
  /**
1446
- * <p>The target stream quality for the stream group.</p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.4, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
1499
+ * <p>The target stream quality for the stream group.</p> <p>A stream class can be one of the following:</p> <ul> <li> <p> <b> <code>gen6n_pro_win2022</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_pro</code> (NVIDIA, pro)</b> Supports applications with extremely high 3D scene complexity which require maximum resources. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 16 vCPUs, 64 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra_win2022</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_medium</code> (NVIDIA, medium)</b> Supports applications with moderate 3D scene complexity. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 2 vCPUs, 8 GB RAM, 6 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 4 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen6n_small</code> (NVIDIA, small)</b> Supports applications with lightweight 3D scene complexity and low CPU usage. Uses NVIDIA L4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 1 vCPUs, 4 GB RAM, 2 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 12 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 12 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen5n_ultra</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Uses dedicated NVIDIA A10G Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 24 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_win2022</code> (NVIDIA, ultra)</b> Supports applications with extremely high 3D scene complexity. Runs applications on Microsoft Windows Server 2022 Base and supports DirectX 12. Compatible with Unreal Engine versions up through 5.6, 32 and 64-bit applications, and anti-cheat technology. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_high</code> (NVIDIA, high)</b> Supports applications with moderate to high 3D scene complexity. Uses NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 4 vCPUs, 16 GB RAM, 8 GB VRAM</p> </li> <li> <p>Tenancy: Supports up to 2 concurrent stream sessions</p> </li> </ul> </li> <li> <p> <b> <code>gen4n_ultra</code> (NVIDIA, ultra)</b> Supports applications with high 3D scene complexity. Uses dedicated NVIDIA T4 Tensor Core GPU.</p> <ul> <li> <p>Reference resolution: 1080p</p> </li> <li> <p>Reference frame rate: 60 fps</p> </li> <li> <p>Workload specifications: 8 vCPUs, 32 GB RAM, 16 GB VRAM</p> </li> <li> <p>Tenancy: Supports 1 concurrent stream session</p> </li> </ul> </li> </ul>
1447
1500
  * @public
1448
1501
  */
1449
1502
  StreamClass?: StreamClass | undefined;
@@ -1458,7 +1511,7 @@ export interface UpdateStreamGroupOutput {
1458
1511
  */
1459
1512
  Status?: StreamGroupStatus | undefined;
1460
1513
  /**
1461
- * <p> A short description of the reason that the stream group is in <code>ERROR</code> status. The possible reasons can be one of the following: </p> <ul> <li> <p> <code>internalError</code>: The request can't process right now because of an issue with the server. Try again later.</p> </li> <li> <p> <code>noAvailableInstances</code>: Amazon GameLift Streams does not currently have enough available on-demand capacity to fulfill your request. Wait a few minutes and retry the request as capacity can shift frequently. You can also try to make the request using a different stream class or in another region.</p> </li> </ul>
1514
+ * <p> A short description of the reason that the stream group is in <code>ERROR</code> status. The possible reasons can be one of the following: </p> <ul> <li> <p> <code>internalError</code>: The request can't process right now because of an issue with the server. Try again later.</p> </li> <li> <p> <code>noAvailableInstances</code>: Amazon GameLift Streams does not currently have enough available capacity to fulfill your request. Wait a few minutes and retry the request as capacity can shift frequently. You can also try to make the request using a different stream class or in another region.</p> </li> </ul>
1462
1515
  * @public
1463
1516
  */
1464
1517
  StatusReason?: StreamGroupStatusReason | undefined;
@@ -39,6 +39,7 @@ export declare var ListTagsForResourceRequest$: StaticStructureSchema;
39
39
  export declare var ListTagsForResourceResponse$: StaticStructureSchema;
40
40
  export declare var LocationConfiguration$: StaticStructureSchema;
41
41
  export declare var LocationState$: StaticStructureSchema;
42
+ export declare var PerformanceStatsConfiguration$: StaticStructureSchema;
42
43
  export declare var RemoveStreamGroupLocationsInput$: StaticStructureSchema;
43
44
  export declare var ReplicationStatus$: StaticStructureSchema;
44
45
  export declare var ResourceNotFoundException$: StaticErrorSchema;
@@ -41,6 +41,13 @@ export declare const StreamClass: {
41
41
  readonly gen5n_high: "gen5n_high";
42
42
  readonly gen5n_ultra: "gen5n_ultra";
43
43
  readonly gen5n_win2022: "gen5n_win2022";
44
+ readonly gen6n_high: "gen6n_high";
45
+ readonly gen6n_medium: "gen6n_medium";
46
+ readonly gen6n_pro: "gen6n_pro";
47
+ readonly gen6n_pro_win2022: "gen6n_pro_win2022";
48
+ readonly gen6n_small: "gen6n_small";
49
+ readonly gen6n_ultra: "gen6n_ultra";
50
+ readonly gen6n_ultra_win2022: "gen6n_ultra_win2022";
44
51
  };
45
52
  export type StreamClass = (typeof StreamClass)[keyof typeof StreamClass];
46
53
  export declare const StreamGroupStatus: {
@@ -16,6 +16,8 @@ export interface LocationConfiguration {
16
16
  LocationName: string | undefined;
17
17
  AlwaysOnCapacity?: number | undefined;
18
18
  OnDemandCapacity?: number | undefined;
19
+ TargetIdleCapacity?: number | undefined;
20
+ MaximumCapacity?: number | undefined;
19
21
  }
20
22
  export interface AddStreamGroupLocationsInput {
21
23
  Identifier: string | undefined;
@@ -26,6 +28,8 @@ export interface LocationState {
26
28
  Status?: StreamGroupLocationStatus | undefined;
27
29
  AlwaysOnCapacity?: number | undefined;
28
30
  OnDemandCapacity?: number | undefined;
31
+ TargetIdleCapacity?: number | undefined;
32
+ MaximumCapacity?: number | undefined;
29
33
  RequestedCapacity?: number | undefined;
30
34
  AllocatedCapacity?: number | undefined;
31
35
  IdleCapacity?: number | undefined;
@@ -198,6 +202,9 @@ export interface GetStreamSessionInput {
198
202
  Identifier: string | undefined;
199
203
  StreamSessionIdentifier: string | undefined;
200
204
  }
205
+ export interface PerformanceStatsConfiguration {
206
+ SharedWithClient?: boolean | undefined;
207
+ }
201
208
  export interface GetStreamSessionOutput {
202
209
  Arn?: string | undefined;
203
210
  Description?: string | undefined;
@@ -213,6 +220,7 @@ export interface GetStreamSessionOutput {
213
220
  SessionLengthSeconds?: number | undefined;
214
221
  AdditionalLaunchArgs?: string[] | undefined;
215
222
  AdditionalEnvironmentVariables?: Record<string, string> | undefined;
223
+ PerformanceStatsConfiguration?: PerformanceStatsConfiguration | undefined;
216
224
  LogFileLocationUri?: string | undefined;
217
225
  WebSdkProtocolUrl?: string | undefined;
218
226
  LastUpdatedAt?: Date | undefined;
@@ -231,6 +239,7 @@ export interface StreamSessionSummary {
231
239
  Arn?: string | undefined;
232
240
  UserId?: string | undefined;
233
241
  Status?: StreamSessionStatus | undefined;
242
+ StatusReason?: StreamSessionStatusReason | undefined;
234
243
  Protocol?: Protocol | undefined;
235
244
  LastUpdatedAt?: Date | undefined;
236
245
  CreatedAt?: Date | undefined;
@@ -275,6 +284,7 @@ export interface StartStreamSessionInput {
275
284
  SessionLengthSeconds?: number | undefined;
276
285
  AdditionalLaunchArgs?: string[] | undefined;
277
286
  AdditionalEnvironmentVariables?: Record<string, string> | undefined;
287
+ PerformanceStatsConfiguration?: PerformanceStatsConfiguration | undefined;
278
288
  }
279
289
  export interface StartStreamSessionOutput {
280
290
  Arn?: string | undefined;
@@ -291,6 +301,7 @@ export interface StartStreamSessionOutput {
291
301
  SessionLengthSeconds?: number | undefined;
292
302
  AdditionalLaunchArgs?: string[] | undefined;
293
303
  AdditionalEnvironmentVariables?: Record<string, string> | undefined;
304
+ PerformanceStatsConfiguration?: PerformanceStatsConfiguration | undefined;
294
305
  LogFileLocationUri?: string | undefined;
295
306
  WebSdkProtocolUrl?: string | undefined;
296
307
  LastUpdatedAt?: Date | undefined;
@@ -43,6 +43,7 @@ export declare var ListTagsForResourceRequest$: StaticStructureSchema;
43
43
  export declare var ListTagsForResourceResponse$: StaticStructureSchema;
44
44
  export declare var LocationConfiguration$: StaticStructureSchema;
45
45
  export declare var LocationState$: StaticStructureSchema;
46
+ export declare var PerformanceStatsConfiguration$: StaticStructureSchema;
46
47
  export declare var RemoveStreamGroupLocationsInput$: StaticStructureSchema;
47
48
  export declare var ReplicationStatus$: StaticStructureSchema;
48
49
  export declare var ResourceNotFoundException$: StaticErrorSchema;
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@aws-sdk/client-gameliftstreams",
3
3
  "description": "AWS SDK for JavaScript Gameliftstreams Client for Node.js, Browser and React Native",
4
- "version": "3.953.0",
4
+ "version": "3.955.0",
5
5
  "scripts": {
6
6
  "build": "concurrently 'yarn:build:types' 'yarn:build:es' && yarn build:cjs",
7
7
  "build:cjs": "node ../../scripts/compilation/inline client-gameliftstreams",
@@ -21,38 +21,38 @@
21
21
  "dependencies": {
22
22
  "@aws-crypto/sha256-browser": "5.2.0",
23
23
  "@aws-crypto/sha256-js": "5.2.0",
24
- "@aws-sdk/core": "3.953.0",
25
- "@aws-sdk/credential-provider-node": "3.953.0",
24
+ "@aws-sdk/core": "3.954.0",
25
+ "@aws-sdk/credential-provider-node": "3.955.0",
26
26
  "@aws-sdk/middleware-host-header": "3.953.0",
27
27
  "@aws-sdk/middleware-logger": "3.953.0",
28
28
  "@aws-sdk/middleware-recursion-detection": "3.953.0",
29
- "@aws-sdk/middleware-user-agent": "3.953.0",
29
+ "@aws-sdk/middleware-user-agent": "3.954.0",
30
30
  "@aws-sdk/region-config-resolver": "3.953.0",
31
31
  "@aws-sdk/types": "3.953.0",
32
32
  "@aws-sdk/util-endpoints": "3.953.0",
33
33
  "@aws-sdk/util-user-agent-browser": "3.953.0",
34
- "@aws-sdk/util-user-agent-node": "3.953.0",
34
+ "@aws-sdk/util-user-agent-node": "3.954.0",
35
35
  "@smithy/config-resolver": "^4.4.4",
36
36
  "@smithy/core": "^3.19.0",
37
37
  "@smithy/fetch-http-handler": "^5.3.7",
38
38
  "@smithy/hash-node": "^4.2.6",
39
39
  "@smithy/invalid-dependency": "^4.2.6",
40
40
  "@smithy/middleware-content-length": "^4.2.6",
41
- "@smithy/middleware-endpoint": "^4.3.15",
42
- "@smithy/middleware-retry": "^4.4.15",
41
+ "@smithy/middleware-endpoint": "^4.4.0",
42
+ "@smithy/middleware-retry": "^4.4.16",
43
43
  "@smithy/middleware-serde": "^4.2.7",
44
44
  "@smithy/middleware-stack": "^4.2.6",
45
45
  "@smithy/node-config-provider": "^4.3.6",
46
46
  "@smithy/node-http-handler": "^4.4.6",
47
47
  "@smithy/protocol-http": "^5.3.6",
48
- "@smithy/smithy-client": "^4.10.0",
48
+ "@smithy/smithy-client": "^4.10.1",
49
49
  "@smithy/types": "^4.10.0",
50
50
  "@smithy/url-parser": "^4.2.6",
51
51
  "@smithy/util-base64": "^4.3.0",
52
52
  "@smithy/util-body-length-browser": "^4.2.0",
53
53
  "@smithy/util-body-length-node": "^4.2.1",
54
- "@smithy/util-defaults-mode-browser": "^4.3.14",
55
- "@smithy/util-defaults-mode-node": "^4.2.17",
54
+ "@smithy/util-defaults-mode-browser": "^4.3.15",
55
+ "@smithy/util-defaults-mode-node": "^4.2.18",
56
56
  "@smithy/util-endpoints": "^3.2.6",
57
57
  "@smithy/util-middleware": "^4.2.6",
58
58
  "@smithy/util-retry": "^4.2.6",