cribl-control-plane 0.0.38a1__py3-none-any.whl → 0.0.39__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of cribl-control-plane might be problematic. Click here for more details.

Files changed (65) hide show
  1. cribl_control_plane/_version.py +3 -3
  2. cribl_control_plane/commits.py +52 -42
  3. cribl_control_plane/commits_files.py +12 -12
  4. cribl_control_plane/destinations.py +3 -3
  5. cribl_control_plane/groups_sdk.py +8 -8
  6. cribl_control_plane/models/__init__.py +46 -368
  7. cribl_control_plane/models/configgroup.py +5 -3
  8. cribl_control_plane/models/createversioncommitop.py +26 -1
  9. cribl_control_plane/models/createversionrevertop.py +4 -2
  10. cribl_control_plane/models/createversionundoop.py +4 -2
  11. cribl_control_plane/models/getversioncountop.py +3 -2
  12. cribl_control_plane/models/getversiondiffop.py +3 -2
  13. cribl_control_plane/models/getversionfilesop.py +3 -2
  14. cribl_control_plane/models/getversionop.py +4 -2
  15. cribl_control_plane/models/getversionshowop.py +3 -2
  16. cribl_control_plane/models/getversionstatusop.py +4 -2
  17. cribl_control_plane/models/inputcribllakehttp.py +5 -4
  18. cribl_control_plane/models/inputcrowdstrike.py +2 -2
  19. cribl_control_plane/models/inputs3.py +2 -2
  20. cribl_control_plane/models/inputs3inventory.py +2 -2
  21. cribl_control_plane/models/inputsecuritylake.py +2 -2
  22. cribl_control_plane/models/outputcrowdstrikenextgensiem.py +1 -1
  23. cribl_control_plane/models/outputgooglepubsub.py +7 -28
  24. cribl_control_plane/models/outputsentineloneaisiem.py +3 -3
  25. cribl_control_plane/models/packinfo.py +5 -5
  26. cribl_control_plane/models/packinstallinfo.py +5 -5
  27. cribl_control_plane/models/packrequestbody_union.py +140 -0
  28. cribl_control_plane/models/packupgraderequest.py +26 -0
  29. cribl_control_plane/models/updatepacksbyidop.py +9 -28
  30. cribl_control_plane/packs.py +62 -116
  31. cribl_control_plane/{destinations_samples.py → samples.py} +1 -1
  32. cribl_control_plane/sdk.py +0 -3
  33. cribl_control_plane/statuses.py +6 -6
  34. {cribl_control_plane-0.0.38a1.dist-info → cribl_control_plane-0.0.39.dist-info}/METADATA +6 -22
  35. {cribl_control_plane-0.0.38a1.dist-info → cribl_control_plane-0.0.39.dist-info}/RECORD +36 -63
  36. cribl_control_plane/cribl.py +0 -513
  37. cribl_control_plane/models/authconfig.py +0 -43
  38. cribl_control_plane/models/commonservicelimitconfigs.py +0 -14
  39. cribl_control_plane/models/edgeheartbeatmetricsmode.py +0 -11
  40. cribl_control_plane/models/getsystemsettingsauthop.py +0 -24
  41. cribl_control_plane/models/getsystemsettingsconfop.py +0 -24
  42. cribl_control_plane/models/getsystemsettingsgitsettingsop.py +0 -24
  43. cribl_control_plane/models/gitopstype.py +0 -10
  44. cribl_control_plane/models/gitsettings.py +0 -70
  45. cribl_control_plane/models/jobsettings.py +0 -83
  46. cribl_control_plane/models/limits.py +0 -127
  47. cribl_control_plane/models/packrequestbody.py +0 -75
  48. cribl_control_plane/models/rediscachelimits.py +0 -38
  49. cribl_control_plane/models/redisconnectionlimits.py +0 -20
  50. cribl_control_plane/models/redislimits.py +0 -14
  51. cribl_control_plane/models/searchsettings.py +0 -71
  52. cribl_control_plane/models/serviceslimits.py +0 -23
  53. cribl_control_plane/models/systemsettings.py +0 -358
  54. cribl_control_plane/models/systemsettingsconf.py +0 -311
  55. cribl_control_plane/models/updatesystemsettingsauthop.py +0 -24
  56. cribl_control_plane/models/updatesystemsettingsconfop.py +0 -24
  57. cribl_control_plane/models/updatesystemsettingsgitsettingsop.py +0 -24
  58. cribl_control_plane/models/upgradegroupsettings.py +0 -24
  59. cribl_control_plane/models/upgradepackageurls.py +0 -20
  60. cribl_control_plane/models/upgradesettings.py +0 -36
  61. cribl_control_plane/settings.py +0 -23
  62. cribl_control_plane/settings_auth.py +0 -339
  63. cribl_control_plane/settings_git.py +0 -339
  64. cribl_control_plane/system_sdk.py +0 -17
  65. {cribl_control_plane-0.0.38a1.dist-info → cribl_control_plane-0.0.39.dist-info}/WHEEL +0 -0
@@ -32,9 +32,9 @@ class ConfigGroupType(str, Enum):
32
32
 
33
33
 
34
34
  class ConfigGroupTypedDict(TypedDict):
35
- config_version: str
36
35
  id: str
37
36
  cloud: NotRequired[ConfigGroupCloudTypedDict]
37
+ config_version: NotRequired[str]
38
38
  deploying_worker_count: NotRequired[float]
39
39
  description: NotRequired[str]
40
40
  estimated_ingest_rate: NotRequired[float]
@@ -57,12 +57,14 @@ class ConfigGroupTypedDict(TypedDict):
57
57
 
58
58
 
59
59
  class ConfigGroup(BaseModel):
60
- config_version: Annotated[str, pydantic.Field(alias="configVersion")]
61
-
62
60
  id: str
63
61
 
64
62
  cloud: Optional[ConfigGroupCloud] = None
65
63
 
64
+ config_version: Annotated[Optional[str], pydantic.Field(alias="configVersion")] = (
65
+ None
66
+ )
67
+
66
68
  deploying_worker_count: Annotated[
67
69
  Optional[float], pydantic.Field(alias="deployingWorkerCount")
68
70
  ] = None
@@ -1,10 +1,35 @@
1
1
  """Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
2
2
 
3
3
  from __future__ import annotations
4
+ from .gitcommitparams import GitCommitParams, GitCommitParamsTypedDict
4
5
  from .gitcommitsummary import GitCommitSummary, GitCommitSummaryTypedDict
5
6
  from cribl_control_plane.types import BaseModel
7
+ from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata, RequestMetadata
8
+ import pydantic
6
9
  from typing import List, Optional
7
- from typing_extensions import NotRequired, TypedDict
10
+ from typing_extensions import Annotated, NotRequired, TypedDict
11
+
12
+
13
+ class CreateVersionCommitRequestTypedDict(TypedDict):
14
+ git_commit_params: GitCommitParamsTypedDict
15
+ r"""GitCommitParams object"""
16
+ group_id: NotRequired[str]
17
+ r"""The <code>id</code> of the Worker Group or Edge Fleet to create a new commit for."""
18
+
19
+
20
+ class CreateVersionCommitRequest(BaseModel):
21
+ git_commit_params: Annotated[
22
+ GitCommitParams,
23
+ FieldMetadata(request=RequestMetadata(media_type="application/json")),
24
+ ]
25
+ r"""GitCommitParams object"""
26
+
27
+ group_id: Annotated[
28
+ Optional[str],
29
+ pydantic.Field(alias="groupId"),
30
+ FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
31
+ ] = None
32
+ r"""The <code>id</code> of the Worker Group or Edge Fleet to create a new commit for."""
8
33
 
9
34
 
10
35
  class CreateVersionCommitResponseTypedDict(TypedDict):
@@ -5,6 +5,7 @@ from .gitrevertparams import GitRevertParams, GitRevertParamsTypedDict
5
5
  from .gitrevertresult import GitRevertResult, GitRevertResultTypedDict
6
6
  from cribl_control_plane.types import BaseModel
7
7
  from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata, RequestMetadata
8
+ import pydantic
8
9
  from typing import List, Optional
9
10
  from typing_extensions import Annotated, NotRequired, TypedDict
10
11
 
@@ -12,7 +13,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
12
13
  class CreateVersionRevertRequestTypedDict(TypedDict):
13
14
  git_revert_params: GitRevertParamsTypedDict
14
15
  r"""GitRevertParams object"""
15
- group: NotRequired[str]
16
+ group_id: NotRequired[str]
16
17
  r"""Group ID"""
17
18
 
18
19
 
@@ -23,8 +24,9 @@ class CreateVersionRevertRequest(BaseModel):
23
24
  ]
24
25
  r"""GitRevertParams object"""
25
26
 
26
- group: Annotated[
27
+ group_id: Annotated[
27
28
  Optional[str],
29
+ pydantic.Field(alias="groupId"),
28
30
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
29
31
  ] = None
30
32
  r"""Group ID"""
@@ -3,18 +3,20 @@
3
3
  from __future__ import annotations
4
4
  from cribl_control_plane.types import BaseModel
5
5
  from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata
6
+ import pydantic
6
7
  from typing import List, Optional
7
8
  from typing_extensions import Annotated, NotRequired, TypedDict
8
9
 
9
10
 
10
11
  class CreateVersionUndoRequestTypedDict(TypedDict):
11
- group: NotRequired[str]
12
+ group_id: NotRequired[str]
12
13
  r"""The <code>id</code> of the Worker Group or Edge Fleet to undo the uncommited changes for."""
13
14
 
14
15
 
15
16
  class CreateVersionUndoRequest(BaseModel):
16
- group: Annotated[
17
+ group_id: Annotated[
17
18
  Optional[str],
19
+ pydantic.Field(alias="groupId"),
18
20
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
19
21
  ] = None
20
22
  r"""The <code>id</code> of the Worker Group or Edge Fleet to undo the uncommited changes for."""
@@ -9,15 +9,16 @@ from typing_extensions import Annotated, NotRequired, TypedDict
9
9
 
10
10
 
11
11
  class GetVersionCountRequestTypedDict(TypedDict):
12
- group: NotRequired[str]
12
+ group_id: NotRequired[str]
13
13
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the count for."""
14
14
  id: NotRequired[str]
15
15
  r"""The Git commit hash to use as the starting point for the count."""
16
16
 
17
17
 
18
18
  class GetVersionCountRequest(BaseModel):
19
- group: Annotated[
19
+ group_id: Annotated[
20
20
  Optional[str],
21
+ pydantic.Field(alias="groupId"),
21
22
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
22
23
  ] = None
23
24
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the count for."""
@@ -11,7 +11,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
11
11
  class GetVersionDiffRequestTypedDict(TypedDict):
12
12
  commit: NotRequired[str]
13
13
  r"""The Git commit hash to get the diff for."""
14
- group: NotRequired[str]
14
+ group_id: NotRequired[str]
15
15
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff for."""
16
16
  filename: NotRequired[str]
17
17
  r"""The relative path of the file to get the diff for."""
@@ -26,8 +26,9 @@ class GetVersionDiffRequest(BaseModel):
26
26
  ] = None
27
27
  r"""The Git commit hash to get the diff for."""
28
28
 
29
- group: Annotated[
29
+ group_id: Annotated[
30
30
  Optional[str],
31
+ pydantic.Field(alias="groupId"),
31
32
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
32
33
  ] = None
33
34
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff for."""
@@ -10,15 +10,16 @@ from typing_extensions import Annotated, NotRequired, TypedDict
10
10
 
11
11
 
12
12
  class GetVersionFilesRequestTypedDict(TypedDict):
13
- group: NotRequired[str]
13
+ group_id: NotRequired[str]
14
14
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get file names and status for."""
15
15
  id: NotRequired[str]
16
16
  r"""The Git commit hash to use as the starting point for the request."""
17
17
 
18
18
 
19
19
  class GetVersionFilesRequest(BaseModel):
20
- group: Annotated[
20
+ group_id: Annotated[
21
21
  Optional[str],
22
+ pydantic.Field(alias="groupId"),
22
23
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
23
24
  ] = None
24
25
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get file names and status for."""
@@ -4,20 +4,22 @@ from __future__ import annotations
4
4
  from .gitlogresult import GitLogResult, GitLogResultTypedDict
5
5
  from cribl_control_plane.types import BaseModel
6
6
  from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata
7
+ import pydantic
7
8
  from typing import List, Optional
8
9
  from typing_extensions import Annotated, NotRequired, TypedDict
9
10
 
10
11
 
11
12
  class GetVersionRequestTypedDict(TypedDict):
12
- group: NotRequired[str]
13
+ group_id: NotRequired[str]
13
14
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the commit history for."""
14
15
  count: NotRequired[float]
15
16
  r"""Maximum number of commits to return in the response for this request."""
16
17
 
17
18
 
18
19
  class GetVersionRequest(BaseModel):
19
- group: Annotated[
20
+ group_id: Annotated[
20
21
  Optional[str],
22
+ pydantic.Field(alias="groupId"),
21
23
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
22
24
  ] = None
23
25
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the commit history for."""
@@ -11,7 +11,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
11
11
  class GetVersionShowRequestTypedDict(TypedDict):
12
12
  commit: NotRequired[str]
13
13
  r"""The Git commit hash to retrieve the diff and log message for."""
14
- group: NotRequired[str]
14
+ group_id: NotRequired[str]
15
15
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff and log message for."""
16
16
  filename: NotRequired[str]
17
17
  r"""The relative path of the file to get the diff and log message for."""
@@ -26,8 +26,9 @@ class GetVersionShowRequest(BaseModel):
26
26
  ] = None
27
27
  r"""The Git commit hash to retrieve the diff and log message for."""
28
28
 
29
- group: Annotated[
29
+ group_id: Annotated[
30
30
  Optional[str],
31
+ pydantic.Field(alias="groupId"),
31
32
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
32
33
  ] = None
33
34
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff and log message for."""
@@ -4,18 +4,20 @@ from __future__ import annotations
4
4
  from .gitstatusresult import GitStatusResult, GitStatusResultTypedDict
5
5
  from cribl_control_plane.types import BaseModel
6
6
  from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata
7
+ import pydantic
7
8
  from typing import List, Optional
8
9
  from typing_extensions import Annotated, NotRequired, TypedDict
9
10
 
10
11
 
11
12
  class GetVersionStatusRequestTypedDict(TypedDict):
12
- group: NotRequired[str]
13
+ group_id: NotRequired[str]
13
14
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the status for."""
14
15
 
15
16
 
16
17
  class GetVersionStatusRequest(BaseModel):
17
- group: Annotated[
18
+ group_id: Annotated[
18
19
  Optional[str],
20
+ pydantic.Field(alias="groupId"),
19
21
  FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
20
22
  ] = None
21
23
  r"""The <code>id</code> of the Worker Group or Edge Fleet to get the status for."""
@@ -189,21 +189,24 @@ class InputCriblLakeHTTPAuthTokensExtMetadatum(BaseModel):
189
189
 
190
190
  class InputCriblLakeHTTPAuthTokensExtTypedDict(TypedDict):
191
191
  token: str
192
- r"""Shared secret to be provided by any client (Authorization: <token>)"""
193
192
  description: NotRequired[str]
194
193
  metadata: NotRequired[List[InputCriblLakeHTTPAuthTokensExtMetadatumTypedDict]]
195
194
  r"""Fields to add to events referencing this token"""
195
+ enable_splunk_hec: NotRequired[bool]
196
196
 
197
197
 
198
198
  class InputCriblLakeHTTPAuthTokensExt(BaseModel):
199
199
  token: str
200
- r"""Shared secret to be provided by any client (Authorization: <token>)"""
201
200
 
202
201
  description: Optional[str] = None
203
202
 
204
203
  metadata: Optional[List[InputCriblLakeHTTPAuthTokensExtMetadatum]] = None
205
204
  r"""Fields to add to events referencing this token"""
206
205
 
206
+ enable_splunk_hec: Annotated[
207
+ Optional[bool], pydantic.Field(alias="enableSplunkHec")
208
+ ] = False
209
+
207
210
 
208
211
  class InputCriblLakeHTTPTypedDict(TypedDict):
209
212
  type: InputCriblLakeHTTPType
@@ -262,7 +265,6 @@ class InputCriblLakeHTTPTypedDict(TypedDict):
262
265
  metadata: NotRequired[List[InputCriblLakeHTTPMetadatumTypedDict]]
263
266
  r"""Fields to add to events from this input"""
264
267
  auth_tokens_ext: NotRequired[List[InputCriblLakeHTTPAuthTokensExtTypedDict]]
265
- r"""Shared secrets to be provided by any client (Authorization: <token>). If empty, unauthorized access is permitted."""
266
268
  description: NotRequired[str]
267
269
 
268
270
 
@@ -388,6 +390,5 @@ class InputCriblLakeHTTP(BaseModel):
388
390
  Optional[List[InputCriblLakeHTTPAuthTokensExt]],
389
391
  pydantic.Field(alias="authTokensExt"),
390
392
  ] = None
391
- r"""Shared secrets to be provided by any client (Authorization: <token>). If empty, unauthorized access is permitted."""
392
393
 
393
394
  description: Optional[str] = None
@@ -201,7 +201,7 @@ class InputCrowdstrikeTypedDict(TypedDict):
201
201
  skip_on_error: NotRequired[bool]
202
202
  r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
203
203
  include_sqs_metadata: NotRequired[bool]
204
- r"""Include metadata from SQS notifications on outgoing events"""
204
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
205
205
  enable_assume_role: NotRequired[bool]
206
206
  r"""Use Assume Role credentials to access Amazon S3"""
207
207
  assume_role_arn: NotRequired[str]
@@ -338,7 +338,7 @@ class InputCrowdstrike(BaseModel):
338
338
  include_sqs_metadata: Annotated[
339
339
  Optional[bool], pydantic.Field(alias="includeSqsMetadata")
340
340
  ] = False
341
- r"""Include metadata from SQS notifications on outgoing events"""
341
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
342
342
 
343
343
  enable_assume_role: Annotated[
344
344
  Optional[bool], pydantic.Field(alias="enableAssumeRole")
@@ -196,7 +196,7 @@ class InputS3TypedDict(TypedDict):
196
196
  skip_on_error: NotRequired[bool]
197
197
  r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
198
198
  include_sqs_metadata: NotRequired[bool]
199
- r"""Include metadata from SQS notifications on outgoing events"""
199
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
200
200
  enable_assume_role: NotRequired[bool]
201
201
  r"""Use Assume Role credentials to access Amazon S3"""
202
202
  assume_role_arn: NotRequired[str]
@@ -337,7 +337,7 @@ class InputS3(BaseModel):
337
337
  include_sqs_metadata: Annotated[
338
338
  Optional[bool], pydantic.Field(alias="includeSqsMetadata")
339
339
  ] = False
340
- r"""Include metadata from SQS notifications on outgoing events"""
340
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
341
341
 
342
342
  enable_assume_role: Annotated[
343
343
  Optional[bool], pydantic.Field(alias="enableAssumeRole")
@@ -201,7 +201,7 @@ class InputS3InventoryTypedDict(TypedDict):
201
201
  skip_on_error: NotRequired[bool]
202
202
  r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
203
203
  include_sqs_metadata: NotRequired[bool]
204
- r"""Include metadata from SQS notifications on outgoing events"""
204
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
205
205
  enable_assume_role: NotRequired[bool]
206
206
  r"""Use Assume Role credentials to access Amazon S3"""
207
207
  assume_role_arn: NotRequired[str]
@@ -346,7 +346,7 @@ class InputS3Inventory(BaseModel):
346
346
  include_sqs_metadata: Annotated[
347
347
  Optional[bool], pydantic.Field(alias="includeSqsMetadata")
348
348
  ] = False
349
- r"""Include metadata from SQS notifications on outgoing events"""
349
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
350
350
 
351
351
  enable_assume_role: Annotated[
352
352
  Optional[bool], pydantic.Field(alias="enableAssumeRole")
@@ -201,7 +201,7 @@ class InputSecurityLakeTypedDict(TypedDict):
201
201
  skip_on_error: NotRequired[bool]
202
202
  r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
203
203
  include_sqs_metadata: NotRequired[bool]
204
- r"""Include metadata from SQS notifications on outgoing events"""
204
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
205
205
  enable_assume_role: NotRequired[bool]
206
206
  r"""Use Assume Role credentials to access Amazon S3"""
207
207
  assume_role_arn: NotRequired[str]
@@ -342,7 +342,7 @@ class InputSecurityLake(BaseModel):
342
342
  include_sqs_metadata: Annotated[
343
343
  Optional[bool], pydantic.Field(alias="includeSqsMetadata")
344
344
  ] = False
345
- r"""Include metadata from SQS notifications on outgoing events"""
345
+ r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
346
346
 
347
347
  enable_assume_role: Annotated[
348
348
  Optional[bool], pydantic.Field(alias="enableAssumeRole")
@@ -298,7 +298,7 @@ class OutputCrowdstrikeNextGenSiem(BaseModel):
298
298
  format_: Annotated[
299
299
  Optional[OutputCrowdstrikeNextGenSiemRequestFormat],
300
300
  pydantic.Field(alias="format"),
301
- ] = OutputCrowdstrikeNextGenSiemRequestFormat.RAW
301
+ ] = OutputCrowdstrikeNextGenSiemRequestFormat.JSON
302
302
  r"""When set to JSON, the event is automatically formatted with required fields before sending. When set to Raw, only the event's `_raw` value is sent."""
303
303
 
304
304
  auth_type: Annotated[
@@ -8,7 +8,7 @@ from typing import List, Optional
8
8
  from typing_extensions import Annotated, NotRequired, TypedDict
9
9
 
10
10
 
11
- class OutputGooglePubsubTypeGooglePubsub(str, Enum):
11
+ class OutputGooglePubsubType(str, Enum):
12
12
  GOOGLE_PUBSUB = "google_pubsub"
13
13
 
14
14
 
@@ -20,25 +20,6 @@ class OutputGooglePubsubGoogleAuthenticationMethod(str, Enum):
20
20
  SECRET = "secret"
21
21
 
22
22
 
23
- class FlushPeriodSecType(str, Enum):
24
- NUMBER = "number"
25
-
26
-
27
- class FlushPeriodSecTypedDict(TypedDict):
28
- r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
29
-
30
- type: NotRequired[FlushPeriodSecType]
31
- default: NotRequired[float]
32
-
33
-
34
- class FlushPeriodSec(BaseModel):
35
- r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
36
-
37
- type: Optional[FlushPeriodSecType] = None
38
-
39
- default: Optional[float] = None
40
-
41
-
42
23
  class OutputGooglePubsubBackpressureBehavior(str, Enum):
43
24
  r"""How to handle events when all receivers are exerting backpressure"""
44
25
 
@@ -78,7 +59,7 @@ class OutputGooglePubsubPqControls(BaseModel):
78
59
 
79
60
 
80
61
  class OutputGooglePubsubTypedDict(TypedDict):
81
- type: OutputGooglePubsubTypeGooglePubsub
62
+ type: OutputGooglePubsubType
82
63
  topic_name: str
83
64
  r"""ID of the topic to send events to."""
84
65
  id: NotRequired[str]
@@ -111,8 +92,8 @@ class OutputGooglePubsubTypedDict(TypedDict):
111
92
  r"""Maximum number of queued batches before blocking."""
112
93
  max_record_size_kb: NotRequired[float]
113
94
  r"""Maximum size (KB) of batches to send."""
114
- flush_period_sec: NotRequired[FlushPeriodSecTypedDict]
115
- r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
95
+ flush_period: NotRequired[float]
96
+ r"""Maximum time to wait before sending a batch (when batch size limit is not reached)"""
116
97
  max_in_progress: NotRequired[float]
117
98
  r"""The maximum number of in-progress API requests before backpressure is applied."""
118
99
  on_backpressure: NotRequired[OutputGooglePubsubBackpressureBehavior]
@@ -134,7 +115,7 @@ class OutputGooglePubsubTypedDict(TypedDict):
134
115
 
135
116
 
136
117
  class OutputGooglePubsub(BaseModel):
137
- type: OutputGooglePubsubTypeGooglePubsub
118
+ type: OutputGooglePubsubType
138
119
 
139
120
  topic_name: Annotated[str, pydantic.Field(alias="topicName")]
140
121
  r"""ID of the topic to send events to."""
@@ -199,10 +180,8 @@ class OutputGooglePubsub(BaseModel):
199
180
  ] = 256
200
181
  r"""Maximum size (KB) of batches to send."""
201
182
 
202
- flush_period_sec: Annotated[
203
- Optional[FlushPeriodSec], pydantic.Field(alias="flushPeriodSec")
204
- ] = None
205
- r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
183
+ flush_period: Annotated[Optional[float], pydantic.Field(alias="flushPeriod")] = 1
184
+ r"""Maximum time to wait before sending a batch (when batch size limit is not reached)"""
206
185
 
207
186
  max_in_progress: Annotated[
208
187
  Optional[float], pydantic.Field(alias="maxInProgress")
@@ -25,7 +25,7 @@ class OutputSentinelOneAiSiemRegion(str, Enum):
25
25
 
26
26
 
27
27
  class AISIEMEndpointPath(str, Enum):
28
- r"""Regional endpoint used to send events to, such as /services/collector/event or /services/collector/raw"""
28
+ r"""Endpoint to send events to. Use /services/collector/event for structured JSON payloads with standard HEC top-level fields. Use /services/collector/raw for unstructured log lines (plain text)."""
29
29
 
30
30
  ROOT_SERVICES_COLLECTOR_EVENT = "/services/collector/event"
31
31
  ROOT_SERVICES_COLLECTOR_RAW = "/services/collector/raw"
@@ -164,7 +164,7 @@ class OutputSentinelOneAiSiemTypedDict(TypedDict):
164
164
  region: NotRequired[OutputSentinelOneAiSiemRegion]
165
165
  r"""The SentinelOne region to send events to. In most cases you can find the region by either looking at your SentinelOne URL or knowing what geographic region your SentinelOne instance is contained in."""
166
166
  endpoint: NotRequired[AISIEMEndpointPath]
167
- r"""Regional endpoint used to send events to, such as /services/collector/event or /services/collector/raw"""
167
+ r"""Endpoint to send events to. Use /services/collector/event for structured JSON payloads with standard HEC top-level fields. Use /services/collector/raw for unstructured log lines (plain text)."""
168
168
  concurrency: NotRequired[float]
169
169
  r"""Maximum number of ongoing requests before blocking"""
170
170
  max_payload_size_kb: NotRequired[float]
@@ -281,7 +281,7 @@ class OutputSentinelOneAiSiem(BaseModel):
281
281
  endpoint: Optional[AISIEMEndpointPath] = (
282
282
  AISIEMEndpointPath.ROOT_SERVICES_COLLECTOR_EVENT
283
283
  )
284
- r"""Regional endpoint used to send events to, such as /services/collector/event or /services/collector/raw"""
284
+ r"""Endpoint to send events to. Use /services/collector/event for structured JSON payloads with standard HEC top-level fields. Use /services/collector/raw for unstructured log lines (plain text)."""
285
285
 
286
286
  concurrency: Optional[float] = 5
287
287
  r"""Maximum number of ongoing requests before blocking"""
@@ -9,19 +9,19 @@ from typing_extensions import Annotated, NotRequired, TypedDict
9
9
 
10
10
  class PackInfoTagsTypedDict(TypedDict):
11
11
  data_type: List[str]
12
- domain: List[str]
13
- streamtags: List[str]
14
12
  technology: List[str]
13
+ domain: NotRequired[List[str]]
14
+ streamtags: NotRequired[List[str]]
15
15
 
16
16
 
17
17
  class PackInfoTags(BaseModel):
18
18
  data_type: Annotated[List[str], pydantic.Field(alias="dataType")]
19
19
 
20
- domain: List[str]
20
+ technology: List[str]
21
21
 
22
- streamtags: List[str]
22
+ domain: Optional[List[str]] = None
23
23
 
24
- technology: List[str]
24
+ streamtags: Optional[List[str]] = None
25
25
 
26
26
 
27
27
  class PackInfoTypedDict(TypedDict):
@@ -9,19 +9,19 @@ from typing_extensions import Annotated, NotRequired, TypedDict
9
9
 
10
10
  class PackInstallInfoTagsTypedDict(TypedDict):
11
11
  data_type: List[str]
12
- domain: List[str]
13
- streamtags: List[str]
14
12
  technology: List[str]
13
+ domain: NotRequired[List[str]]
14
+ streamtags: NotRequired[List[str]]
15
15
 
16
16
 
17
17
  class PackInstallInfoTags(BaseModel):
18
18
  data_type: Annotated[List[str], pydantic.Field(alias="dataType")]
19
19
 
20
- domain: List[str]
20
+ technology: List[str]
21
21
 
22
- streamtags: List[str]
22
+ domain: Optional[List[str]] = None
23
23
 
24
- technology: List[str]
24
+ streamtags: Optional[List[str]] = None
25
25
 
26
26
 
27
27
  class PackInstallInfoTypedDict(TypedDict):