cribl-control-plane 0.0.38a1__py3-none-any.whl → 0.0.39__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of cribl-control-plane might be problematic. Click here for more details.
- cribl_control_plane/_version.py +3 -3
- cribl_control_plane/commits.py +52 -42
- cribl_control_plane/commits_files.py +12 -12
- cribl_control_plane/destinations.py +3 -3
- cribl_control_plane/groups_sdk.py +8 -8
- cribl_control_plane/models/__init__.py +46 -368
- cribl_control_plane/models/configgroup.py +5 -3
- cribl_control_plane/models/createversioncommitop.py +26 -1
- cribl_control_plane/models/createversionrevertop.py +4 -2
- cribl_control_plane/models/createversionundoop.py +4 -2
- cribl_control_plane/models/getversioncountop.py +3 -2
- cribl_control_plane/models/getversiondiffop.py +3 -2
- cribl_control_plane/models/getversionfilesop.py +3 -2
- cribl_control_plane/models/getversionop.py +4 -2
- cribl_control_plane/models/getversionshowop.py +3 -2
- cribl_control_plane/models/getversionstatusop.py +4 -2
- cribl_control_plane/models/inputcribllakehttp.py +5 -4
- cribl_control_plane/models/inputcrowdstrike.py +2 -2
- cribl_control_plane/models/inputs3.py +2 -2
- cribl_control_plane/models/inputs3inventory.py +2 -2
- cribl_control_plane/models/inputsecuritylake.py +2 -2
- cribl_control_plane/models/outputcrowdstrikenextgensiem.py +1 -1
- cribl_control_plane/models/outputgooglepubsub.py +7 -28
- cribl_control_plane/models/outputsentineloneaisiem.py +3 -3
- cribl_control_plane/models/packinfo.py +5 -5
- cribl_control_plane/models/packinstallinfo.py +5 -5
- cribl_control_plane/models/packrequestbody_union.py +140 -0
- cribl_control_plane/models/packupgraderequest.py +26 -0
- cribl_control_plane/models/updatepacksbyidop.py +9 -28
- cribl_control_plane/packs.py +62 -116
- cribl_control_plane/{destinations_samples.py → samples.py} +1 -1
- cribl_control_plane/sdk.py +0 -3
- cribl_control_plane/statuses.py +6 -6
- {cribl_control_plane-0.0.38a1.dist-info → cribl_control_plane-0.0.39.dist-info}/METADATA +6 -22
- {cribl_control_plane-0.0.38a1.dist-info → cribl_control_plane-0.0.39.dist-info}/RECORD +36 -63
- cribl_control_plane/cribl.py +0 -513
- cribl_control_plane/models/authconfig.py +0 -43
- cribl_control_plane/models/commonservicelimitconfigs.py +0 -14
- cribl_control_plane/models/edgeheartbeatmetricsmode.py +0 -11
- cribl_control_plane/models/getsystemsettingsauthop.py +0 -24
- cribl_control_plane/models/getsystemsettingsconfop.py +0 -24
- cribl_control_plane/models/getsystemsettingsgitsettingsop.py +0 -24
- cribl_control_plane/models/gitopstype.py +0 -10
- cribl_control_plane/models/gitsettings.py +0 -70
- cribl_control_plane/models/jobsettings.py +0 -83
- cribl_control_plane/models/limits.py +0 -127
- cribl_control_plane/models/packrequestbody.py +0 -75
- cribl_control_plane/models/rediscachelimits.py +0 -38
- cribl_control_plane/models/redisconnectionlimits.py +0 -20
- cribl_control_plane/models/redislimits.py +0 -14
- cribl_control_plane/models/searchsettings.py +0 -71
- cribl_control_plane/models/serviceslimits.py +0 -23
- cribl_control_plane/models/systemsettings.py +0 -358
- cribl_control_plane/models/systemsettingsconf.py +0 -311
- cribl_control_plane/models/updatesystemsettingsauthop.py +0 -24
- cribl_control_plane/models/updatesystemsettingsconfop.py +0 -24
- cribl_control_plane/models/updatesystemsettingsgitsettingsop.py +0 -24
- cribl_control_plane/models/upgradegroupsettings.py +0 -24
- cribl_control_plane/models/upgradepackageurls.py +0 -20
- cribl_control_plane/models/upgradesettings.py +0 -36
- cribl_control_plane/settings.py +0 -23
- cribl_control_plane/settings_auth.py +0 -339
- cribl_control_plane/settings_git.py +0 -339
- cribl_control_plane/system_sdk.py +0 -17
- {cribl_control_plane-0.0.38a1.dist-info → cribl_control_plane-0.0.39.dist-info}/WHEEL +0 -0
|
@@ -32,9 +32,9 @@ class ConfigGroupType(str, Enum):
|
|
|
32
32
|
|
|
33
33
|
|
|
34
34
|
class ConfigGroupTypedDict(TypedDict):
|
|
35
|
-
config_version: str
|
|
36
35
|
id: str
|
|
37
36
|
cloud: NotRequired[ConfigGroupCloudTypedDict]
|
|
37
|
+
config_version: NotRequired[str]
|
|
38
38
|
deploying_worker_count: NotRequired[float]
|
|
39
39
|
description: NotRequired[str]
|
|
40
40
|
estimated_ingest_rate: NotRequired[float]
|
|
@@ -57,12 +57,14 @@ class ConfigGroupTypedDict(TypedDict):
|
|
|
57
57
|
|
|
58
58
|
|
|
59
59
|
class ConfigGroup(BaseModel):
|
|
60
|
-
config_version: Annotated[str, pydantic.Field(alias="configVersion")]
|
|
61
|
-
|
|
62
60
|
id: str
|
|
63
61
|
|
|
64
62
|
cloud: Optional[ConfigGroupCloud] = None
|
|
65
63
|
|
|
64
|
+
config_version: Annotated[Optional[str], pydantic.Field(alias="configVersion")] = (
|
|
65
|
+
None
|
|
66
|
+
)
|
|
67
|
+
|
|
66
68
|
deploying_worker_count: Annotated[
|
|
67
69
|
Optional[float], pydantic.Field(alias="deployingWorkerCount")
|
|
68
70
|
] = None
|
|
@@ -1,10 +1,35 @@
|
|
|
1
1
|
"""Code generated by Speakeasy (https://speakeasy.com). DO NOT EDIT."""
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
|
+
from .gitcommitparams import GitCommitParams, GitCommitParamsTypedDict
|
|
4
5
|
from .gitcommitsummary import GitCommitSummary, GitCommitSummaryTypedDict
|
|
5
6
|
from cribl_control_plane.types import BaseModel
|
|
7
|
+
from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata, RequestMetadata
|
|
8
|
+
import pydantic
|
|
6
9
|
from typing import List, Optional
|
|
7
|
-
from typing_extensions import NotRequired, TypedDict
|
|
10
|
+
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
11
|
+
|
|
12
|
+
|
|
13
|
+
class CreateVersionCommitRequestTypedDict(TypedDict):
|
|
14
|
+
git_commit_params: GitCommitParamsTypedDict
|
|
15
|
+
r"""GitCommitParams object"""
|
|
16
|
+
group_id: NotRequired[str]
|
|
17
|
+
r"""The <code>id</code> of the Worker Group or Edge Fleet to create a new commit for."""
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
class CreateVersionCommitRequest(BaseModel):
|
|
21
|
+
git_commit_params: Annotated[
|
|
22
|
+
GitCommitParams,
|
|
23
|
+
FieldMetadata(request=RequestMetadata(media_type="application/json")),
|
|
24
|
+
]
|
|
25
|
+
r"""GitCommitParams object"""
|
|
26
|
+
|
|
27
|
+
group_id: Annotated[
|
|
28
|
+
Optional[str],
|
|
29
|
+
pydantic.Field(alias="groupId"),
|
|
30
|
+
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
31
|
+
] = None
|
|
32
|
+
r"""The <code>id</code> of the Worker Group or Edge Fleet to create a new commit for."""
|
|
8
33
|
|
|
9
34
|
|
|
10
35
|
class CreateVersionCommitResponseTypedDict(TypedDict):
|
|
@@ -5,6 +5,7 @@ from .gitrevertparams import GitRevertParams, GitRevertParamsTypedDict
|
|
|
5
5
|
from .gitrevertresult import GitRevertResult, GitRevertResultTypedDict
|
|
6
6
|
from cribl_control_plane.types import BaseModel
|
|
7
7
|
from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata, RequestMetadata
|
|
8
|
+
import pydantic
|
|
8
9
|
from typing import List, Optional
|
|
9
10
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
10
11
|
|
|
@@ -12,7 +13,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
12
13
|
class CreateVersionRevertRequestTypedDict(TypedDict):
|
|
13
14
|
git_revert_params: GitRevertParamsTypedDict
|
|
14
15
|
r"""GitRevertParams object"""
|
|
15
|
-
|
|
16
|
+
group_id: NotRequired[str]
|
|
16
17
|
r"""Group ID"""
|
|
17
18
|
|
|
18
19
|
|
|
@@ -23,8 +24,9 @@ class CreateVersionRevertRequest(BaseModel):
|
|
|
23
24
|
]
|
|
24
25
|
r"""GitRevertParams object"""
|
|
25
26
|
|
|
26
|
-
|
|
27
|
+
group_id: Annotated[
|
|
27
28
|
Optional[str],
|
|
29
|
+
pydantic.Field(alias="groupId"),
|
|
28
30
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
29
31
|
] = None
|
|
30
32
|
r"""Group ID"""
|
|
@@ -3,18 +3,20 @@
|
|
|
3
3
|
from __future__ import annotations
|
|
4
4
|
from cribl_control_plane.types import BaseModel
|
|
5
5
|
from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata
|
|
6
|
+
import pydantic
|
|
6
7
|
from typing import List, Optional
|
|
7
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
8
9
|
|
|
9
10
|
|
|
10
11
|
class CreateVersionUndoRequestTypedDict(TypedDict):
|
|
11
|
-
|
|
12
|
+
group_id: NotRequired[str]
|
|
12
13
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to undo the uncommited changes for."""
|
|
13
14
|
|
|
14
15
|
|
|
15
16
|
class CreateVersionUndoRequest(BaseModel):
|
|
16
|
-
|
|
17
|
+
group_id: Annotated[
|
|
17
18
|
Optional[str],
|
|
19
|
+
pydantic.Field(alias="groupId"),
|
|
18
20
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
19
21
|
] = None
|
|
20
22
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to undo the uncommited changes for."""
|
|
@@ -9,15 +9,16 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
class GetVersionCountRequestTypedDict(TypedDict):
|
|
12
|
-
|
|
12
|
+
group_id: NotRequired[str]
|
|
13
13
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the count for."""
|
|
14
14
|
id: NotRequired[str]
|
|
15
15
|
r"""The Git commit hash to use as the starting point for the count."""
|
|
16
16
|
|
|
17
17
|
|
|
18
18
|
class GetVersionCountRequest(BaseModel):
|
|
19
|
-
|
|
19
|
+
group_id: Annotated[
|
|
20
20
|
Optional[str],
|
|
21
|
+
pydantic.Field(alias="groupId"),
|
|
21
22
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
22
23
|
] = None
|
|
23
24
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the count for."""
|
|
@@ -11,7 +11,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
11
11
|
class GetVersionDiffRequestTypedDict(TypedDict):
|
|
12
12
|
commit: NotRequired[str]
|
|
13
13
|
r"""The Git commit hash to get the diff for."""
|
|
14
|
-
|
|
14
|
+
group_id: NotRequired[str]
|
|
15
15
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff for."""
|
|
16
16
|
filename: NotRequired[str]
|
|
17
17
|
r"""The relative path of the file to get the diff for."""
|
|
@@ -26,8 +26,9 @@ class GetVersionDiffRequest(BaseModel):
|
|
|
26
26
|
] = None
|
|
27
27
|
r"""The Git commit hash to get the diff for."""
|
|
28
28
|
|
|
29
|
-
|
|
29
|
+
group_id: Annotated[
|
|
30
30
|
Optional[str],
|
|
31
|
+
pydantic.Field(alias="groupId"),
|
|
31
32
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
32
33
|
] = None
|
|
33
34
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff for."""
|
|
@@ -10,15 +10,16 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
10
10
|
|
|
11
11
|
|
|
12
12
|
class GetVersionFilesRequestTypedDict(TypedDict):
|
|
13
|
-
|
|
13
|
+
group_id: NotRequired[str]
|
|
14
14
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get file names and status for."""
|
|
15
15
|
id: NotRequired[str]
|
|
16
16
|
r"""The Git commit hash to use as the starting point for the request."""
|
|
17
17
|
|
|
18
18
|
|
|
19
19
|
class GetVersionFilesRequest(BaseModel):
|
|
20
|
-
|
|
20
|
+
group_id: Annotated[
|
|
21
21
|
Optional[str],
|
|
22
|
+
pydantic.Field(alias="groupId"),
|
|
22
23
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
23
24
|
] = None
|
|
24
25
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get file names and status for."""
|
|
@@ -4,20 +4,22 @@ from __future__ import annotations
|
|
|
4
4
|
from .gitlogresult import GitLogResult, GitLogResultTypedDict
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata
|
|
7
|
+
import pydantic
|
|
7
8
|
from typing import List, Optional
|
|
8
9
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
10
|
|
|
10
11
|
|
|
11
12
|
class GetVersionRequestTypedDict(TypedDict):
|
|
12
|
-
|
|
13
|
+
group_id: NotRequired[str]
|
|
13
14
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the commit history for."""
|
|
14
15
|
count: NotRequired[float]
|
|
15
16
|
r"""Maximum number of commits to return in the response for this request."""
|
|
16
17
|
|
|
17
18
|
|
|
18
19
|
class GetVersionRequest(BaseModel):
|
|
19
|
-
|
|
20
|
+
group_id: Annotated[
|
|
20
21
|
Optional[str],
|
|
22
|
+
pydantic.Field(alias="groupId"),
|
|
21
23
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
22
24
|
] = None
|
|
23
25
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the commit history for."""
|
|
@@ -11,7 +11,7 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
11
11
|
class GetVersionShowRequestTypedDict(TypedDict):
|
|
12
12
|
commit: NotRequired[str]
|
|
13
13
|
r"""The Git commit hash to retrieve the diff and log message for."""
|
|
14
|
-
|
|
14
|
+
group_id: NotRequired[str]
|
|
15
15
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff and log message for."""
|
|
16
16
|
filename: NotRequired[str]
|
|
17
17
|
r"""The relative path of the file to get the diff and log message for."""
|
|
@@ -26,8 +26,9 @@ class GetVersionShowRequest(BaseModel):
|
|
|
26
26
|
] = None
|
|
27
27
|
r"""The Git commit hash to retrieve the diff and log message for."""
|
|
28
28
|
|
|
29
|
-
|
|
29
|
+
group_id: Annotated[
|
|
30
30
|
Optional[str],
|
|
31
|
+
pydantic.Field(alias="groupId"),
|
|
31
32
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
32
33
|
] = None
|
|
33
34
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the diff and log message for."""
|
|
@@ -4,18 +4,20 @@ from __future__ import annotations
|
|
|
4
4
|
from .gitstatusresult import GitStatusResult, GitStatusResultTypedDict
|
|
5
5
|
from cribl_control_plane.types import BaseModel
|
|
6
6
|
from cribl_control_plane.utils import FieldMetadata, QueryParamMetadata
|
|
7
|
+
import pydantic
|
|
7
8
|
from typing import List, Optional
|
|
8
9
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
10
|
|
|
10
11
|
|
|
11
12
|
class GetVersionStatusRequestTypedDict(TypedDict):
|
|
12
|
-
|
|
13
|
+
group_id: NotRequired[str]
|
|
13
14
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the status for."""
|
|
14
15
|
|
|
15
16
|
|
|
16
17
|
class GetVersionStatusRequest(BaseModel):
|
|
17
|
-
|
|
18
|
+
group_id: Annotated[
|
|
18
19
|
Optional[str],
|
|
20
|
+
pydantic.Field(alias="groupId"),
|
|
19
21
|
FieldMetadata(query=QueryParamMetadata(style="form", explode=True)),
|
|
20
22
|
] = None
|
|
21
23
|
r"""The <code>id</code> of the Worker Group or Edge Fleet to get the status for."""
|
|
@@ -189,21 +189,24 @@ class InputCriblLakeHTTPAuthTokensExtMetadatum(BaseModel):
|
|
|
189
189
|
|
|
190
190
|
class InputCriblLakeHTTPAuthTokensExtTypedDict(TypedDict):
|
|
191
191
|
token: str
|
|
192
|
-
r"""Shared secret to be provided by any client (Authorization: <token>)"""
|
|
193
192
|
description: NotRequired[str]
|
|
194
193
|
metadata: NotRequired[List[InputCriblLakeHTTPAuthTokensExtMetadatumTypedDict]]
|
|
195
194
|
r"""Fields to add to events referencing this token"""
|
|
195
|
+
enable_splunk_hec: NotRequired[bool]
|
|
196
196
|
|
|
197
197
|
|
|
198
198
|
class InputCriblLakeHTTPAuthTokensExt(BaseModel):
|
|
199
199
|
token: str
|
|
200
|
-
r"""Shared secret to be provided by any client (Authorization: <token>)"""
|
|
201
200
|
|
|
202
201
|
description: Optional[str] = None
|
|
203
202
|
|
|
204
203
|
metadata: Optional[List[InputCriblLakeHTTPAuthTokensExtMetadatum]] = None
|
|
205
204
|
r"""Fields to add to events referencing this token"""
|
|
206
205
|
|
|
206
|
+
enable_splunk_hec: Annotated[
|
|
207
|
+
Optional[bool], pydantic.Field(alias="enableSplunkHec")
|
|
208
|
+
] = False
|
|
209
|
+
|
|
207
210
|
|
|
208
211
|
class InputCriblLakeHTTPTypedDict(TypedDict):
|
|
209
212
|
type: InputCriblLakeHTTPType
|
|
@@ -262,7 +265,6 @@ class InputCriblLakeHTTPTypedDict(TypedDict):
|
|
|
262
265
|
metadata: NotRequired[List[InputCriblLakeHTTPMetadatumTypedDict]]
|
|
263
266
|
r"""Fields to add to events from this input"""
|
|
264
267
|
auth_tokens_ext: NotRequired[List[InputCriblLakeHTTPAuthTokensExtTypedDict]]
|
|
265
|
-
r"""Shared secrets to be provided by any client (Authorization: <token>). If empty, unauthorized access is permitted."""
|
|
266
268
|
description: NotRequired[str]
|
|
267
269
|
|
|
268
270
|
|
|
@@ -388,6 +390,5 @@ class InputCriblLakeHTTP(BaseModel):
|
|
|
388
390
|
Optional[List[InputCriblLakeHTTPAuthTokensExt]],
|
|
389
391
|
pydantic.Field(alias="authTokensExt"),
|
|
390
392
|
] = None
|
|
391
|
-
r"""Shared secrets to be provided by any client (Authorization: <token>). If empty, unauthorized access is permitted."""
|
|
392
393
|
|
|
393
394
|
description: Optional[str] = None
|
|
@@ -201,7 +201,7 @@ class InputCrowdstrikeTypedDict(TypedDict):
|
|
|
201
201
|
skip_on_error: NotRequired[bool]
|
|
202
202
|
r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
|
|
203
203
|
include_sqs_metadata: NotRequired[bool]
|
|
204
|
-
r"""
|
|
204
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
205
205
|
enable_assume_role: NotRequired[bool]
|
|
206
206
|
r"""Use Assume Role credentials to access Amazon S3"""
|
|
207
207
|
assume_role_arn: NotRequired[str]
|
|
@@ -338,7 +338,7 @@ class InputCrowdstrike(BaseModel):
|
|
|
338
338
|
include_sqs_metadata: Annotated[
|
|
339
339
|
Optional[bool], pydantic.Field(alias="includeSqsMetadata")
|
|
340
340
|
] = False
|
|
341
|
-
r"""
|
|
341
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
342
342
|
|
|
343
343
|
enable_assume_role: Annotated[
|
|
344
344
|
Optional[bool], pydantic.Field(alias="enableAssumeRole")
|
|
@@ -196,7 +196,7 @@ class InputS3TypedDict(TypedDict):
|
|
|
196
196
|
skip_on_error: NotRequired[bool]
|
|
197
197
|
r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
|
|
198
198
|
include_sqs_metadata: NotRequired[bool]
|
|
199
|
-
r"""
|
|
199
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
200
200
|
enable_assume_role: NotRequired[bool]
|
|
201
201
|
r"""Use Assume Role credentials to access Amazon S3"""
|
|
202
202
|
assume_role_arn: NotRequired[str]
|
|
@@ -337,7 +337,7 @@ class InputS3(BaseModel):
|
|
|
337
337
|
include_sqs_metadata: Annotated[
|
|
338
338
|
Optional[bool], pydantic.Field(alias="includeSqsMetadata")
|
|
339
339
|
] = False
|
|
340
|
-
r"""
|
|
340
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
341
341
|
|
|
342
342
|
enable_assume_role: Annotated[
|
|
343
343
|
Optional[bool], pydantic.Field(alias="enableAssumeRole")
|
|
@@ -201,7 +201,7 @@ class InputS3InventoryTypedDict(TypedDict):
|
|
|
201
201
|
skip_on_error: NotRequired[bool]
|
|
202
202
|
r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
|
|
203
203
|
include_sqs_metadata: NotRequired[bool]
|
|
204
|
-
r"""
|
|
204
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
205
205
|
enable_assume_role: NotRequired[bool]
|
|
206
206
|
r"""Use Assume Role credentials to access Amazon S3"""
|
|
207
207
|
assume_role_arn: NotRequired[str]
|
|
@@ -346,7 +346,7 @@ class InputS3Inventory(BaseModel):
|
|
|
346
346
|
include_sqs_metadata: Annotated[
|
|
347
347
|
Optional[bool], pydantic.Field(alias="includeSqsMetadata")
|
|
348
348
|
] = False
|
|
349
|
-
r"""
|
|
349
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
350
350
|
|
|
351
351
|
enable_assume_role: Annotated[
|
|
352
352
|
Optional[bool], pydantic.Field(alias="enableAssumeRole")
|
|
@@ -201,7 +201,7 @@ class InputSecurityLakeTypedDict(TypedDict):
|
|
|
201
201
|
skip_on_error: NotRequired[bool]
|
|
202
202
|
r"""Skip files that trigger a processing error. Disabled by default, which allows retries after processing errors."""
|
|
203
203
|
include_sqs_metadata: NotRequired[bool]
|
|
204
|
-
r"""
|
|
204
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
205
205
|
enable_assume_role: NotRequired[bool]
|
|
206
206
|
r"""Use Assume Role credentials to access Amazon S3"""
|
|
207
207
|
assume_role_arn: NotRequired[str]
|
|
@@ -342,7 +342,7 @@ class InputSecurityLake(BaseModel):
|
|
|
342
342
|
include_sqs_metadata: Annotated[
|
|
343
343
|
Optional[bool], pydantic.Field(alias="includeSqsMetadata")
|
|
344
344
|
] = False
|
|
345
|
-
r"""
|
|
345
|
+
r"""Attach SQS notification metadata to a __sqsMetadata field on each event"""
|
|
346
346
|
|
|
347
347
|
enable_assume_role: Annotated[
|
|
348
348
|
Optional[bool], pydantic.Field(alias="enableAssumeRole")
|
|
@@ -298,7 +298,7 @@ class OutputCrowdstrikeNextGenSiem(BaseModel):
|
|
|
298
298
|
format_: Annotated[
|
|
299
299
|
Optional[OutputCrowdstrikeNextGenSiemRequestFormat],
|
|
300
300
|
pydantic.Field(alias="format"),
|
|
301
|
-
] = OutputCrowdstrikeNextGenSiemRequestFormat.
|
|
301
|
+
] = OutputCrowdstrikeNextGenSiemRequestFormat.JSON
|
|
302
302
|
r"""When set to JSON, the event is automatically formatted with required fields before sending. When set to Raw, only the event's `_raw` value is sent."""
|
|
303
303
|
|
|
304
304
|
auth_type: Annotated[
|
|
@@ -8,7 +8,7 @@ from typing import List, Optional
|
|
|
8
8
|
from typing_extensions import Annotated, NotRequired, TypedDict
|
|
9
9
|
|
|
10
10
|
|
|
11
|
-
class
|
|
11
|
+
class OutputGooglePubsubType(str, Enum):
|
|
12
12
|
GOOGLE_PUBSUB = "google_pubsub"
|
|
13
13
|
|
|
14
14
|
|
|
@@ -20,25 +20,6 @@ class OutputGooglePubsubGoogleAuthenticationMethod(str, Enum):
|
|
|
20
20
|
SECRET = "secret"
|
|
21
21
|
|
|
22
22
|
|
|
23
|
-
class FlushPeriodSecType(str, Enum):
|
|
24
|
-
NUMBER = "number"
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
class FlushPeriodSecTypedDict(TypedDict):
|
|
28
|
-
r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
|
|
29
|
-
|
|
30
|
-
type: NotRequired[FlushPeriodSecType]
|
|
31
|
-
default: NotRequired[float]
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
class FlushPeriodSec(BaseModel):
|
|
35
|
-
r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
|
|
36
|
-
|
|
37
|
-
type: Optional[FlushPeriodSecType] = None
|
|
38
|
-
|
|
39
|
-
default: Optional[float] = None
|
|
40
|
-
|
|
41
|
-
|
|
42
23
|
class OutputGooglePubsubBackpressureBehavior(str, Enum):
|
|
43
24
|
r"""How to handle events when all receivers are exerting backpressure"""
|
|
44
25
|
|
|
@@ -78,7 +59,7 @@ class OutputGooglePubsubPqControls(BaseModel):
|
|
|
78
59
|
|
|
79
60
|
|
|
80
61
|
class OutputGooglePubsubTypedDict(TypedDict):
|
|
81
|
-
type:
|
|
62
|
+
type: OutputGooglePubsubType
|
|
82
63
|
topic_name: str
|
|
83
64
|
r"""ID of the topic to send events to."""
|
|
84
65
|
id: NotRequired[str]
|
|
@@ -111,8 +92,8 @@ class OutputGooglePubsubTypedDict(TypedDict):
|
|
|
111
92
|
r"""Maximum number of queued batches before blocking."""
|
|
112
93
|
max_record_size_kb: NotRequired[float]
|
|
113
94
|
r"""Maximum size (KB) of batches to send."""
|
|
114
|
-
|
|
115
|
-
r"""Maximum time to wait before sending a batch (when batch size limit is not reached)
|
|
95
|
+
flush_period: NotRequired[float]
|
|
96
|
+
r"""Maximum time to wait before sending a batch (when batch size limit is not reached)"""
|
|
116
97
|
max_in_progress: NotRequired[float]
|
|
117
98
|
r"""The maximum number of in-progress API requests before backpressure is applied."""
|
|
118
99
|
on_backpressure: NotRequired[OutputGooglePubsubBackpressureBehavior]
|
|
@@ -134,7 +115,7 @@ class OutputGooglePubsubTypedDict(TypedDict):
|
|
|
134
115
|
|
|
135
116
|
|
|
136
117
|
class OutputGooglePubsub(BaseModel):
|
|
137
|
-
type:
|
|
118
|
+
type: OutputGooglePubsubType
|
|
138
119
|
|
|
139
120
|
topic_name: Annotated[str, pydantic.Field(alias="topicName")]
|
|
140
121
|
r"""ID of the topic to send events to."""
|
|
@@ -199,10 +180,8 @@ class OutputGooglePubsub(BaseModel):
|
|
|
199
180
|
] = 256
|
|
200
181
|
r"""Maximum size (KB) of batches to send."""
|
|
201
182
|
|
|
202
|
-
|
|
203
|
-
|
|
204
|
-
] = None
|
|
205
|
-
r"""Maximum time to wait before sending a batch (when batch size limit is not reached)."""
|
|
183
|
+
flush_period: Annotated[Optional[float], pydantic.Field(alias="flushPeriod")] = 1
|
|
184
|
+
r"""Maximum time to wait before sending a batch (when batch size limit is not reached)"""
|
|
206
185
|
|
|
207
186
|
max_in_progress: Annotated[
|
|
208
187
|
Optional[float], pydantic.Field(alias="maxInProgress")
|
|
@@ -25,7 +25,7 @@ class OutputSentinelOneAiSiemRegion(str, Enum):
|
|
|
25
25
|
|
|
26
26
|
|
|
27
27
|
class AISIEMEndpointPath(str, Enum):
|
|
28
|
-
r"""
|
|
28
|
+
r"""Endpoint to send events to. Use /services/collector/event for structured JSON payloads with standard HEC top-level fields. Use /services/collector/raw for unstructured log lines (plain text)."""
|
|
29
29
|
|
|
30
30
|
ROOT_SERVICES_COLLECTOR_EVENT = "/services/collector/event"
|
|
31
31
|
ROOT_SERVICES_COLLECTOR_RAW = "/services/collector/raw"
|
|
@@ -164,7 +164,7 @@ class OutputSentinelOneAiSiemTypedDict(TypedDict):
|
|
|
164
164
|
region: NotRequired[OutputSentinelOneAiSiemRegion]
|
|
165
165
|
r"""The SentinelOne region to send events to. In most cases you can find the region by either looking at your SentinelOne URL or knowing what geographic region your SentinelOne instance is contained in."""
|
|
166
166
|
endpoint: NotRequired[AISIEMEndpointPath]
|
|
167
|
-
r"""
|
|
167
|
+
r"""Endpoint to send events to. Use /services/collector/event for structured JSON payloads with standard HEC top-level fields. Use /services/collector/raw for unstructured log lines (plain text)."""
|
|
168
168
|
concurrency: NotRequired[float]
|
|
169
169
|
r"""Maximum number of ongoing requests before blocking"""
|
|
170
170
|
max_payload_size_kb: NotRequired[float]
|
|
@@ -281,7 +281,7 @@ class OutputSentinelOneAiSiem(BaseModel):
|
|
|
281
281
|
endpoint: Optional[AISIEMEndpointPath] = (
|
|
282
282
|
AISIEMEndpointPath.ROOT_SERVICES_COLLECTOR_EVENT
|
|
283
283
|
)
|
|
284
|
-
r"""
|
|
284
|
+
r"""Endpoint to send events to. Use /services/collector/event for structured JSON payloads with standard HEC top-level fields. Use /services/collector/raw for unstructured log lines (plain text)."""
|
|
285
285
|
|
|
286
286
|
concurrency: Optional[float] = 5
|
|
287
287
|
r"""Maximum number of ongoing requests before blocking"""
|
|
@@ -9,19 +9,19 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
9
9
|
|
|
10
10
|
class PackInfoTagsTypedDict(TypedDict):
|
|
11
11
|
data_type: List[str]
|
|
12
|
-
domain: List[str]
|
|
13
|
-
streamtags: List[str]
|
|
14
12
|
technology: List[str]
|
|
13
|
+
domain: NotRequired[List[str]]
|
|
14
|
+
streamtags: NotRequired[List[str]]
|
|
15
15
|
|
|
16
16
|
|
|
17
17
|
class PackInfoTags(BaseModel):
|
|
18
18
|
data_type: Annotated[List[str], pydantic.Field(alias="dataType")]
|
|
19
19
|
|
|
20
|
-
|
|
20
|
+
technology: List[str]
|
|
21
21
|
|
|
22
|
-
|
|
22
|
+
domain: Optional[List[str]] = None
|
|
23
23
|
|
|
24
|
-
|
|
24
|
+
streamtags: Optional[List[str]] = None
|
|
25
25
|
|
|
26
26
|
|
|
27
27
|
class PackInfoTypedDict(TypedDict):
|
|
@@ -9,19 +9,19 @@ from typing_extensions import Annotated, NotRequired, TypedDict
|
|
|
9
9
|
|
|
10
10
|
class PackInstallInfoTagsTypedDict(TypedDict):
|
|
11
11
|
data_type: List[str]
|
|
12
|
-
domain: List[str]
|
|
13
|
-
streamtags: List[str]
|
|
14
12
|
technology: List[str]
|
|
13
|
+
domain: NotRequired[List[str]]
|
|
14
|
+
streamtags: NotRequired[List[str]]
|
|
15
15
|
|
|
16
16
|
|
|
17
17
|
class PackInstallInfoTags(BaseModel):
|
|
18
18
|
data_type: Annotated[List[str], pydantic.Field(alias="dataType")]
|
|
19
19
|
|
|
20
|
-
|
|
20
|
+
technology: List[str]
|
|
21
21
|
|
|
22
|
-
|
|
22
|
+
domain: Optional[List[str]] = None
|
|
23
23
|
|
|
24
|
-
|
|
24
|
+
streamtags: Optional[List[str]] = None
|
|
25
25
|
|
|
26
26
|
|
|
27
27
|
class PackInstallInfoTypedDict(TypedDict):
|