robosystems-client 0.2.5__py3-none-any.whl → 0.2.7__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of robosystems-client might be problematic. Click here for more details.
- robosystems_client/api/agent/auto_select_agent.py +164 -32
- robosystems_client/api/backup/create_backup.py +72 -0
- robosystems_client/api/backup/get_backup_download_url.py +12 -28
- robosystems_client/api/backup/restore_backup.py +92 -0
- robosystems_client/api/graph_limits/get_graph_limits.py +12 -14
- robosystems_client/api/graphs/create_graph.py +136 -36
- robosystems_client/api/graphs/get_available_graph_tiers.py +281 -0
- robosystems_client/api/query/execute_cypher_query.py +13 -11
- robosystems_client/api/service_offerings/get_service_offerings.py +13 -11
- robosystems_client/models/__init__.py +66 -8
- robosystems_client/models/agent_response.py +1 -1
- robosystems_client/models/available_graph_tiers_response.py +74 -0
- robosystems_client/models/backup_download_url_response.py +92 -0
- robosystems_client/models/backup_limits.py +76 -0
- robosystems_client/models/batch_agent_request.py +1 -1
- robosystems_client/models/batch_agent_response.py +2 -2
- robosystems_client/models/copy_operation_limits.py +100 -0
- robosystems_client/models/create_graph_request.py +16 -17
- robosystems_client/models/credit_limits.py +84 -0
- robosystems_client/models/custom_schema_definition.py +14 -10
- robosystems_client/models/execute_cypher_query_response_200.py +135 -0
- robosystems_client/models/{get_graph_limits_response_getgraphlimits.py → execute_cypher_query_response_200_data_item.py} +5 -5
- robosystems_client/models/graph_limits_response.py +174 -0
- robosystems_client/models/graph_subscription_tier.py +220 -0
- robosystems_client/models/graph_subscriptions.py +100 -0
- robosystems_client/models/graph_tier_backup.py +76 -0
- robosystems_client/models/graph_tier_copy_operations.py +92 -0
- robosystems_client/models/graph_tier_info.py +192 -0
- robosystems_client/models/graph_tier_instance.py +76 -0
- robosystems_client/models/graph_tier_limits.py +106 -0
- robosystems_client/models/initial_entity_data.py +15 -12
- robosystems_client/models/offering_repository_plan.py +148 -0
- robosystems_client/models/offering_repository_plan_rate_limits_type_0.py +61 -0
- robosystems_client/models/operation_costs.py +100 -0
- robosystems_client/models/operation_costs_ai_operations.py +44 -0
- robosystems_client/models/operation_costs_token_pricing.py +59 -0
- robosystems_client/models/query_limits.py +84 -0
- robosystems_client/models/rate_limits.py +76 -0
- robosystems_client/models/repository_info.py +114 -0
- robosystems_client/models/repository_subscriptions.py +90 -0
- robosystems_client/models/service_offering_summary.py +84 -0
- robosystems_client/models/service_offerings_response.py +98 -0
- robosystems_client/models/storage_info.py +76 -0
- robosystems_client/models/{get_backup_download_url_response_getbackupdownloadurl.py → storage_info_included_per_tier.py} +9 -9
- robosystems_client/models/storage_info_overage_pricing.py +44 -0
- robosystems_client/models/storage_limits.py +90 -0
- robosystems_client/models/token_pricing.py +68 -0
- {robosystems_client-0.2.5.dist-info → robosystems_client-0.2.7.dist-info}/METADATA +1 -1
- {robosystems_client-0.2.5.dist-info → robosystems_client-0.2.7.dist-info}/RECORD +51 -21
- {robosystems_client-0.2.5.dist-info → robosystems_client-0.2.7.dist-info}/WHEEL +0 -0
- {robosystems_client-0.2.5.dist-info → robosystems_client-0.2.7.dist-info}/licenses/LICENSE +0 -0
|
@@ -0,0 +1,76 @@
|
|
|
1
|
+
from collections.abc import Mapping
|
|
2
|
+
from typing import Any, TypeVar
|
|
3
|
+
|
|
4
|
+
from attrs import define as _attrs_define
|
|
5
|
+
from attrs import field as _attrs_field
|
|
6
|
+
|
|
7
|
+
T = TypeVar("T", bound="BackupLimits")
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
@_attrs_define
|
|
11
|
+
class BackupLimits:
|
|
12
|
+
"""Backup operation limits.
|
|
13
|
+
|
|
14
|
+
Attributes:
|
|
15
|
+
max_backup_size_gb (float): Maximum backup size in GB
|
|
16
|
+
backup_retention_days (int): Backup retention period in days
|
|
17
|
+
max_backups_per_day (int): Maximum backups per day
|
|
18
|
+
"""
|
|
19
|
+
|
|
20
|
+
max_backup_size_gb: float
|
|
21
|
+
backup_retention_days: int
|
|
22
|
+
max_backups_per_day: int
|
|
23
|
+
additional_properties: dict[str, Any] = _attrs_field(init=False, factory=dict)
|
|
24
|
+
|
|
25
|
+
def to_dict(self) -> dict[str, Any]:
|
|
26
|
+
max_backup_size_gb = self.max_backup_size_gb
|
|
27
|
+
|
|
28
|
+
backup_retention_days = self.backup_retention_days
|
|
29
|
+
|
|
30
|
+
max_backups_per_day = self.max_backups_per_day
|
|
31
|
+
|
|
32
|
+
field_dict: dict[str, Any] = {}
|
|
33
|
+
field_dict.update(self.additional_properties)
|
|
34
|
+
field_dict.update(
|
|
35
|
+
{
|
|
36
|
+
"max_backup_size_gb": max_backup_size_gb,
|
|
37
|
+
"backup_retention_days": backup_retention_days,
|
|
38
|
+
"max_backups_per_day": max_backups_per_day,
|
|
39
|
+
}
|
|
40
|
+
)
|
|
41
|
+
|
|
42
|
+
return field_dict
|
|
43
|
+
|
|
44
|
+
@classmethod
|
|
45
|
+
def from_dict(cls: type[T], src_dict: Mapping[str, Any]) -> T:
|
|
46
|
+
d = dict(src_dict)
|
|
47
|
+
max_backup_size_gb = d.pop("max_backup_size_gb")
|
|
48
|
+
|
|
49
|
+
backup_retention_days = d.pop("backup_retention_days")
|
|
50
|
+
|
|
51
|
+
max_backups_per_day = d.pop("max_backups_per_day")
|
|
52
|
+
|
|
53
|
+
backup_limits = cls(
|
|
54
|
+
max_backup_size_gb=max_backup_size_gb,
|
|
55
|
+
backup_retention_days=backup_retention_days,
|
|
56
|
+
max_backups_per_day=max_backups_per_day,
|
|
57
|
+
)
|
|
58
|
+
|
|
59
|
+
backup_limits.additional_properties = d
|
|
60
|
+
return backup_limits
|
|
61
|
+
|
|
62
|
+
@property
|
|
63
|
+
def additional_keys(self) -> list[str]:
|
|
64
|
+
return list(self.additional_properties.keys())
|
|
65
|
+
|
|
66
|
+
def __getitem__(self, key: str) -> Any:
|
|
67
|
+
return self.additional_properties[key]
|
|
68
|
+
|
|
69
|
+
def __setitem__(self, key: str, value: Any) -> None:
|
|
70
|
+
self.additional_properties[key] = value
|
|
71
|
+
|
|
72
|
+
def __delitem__(self, key: str) -> None:
|
|
73
|
+
del self.additional_properties[key]
|
|
74
|
+
|
|
75
|
+
def __contains__(self, key: str) -> bool:
|
|
76
|
+
return key in self.additional_properties
|
|
@@ -18,7 +18,7 @@ class BatchAgentRequest:
|
|
|
18
18
|
"""Request for batch processing multiple queries.
|
|
19
19
|
|
|
20
20
|
Attributes:
|
|
21
|
-
queries (list['AgentRequest']): List of queries to process
|
|
21
|
+
queries (list['AgentRequest']): List of queries to process (max 10)
|
|
22
22
|
parallel (Union[Unset, bool]): Process queries in parallel Default: False.
|
|
23
23
|
"""
|
|
24
24
|
|
|
@@ -16,8 +16,8 @@ class BatchAgentResponse:
|
|
|
16
16
|
"""Response for batch processing.
|
|
17
17
|
|
|
18
18
|
Attributes:
|
|
19
|
-
results (list['AgentResponse']): List of agent responses
|
|
20
|
-
total_execution_time (float): Total execution time
|
|
19
|
+
results (list['AgentResponse']): List of agent responses (includes successes and failures)
|
|
20
|
+
total_execution_time (float): Total execution time in seconds
|
|
21
21
|
parallel_processed (bool): Whether queries were processed in parallel
|
|
22
22
|
"""
|
|
23
23
|
|
|
@@ -0,0 +1,100 @@
|
|
|
1
|
+
from collections.abc import Mapping
|
|
2
|
+
from typing import Any, TypeVar, cast
|
|
3
|
+
|
|
4
|
+
from attrs import define as _attrs_define
|
|
5
|
+
from attrs import field as _attrs_field
|
|
6
|
+
|
|
7
|
+
T = TypeVar("T", bound="CopyOperationLimits")
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
@_attrs_define
|
|
11
|
+
class CopyOperationLimits:
|
|
12
|
+
"""Copy/ingestion operation limits.
|
|
13
|
+
|
|
14
|
+
Attributes:
|
|
15
|
+
max_file_size_gb (float): Maximum file size in GB
|
|
16
|
+
timeout_seconds (int): Operation timeout in seconds
|
|
17
|
+
concurrent_operations (int): Maximum concurrent operations
|
|
18
|
+
max_files_per_operation (int): Maximum files per operation
|
|
19
|
+
daily_copy_operations (int): Daily operation limit
|
|
20
|
+
supported_formats (list[str]): Supported file formats
|
|
21
|
+
"""
|
|
22
|
+
|
|
23
|
+
max_file_size_gb: float
|
|
24
|
+
timeout_seconds: int
|
|
25
|
+
concurrent_operations: int
|
|
26
|
+
max_files_per_operation: int
|
|
27
|
+
daily_copy_operations: int
|
|
28
|
+
supported_formats: list[str]
|
|
29
|
+
additional_properties: dict[str, Any] = _attrs_field(init=False, factory=dict)
|
|
30
|
+
|
|
31
|
+
def to_dict(self) -> dict[str, Any]:
|
|
32
|
+
max_file_size_gb = self.max_file_size_gb
|
|
33
|
+
|
|
34
|
+
timeout_seconds = self.timeout_seconds
|
|
35
|
+
|
|
36
|
+
concurrent_operations = self.concurrent_operations
|
|
37
|
+
|
|
38
|
+
max_files_per_operation = self.max_files_per_operation
|
|
39
|
+
|
|
40
|
+
daily_copy_operations = self.daily_copy_operations
|
|
41
|
+
|
|
42
|
+
supported_formats = self.supported_formats
|
|
43
|
+
|
|
44
|
+
field_dict: dict[str, Any] = {}
|
|
45
|
+
field_dict.update(self.additional_properties)
|
|
46
|
+
field_dict.update(
|
|
47
|
+
{
|
|
48
|
+
"max_file_size_gb": max_file_size_gb,
|
|
49
|
+
"timeout_seconds": timeout_seconds,
|
|
50
|
+
"concurrent_operations": concurrent_operations,
|
|
51
|
+
"max_files_per_operation": max_files_per_operation,
|
|
52
|
+
"daily_copy_operations": daily_copy_operations,
|
|
53
|
+
"supported_formats": supported_formats,
|
|
54
|
+
}
|
|
55
|
+
)
|
|
56
|
+
|
|
57
|
+
return field_dict
|
|
58
|
+
|
|
59
|
+
@classmethod
|
|
60
|
+
def from_dict(cls: type[T], src_dict: Mapping[str, Any]) -> T:
|
|
61
|
+
d = dict(src_dict)
|
|
62
|
+
max_file_size_gb = d.pop("max_file_size_gb")
|
|
63
|
+
|
|
64
|
+
timeout_seconds = d.pop("timeout_seconds")
|
|
65
|
+
|
|
66
|
+
concurrent_operations = d.pop("concurrent_operations")
|
|
67
|
+
|
|
68
|
+
max_files_per_operation = d.pop("max_files_per_operation")
|
|
69
|
+
|
|
70
|
+
daily_copy_operations = d.pop("daily_copy_operations")
|
|
71
|
+
|
|
72
|
+
supported_formats = cast(list[str], d.pop("supported_formats"))
|
|
73
|
+
|
|
74
|
+
copy_operation_limits = cls(
|
|
75
|
+
max_file_size_gb=max_file_size_gb,
|
|
76
|
+
timeout_seconds=timeout_seconds,
|
|
77
|
+
concurrent_operations=concurrent_operations,
|
|
78
|
+
max_files_per_operation=max_files_per_operation,
|
|
79
|
+
daily_copy_operations=daily_copy_operations,
|
|
80
|
+
supported_formats=supported_formats,
|
|
81
|
+
)
|
|
82
|
+
|
|
83
|
+
copy_operation_limits.additional_properties = d
|
|
84
|
+
return copy_operation_limits
|
|
85
|
+
|
|
86
|
+
@property
|
|
87
|
+
def additional_keys(self) -> list[str]:
|
|
88
|
+
return list(self.additional_properties.keys())
|
|
89
|
+
|
|
90
|
+
def __getitem__(self, key: str) -> Any:
|
|
91
|
+
return self.additional_properties[key]
|
|
92
|
+
|
|
93
|
+
def __setitem__(self, key: str, value: Any) -> None:
|
|
94
|
+
self.additional_properties[key] = value
|
|
95
|
+
|
|
96
|
+
def __delitem__(self, key: str) -> None:
|
|
97
|
+
del self.additional_properties[key]
|
|
98
|
+
|
|
99
|
+
def __contains__(self, key: str) -> bool:
|
|
100
|
+
return key in self.additional_properties
|
|
@@ -19,23 +19,22 @@ T = TypeVar("T", bound="CreateGraphRequest")
|
|
|
19
19
|
class CreateGraphRequest:
|
|
20
20
|
"""Request model for creating a new graph.
|
|
21
21
|
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
tags (Union[Unset, list[str]]): Optional tags for organization
|
|
22
|
+
Use this to create either:
|
|
23
|
+
- **Entity graphs**: Standard graphs with entity schema and optional extensions
|
|
24
|
+
- **Custom graphs**: Generic graphs with fully custom schema definitions
|
|
25
|
+
|
|
26
|
+
Attributes:
|
|
27
|
+
metadata (GraphMetadata): Metadata for graph creation.
|
|
28
|
+
instance_tier (Union[Unset, str]): Instance tier: kuzu-standard, kuzu-large, kuzu-xlarge, neo4j-community-large,
|
|
29
|
+
neo4j-enterprise-xlarge Default: 'kuzu-standard'.
|
|
30
|
+
custom_schema (Union['CustomSchemaDefinition', None, Unset]): Custom schema definition to apply. If provided,
|
|
31
|
+
creates a generic custom graph. If omitted, creates an entity graph using schema_extensions.
|
|
32
|
+
initial_entity (Union['InitialEntityData', None, Unset]): Optional initial entity to create in the graph. If
|
|
33
|
+
provided with entity graph, populates the first entity node.
|
|
34
|
+
create_entity (Union[Unset, bool]): Whether to create the entity node and upload initial data. Only applies when
|
|
35
|
+
initial_entity is provided. Set to False to create graph without populating entity data (useful for file-based
|
|
36
|
+
ingestion workflows). Default: True.
|
|
37
|
+
tags (Union[Unset, list[str]]): Optional tags for organization
|
|
39
38
|
"""
|
|
40
39
|
|
|
41
40
|
metadata: "GraphMetadata"
|
|
@@ -0,0 +1,84 @@
|
|
|
1
|
+
from collections.abc import Mapping
|
|
2
|
+
from typing import Any, TypeVar
|
|
3
|
+
|
|
4
|
+
from attrs import define as _attrs_define
|
|
5
|
+
from attrs import field as _attrs_field
|
|
6
|
+
|
|
7
|
+
T = TypeVar("T", bound="CreditLimits")
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
@_attrs_define
|
|
11
|
+
class CreditLimits:
|
|
12
|
+
"""AI credit limits (optional).
|
|
13
|
+
|
|
14
|
+
Attributes:
|
|
15
|
+
monthly_ai_credits (int): Monthly AI credits allocation
|
|
16
|
+
current_balance (int): Current credit balance
|
|
17
|
+
storage_billing_enabled (bool): Whether storage billing is enabled
|
|
18
|
+
storage_rate_per_gb_per_day (int): Storage billing rate per GB per day
|
|
19
|
+
"""
|
|
20
|
+
|
|
21
|
+
monthly_ai_credits: int
|
|
22
|
+
current_balance: int
|
|
23
|
+
storage_billing_enabled: bool
|
|
24
|
+
storage_rate_per_gb_per_day: int
|
|
25
|
+
additional_properties: dict[str, Any] = _attrs_field(init=False, factory=dict)
|
|
26
|
+
|
|
27
|
+
def to_dict(self) -> dict[str, Any]:
|
|
28
|
+
monthly_ai_credits = self.monthly_ai_credits
|
|
29
|
+
|
|
30
|
+
current_balance = self.current_balance
|
|
31
|
+
|
|
32
|
+
storage_billing_enabled = self.storage_billing_enabled
|
|
33
|
+
|
|
34
|
+
storage_rate_per_gb_per_day = self.storage_rate_per_gb_per_day
|
|
35
|
+
|
|
36
|
+
field_dict: dict[str, Any] = {}
|
|
37
|
+
field_dict.update(self.additional_properties)
|
|
38
|
+
field_dict.update(
|
|
39
|
+
{
|
|
40
|
+
"monthly_ai_credits": monthly_ai_credits,
|
|
41
|
+
"current_balance": current_balance,
|
|
42
|
+
"storage_billing_enabled": storage_billing_enabled,
|
|
43
|
+
"storage_rate_per_gb_per_day": storage_rate_per_gb_per_day,
|
|
44
|
+
}
|
|
45
|
+
)
|
|
46
|
+
|
|
47
|
+
return field_dict
|
|
48
|
+
|
|
49
|
+
@classmethod
|
|
50
|
+
def from_dict(cls: type[T], src_dict: Mapping[str, Any]) -> T:
|
|
51
|
+
d = dict(src_dict)
|
|
52
|
+
monthly_ai_credits = d.pop("monthly_ai_credits")
|
|
53
|
+
|
|
54
|
+
current_balance = d.pop("current_balance")
|
|
55
|
+
|
|
56
|
+
storage_billing_enabled = d.pop("storage_billing_enabled")
|
|
57
|
+
|
|
58
|
+
storage_rate_per_gb_per_day = d.pop("storage_rate_per_gb_per_day")
|
|
59
|
+
|
|
60
|
+
credit_limits = cls(
|
|
61
|
+
monthly_ai_credits=monthly_ai_credits,
|
|
62
|
+
current_balance=current_balance,
|
|
63
|
+
storage_billing_enabled=storage_billing_enabled,
|
|
64
|
+
storage_rate_per_gb_per_day=storage_rate_per_gb_per_day,
|
|
65
|
+
)
|
|
66
|
+
|
|
67
|
+
credit_limits.additional_properties = d
|
|
68
|
+
return credit_limits
|
|
69
|
+
|
|
70
|
+
@property
|
|
71
|
+
def additional_keys(self) -> list[str]:
|
|
72
|
+
return list(self.additional_properties.keys())
|
|
73
|
+
|
|
74
|
+
def __getitem__(self, key: str) -> Any:
|
|
75
|
+
return self.additional_properties[key]
|
|
76
|
+
|
|
77
|
+
def __setitem__(self, key: str, value: Any) -> None:
|
|
78
|
+
self.additional_properties[key] = value
|
|
79
|
+
|
|
80
|
+
def __delitem__(self, key: str) -> None:
|
|
81
|
+
del self.additional_properties[key]
|
|
82
|
+
|
|
83
|
+
def __contains__(self, key: str) -> bool:
|
|
84
|
+
return key in self.additional_properties
|
|
@@ -21,16 +21,20 @@ T = TypeVar("T", bound="CustomSchemaDefinition")
|
|
|
21
21
|
|
|
22
22
|
@_attrs_define
|
|
23
23
|
class CustomSchemaDefinition:
|
|
24
|
-
"""Custom schema definition for
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
24
|
+
"""Custom schema definition for generic graphs.
|
|
25
|
+
|
|
26
|
+
This model allows you to define custom node types, relationship types, and properties
|
|
27
|
+
for graphs that don't fit the standard entity-based schema. Perfect for domain-specific
|
|
28
|
+
applications like inventory systems, org charts, project management, etc.
|
|
29
|
+
|
|
30
|
+
Attributes:
|
|
31
|
+
name (str): Schema name
|
|
32
|
+
version (Union[Unset, str]): Schema version Default: '1.0.0'.
|
|
33
|
+
description (Union[None, Unset, str]): Schema description
|
|
34
|
+
extends (Union[None, Unset, str]): Base schema to extend (e.g., 'base' for common utilities)
|
|
35
|
+
nodes (Union[Unset, list['CustomSchemaDefinitionNodesItem']]): List of node definitions with properties
|
|
36
|
+
relationships (Union[Unset, list['CustomSchemaDefinitionRelationshipsItem']]): List of relationship definitions
|
|
37
|
+
metadata (Union[Unset, CustomSchemaDefinitionMetadata]): Additional schema metadata
|
|
34
38
|
"""
|
|
35
39
|
|
|
36
40
|
name: str
|
|
@@ -0,0 +1,135 @@
|
|
|
1
|
+
from collections.abc import Mapping
|
|
2
|
+
from typing import TYPE_CHECKING, Any, TypeVar, Union, cast
|
|
3
|
+
|
|
4
|
+
from attrs import define as _attrs_define
|
|
5
|
+
from attrs import field as _attrs_field
|
|
6
|
+
|
|
7
|
+
from ..types import UNSET, Unset
|
|
8
|
+
|
|
9
|
+
if TYPE_CHECKING:
|
|
10
|
+
from ..models.execute_cypher_query_response_200_data_item import (
|
|
11
|
+
ExecuteCypherQueryResponse200DataItem,
|
|
12
|
+
)
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
T = TypeVar("T", bound="ExecuteCypherQueryResponse200")
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
@_attrs_define
|
|
19
|
+
class ExecuteCypherQueryResponse200:
|
|
20
|
+
"""
|
|
21
|
+
Attributes:
|
|
22
|
+
success (Union[Unset, bool]):
|
|
23
|
+
data (Union[Unset, list['ExecuteCypherQueryResponse200DataItem']]):
|
|
24
|
+
columns (Union[Unset, list[str]]):
|
|
25
|
+
row_count (Union[Unset, int]):
|
|
26
|
+
execution_time_ms (Union[Unset, float]):
|
|
27
|
+
graph_id (Union[Unset, str]):
|
|
28
|
+
timestamp (Union[Unset, str]):
|
|
29
|
+
"""
|
|
30
|
+
|
|
31
|
+
success: Union[Unset, bool] = UNSET
|
|
32
|
+
data: Union[Unset, list["ExecuteCypherQueryResponse200DataItem"]] = UNSET
|
|
33
|
+
columns: Union[Unset, list[str]] = UNSET
|
|
34
|
+
row_count: Union[Unset, int] = UNSET
|
|
35
|
+
execution_time_ms: Union[Unset, float] = UNSET
|
|
36
|
+
graph_id: Union[Unset, str] = UNSET
|
|
37
|
+
timestamp: Union[Unset, str] = UNSET
|
|
38
|
+
additional_properties: dict[str, Any] = _attrs_field(init=False, factory=dict)
|
|
39
|
+
|
|
40
|
+
def to_dict(self) -> dict[str, Any]:
|
|
41
|
+
success = self.success
|
|
42
|
+
|
|
43
|
+
data: Union[Unset, list[dict[str, Any]]] = UNSET
|
|
44
|
+
if not isinstance(self.data, Unset):
|
|
45
|
+
data = []
|
|
46
|
+
for data_item_data in self.data:
|
|
47
|
+
data_item = data_item_data.to_dict()
|
|
48
|
+
data.append(data_item)
|
|
49
|
+
|
|
50
|
+
columns: Union[Unset, list[str]] = UNSET
|
|
51
|
+
if not isinstance(self.columns, Unset):
|
|
52
|
+
columns = self.columns
|
|
53
|
+
|
|
54
|
+
row_count = self.row_count
|
|
55
|
+
|
|
56
|
+
execution_time_ms = self.execution_time_ms
|
|
57
|
+
|
|
58
|
+
graph_id = self.graph_id
|
|
59
|
+
|
|
60
|
+
timestamp = self.timestamp
|
|
61
|
+
|
|
62
|
+
field_dict: dict[str, Any] = {}
|
|
63
|
+
field_dict.update(self.additional_properties)
|
|
64
|
+
field_dict.update({})
|
|
65
|
+
if success is not UNSET:
|
|
66
|
+
field_dict["success"] = success
|
|
67
|
+
if data is not UNSET:
|
|
68
|
+
field_dict["data"] = data
|
|
69
|
+
if columns is not UNSET:
|
|
70
|
+
field_dict["columns"] = columns
|
|
71
|
+
if row_count is not UNSET:
|
|
72
|
+
field_dict["row_count"] = row_count
|
|
73
|
+
if execution_time_ms is not UNSET:
|
|
74
|
+
field_dict["execution_time_ms"] = execution_time_ms
|
|
75
|
+
if graph_id is not UNSET:
|
|
76
|
+
field_dict["graph_id"] = graph_id
|
|
77
|
+
if timestamp is not UNSET:
|
|
78
|
+
field_dict["timestamp"] = timestamp
|
|
79
|
+
|
|
80
|
+
return field_dict
|
|
81
|
+
|
|
82
|
+
@classmethod
|
|
83
|
+
def from_dict(cls: type[T], src_dict: Mapping[str, Any]) -> T:
|
|
84
|
+
from ..models.execute_cypher_query_response_200_data_item import (
|
|
85
|
+
ExecuteCypherQueryResponse200DataItem,
|
|
86
|
+
)
|
|
87
|
+
|
|
88
|
+
d = dict(src_dict)
|
|
89
|
+
success = d.pop("success", UNSET)
|
|
90
|
+
|
|
91
|
+
data = []
|
|
92
|
+
_data = d.pop("data", UNSET)
|
|
93
|
+
for data_item_data in _data or []:
|
|
94
|
+
data_item = ExecuteCypherQueryResponse200DataItem.from_dict(data_item_data)
|
|
95
|
+
|
|
96
|
+
data.append(data_item)
|
|
97
|
+
|
|
98
|
+
columns = cast(list[str], d.pop("columns", UNSET))
|
|
99
|
+
|
|
100
|
+
row_count = d.pop("row_count", UNSET)
|
|
101
|
+
|
|
102
|
+
execution_time_ms = d.pop("execution_time_ms", UNSET)
|
|
103
|
+
|
|
104
|
+
graph_id = d.pop("graph_id", UNSET)
|
|
105
|
+
|
|
106
|
+
timestamp = d.pop("timestamp", UNSET)
|
|
107
|
+
|
|
108
|
+
execute_cypher_query_response_200 = cls(
|
|
109
|
+
success=success,
|
|
110
|
+
data=data,
|
|
111
|
+
columns=columns,
|
|
112
|
+
row_count=row_count,
|
|
113
|
+
execution_time_ms=execution_time_ms,
|
|
114
|
+
graph_id=graph_id,
|
|
115
|
+
timestamp=timestamp,
|
|
116
|
+
)
|
|
117
|
+
|
|
118
|
+
execute_cypher_query_response_200.additional_properties = d
|
|
119
|
+
return execute_cypher_query_response_200
|
|
120
|
+
|
|
121
|
+
@property
|
|
122
|
+
def additional_keys(self) -> list[str]:
|
|
123
|
+
return list(self.additional_properties.keys())
|
|
124
|
+
|
|
125
|
+
def __getitem__(self, key: str) -> Any:
|
|
126
|
+
return self.additional_properties[key]
|
|
127
|
+
|
|
128
|
+
def __setitem__(self, key: str, value: Any) -> None:
|
|
129
|
+
self.additional_properties[key] = value
|
|
130
|
+
|
|
131
|
+
def __delitem__(self, key: str) -> None:
|
|
132
|
+
del self.additional_properties[key]
|
|
133
|
+
|
|
134
|
+
def __contains__(self, key: str) -> bool:
|
|
135
|
+
return key in self.additional_properties
|
|
@@ -4,11 +4,11 @@ from typing import Any, TypeVar
|
|
|
4
4
|
from attrs import define as _attrs_define
|
|
5
5
|
from attrs import field as _attrs_field
|
|
6
6
|
|
|
7
|
-
T = TypeVar("T", bound="
|
|
7
|
+
T = TypeVar("T", bound="ExecuteCypherQueryResponse200DataItem")
|
|
8
8
|
|
|
9
9
|
|
|
10
10
|
@_attrs_define
|
|
11
|
-
class
|
|
11
|
+
class ExecuteCypherQueryResponse200DataItem:
|
|
12
12
|
""" """
|
|
13
13
|
|
|
14
14
|
additional_properties: dict[str, Any] = _attrs_field(init=False, factory=dict)
|
|
@@ -22,10 +22,10 @@ class GetGraphLimitsResponseGetgraphlimits:
|
|
|
22
22
|
@classmethod
|
|
23
23
|
def from_dict(cls: type[T], src_dict: Mapping[str, Any]) -> T:
|
|
24
24
|
d = dict(src_dict)
|
|
25
|
-
|
|
25
|
+
execute_cypher_query_response_200_data_item = cls()
|
|
26
26
|
|
|
27
|
-
|
|
28
|
-
return
|
|
27
|
+
execute_cypher_query_response_200_data_item.additional_properties = d
|
|
28
|
+
return execute_cypher_query_response_200_data_item
|
|
29
29
|
|
|
30
30
|
@property
|
|
31
31
|
def additional_keys(self) -> list[str]:
|