evermemos 0.3.8__py3-none-any.whl → 0.3.10__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- evermemos/_base_client.py +5 -2
- evermemos/_compat.py +3 -3
- evermemos/_utils/_json.py +35 -0
- evermemos/_version.py +1 -1
- evermemos/resources/v1/__init__.py +13 -13
- evermemos/resources/v1/memories/conversation_meta.py +82 -40
- evermemos/resources/v1/memories/memories.py +133 -228
- evermemos/resources/v1/{stats → status}/__init__.py +13 -13
- evermemos/resources/v1/{stats → status}/request.py +6 -6
- evermemos/resources/v1/{stats/stats.py → status/status.py} +27 -27
- evermemos/resources/v1/v1.py +19 -19
- evermemos/types/v1/__init__.py +2 -4
- evermemos/types/v1/memories/conversation_meta_create_params.py +90 -19
- evermemos/types/v1/memories/conversation_meta_create_response.py +60 -17
- evermemos/types/v1/memories/conversation_meta_get_response.py +60 -17
- evermemos/types/v1/memories/conversation_meta_update_params.py +68 -9
- evermemos/types/v1/memories/conversation_meta_update_response.py +0 -5
- evermemos/types/v1/{memory_create_params.py → memory_add_params.py} +2 -2
- evermemos/types/v1/{memory_create_response.py → memory_add_response.py} +2 -2
- evermemos/types/v1/memory_delete_params.py +6 -0
- evermemos/types/v1/memory_delete_response.py +0 -5
- evermemos/types/v1/memory_get_response.py +7 -10
- evermemos/types/v1/memory_search_response.py +5 -18
- evermemos/types/v1/{stats → status}/request_get_params.py +1 -1
- evermemos/types/v1/{stats → status}/request_get_response.py +0 -5
- {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/METADATA +24 -27
- {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/RECORD +30 -32
- evermemos/types/v1/global_user_profile/__init__.py +0 -3
- evermemos/types/v1/memory_load_params.py +0 -83
- evermemos/types/v1/memory_load_response.py +0 -26
- /evermemos/types/v1/{stats → status}/__init__.py +0 -0
- {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/WHEEL +0 -0
- {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/licenses/LICENSE +0 -0
evermemos/resources/v1/v1.py
CHANGED
|
@@ -4,13 +4,13 @@ from __future__ import annotations
|
|
|
4
4
|
|
|
5
5
|
from ..._compat import cached_property
|
|
6
6
|
from ..._resource import SyncAPIResource, AsyncAPIResource
|
|
7
|
-
from .
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
7
|
+
from .status.status import (
|
|
8
|
+
StatusResource,
|
|
9
|
+
AsyncStatusResource,
|
|
10
|
+
StatusResourceWithRawResponse,
|
|
11
|
+
AsyncStatusResourceWithRawResponse,
|
|
12
|
+
StatusResourceWithStreamingResponse,
|
|
13
|
+
AsyncStatusResourceWithStreamingResponse,
|
|
14
14
|
)
|
|
15
15
|
from .memories.memories import (
|
|
16
16
|
MemoriesResource,
|
|
@@ -30,8 +30,8 @@ class V1Resource(SyncAPIResource):
|
|
|
30
30
|
return MemoriesResource(self._client)
|
|
31
31
|
|
|
32
32
|
@cached_property
|
|
33
|
-
def
|
|
34
|
-
return
|
|
33
|
+
def status(self) -> StatusResource:
|
|
34
|
+
return StatusResource(self._client)
|
|
35
35
|
|
|
36
36
|
@cached_property
|
|
37
37
|
def with_raw_response(self) -> V1ResourceWithRawResponse:
|
|
@@ -59,8 +59,8 @@ class AsyncV1Resource(AsyncAPIResource):
|
|
|
59
59
|
return AsyncMemoriesResource(self._client)
|
|
60
60
|
|
|
61
61
|
@cached_property
|
|
62
|
-
def
|
|
63
|
-
return
|
|
62
|
+
def status(self) -> AsyncStatusResource:
|
|
63
|
+
return AsyncStatusResource(self._client)
|
|
64
64
|
|
|
65
65
|
@cached_property
|
|
66
66
|
def with_raw_response(self) -> AsyncV1ResourceWithRawResponse:
|
|
@@ -91,8 +91,8 @@ class V1ResourceWithRawResponse:
|
|
|
91
91
|
return MemoriesResourceWithRawResponse(self._v1.memories)
|
|
92
92
|
|
|
93
93
|
@cached_property
|
|
94
|
-
def
|
|
95
|
-
return
|
|
94
|
+
def status(self) -> StatusResourceWithRawResponse:
|
|
95
|
+
return StatusResourceWithRawResponse(self._v1.status)
|
|
96
96
|
|
|
97
97
|
|
|
98
98
|
class AsyncV1ResourceWithRawResponse:
|
|
@@ -104,8 +104,8 @@ class AsyncV1ResourceWithRawResponse:
|
|
|
104
104
|
return AsyncMemoriesResourceWithRawResponse(self._v1.memories)
|
|
105
105
|
|
|
106
106
|
@cached_property
|
|
107
|
-
def
|
|
108
|
-
return
|
|
107
|
+
def status(self) -> AsyncStatusResourceWithRawResponse:
|
|
108
|
+
return AsyncStatusResourceWithRawResponse(self._v1.status)
|
|
109
109
|
|
|
110
110
|
|
|
111
111
|
class V1ResourceWithStreamingResponse:
|
|
@@ -117,8 +117,8 @@ class V1ResourceWithStreamingResponse:
|
|
|
117
117
|
return MemoriesResourceWithStreamingResponse(self._v1.memories)
|
|
118
118
|
|
|
119
119
|
@cached_property
|
|
120
|
-
def
|
|
121
|
-
return
|
|
120
|
+
def status(self) -> StatusResourceWithStreamingResponse:
|
|
121
|
+
return StatusResourceWithStreamingResponse(self._v1.status)
|
|
122
122
|
|
|
123
123
|
|
|
124
124
|
class AsyncV1ResourceWithStreamingResponse:
|
|
@@ -130,5 +130,5 @@ class AsyncV1ResourceWithStreamingResponse:
|
|
|
130
130
|
return AsyncMemoriesResourceWithStreamingResponse(self._v1.memories)
|
|
131
131
|
|
|
132
132
|
@cached_property
|
|
133
|
-
def
|
|
134
|
-
return
|
|
133
|
+
def status(self) -> AsyncStatusResourceWithStreamingResponse:
|
|
134
|
+
return AsyncStatusResourceWithStreamingResponse(self._v1.status)
|
evermemos/types/v1/__init__.py
CHANGED
|
@@ -4,11 +4,9 @@ from __future__ import annotations
|
|
|
4
4
|
|
|
5
5
|
from .metadata import Metadata as Metadata
|
|
6
6
|
from .memory_type import MemoryType as MemoryType
|
|
7
|
-
from .
|
|
7
|
+
from .memory_add_params import MemoryAddParams as MemoryAddParams
|
|
8
|
+
from .memory_add_response import MemoryAddResponse as MemoryAddResponse
|
|
8
9
|
from .memory_get_response import MemoryGetResponse as MemoryGetResponse
|
|
9
|
-
from .memory_create_params import MemoryCreateParams as MemoryCreateParams
|
|
10
10
|
from .memory_delete_params import MemoryDeleteParams as MemoryDeleteParams
|
|
11
|
-
from .memory_load_response import MemoryLoadResponse as MemoryLoadResponse
|
|
12
|
-
from .memory_create_response import MemoryCreateResponse as MemoryCreateResponse
|
|
13
11
|
from .memory_delete_response import MemoryDeleteResponse as MemoryDeleteResponse
|
|
14
12
|
from .memory_search_response import MemorySearchResponse as MemorySearchResponse
|
|
@@ -7,28 +7,19 @@ from typing_extensions import Required, TypedDict
|
|
|
7
7
|
|
|
8
8
|
from ...._types import SequenceNotStr
|
|
9
9
|
|
|
10
|
-
__all__ = [
|
|
10
|
+
__all__ = [
|
|
11
|
+
"ConversationMetaCreateParams",
|
|
12
|
+
"LlmCustomSetting",
|
|
13
|
+
"LlmCustomSettingBoundary",
|
|
14
|
+
"LlmCustomSettingExtraction",
|
|
15
|
+
"UserDetails",
|
|
16
|
+
]
|
|
11
17
|
|
|
12
18
|
|
|
13
19
|
class ConversationMetaCreateParams(TypedDict, total=False):
|
|
14
20
|
created_at: Required[str]
|
|
15
21
|
"""Conversation creation time (ISO 8601 format)"""
|
|
16
22
|
|
|
17
|
-
name: Required[str]
|
|
18
|
-
"""Conversation name"""
|
|
19
|
-
|
|
20
|
-
scene: Required[str]
|
|
21
|
-
"""Scene identifier, enum values from ScenarioType:
|
|
22
|
-
|
|
23
|
-
- group_chat: work/group chat scenario, suitable for group conversations such as
|
|
24
|
-
multi-person collaboration and project discussions
|
|
25
|
-
- assistant: assistant scenario, suitable for one-on-one AI assistant
|
|
26
|
-
conversations
|
|
27
|
-
"""
|
|
28
|
-
|
|
29
|
-
scene_desc: Required[Dict[str, object]]
|
|
30
|
-
"""Scene description object, can include fields like description"""
|
|
31
|
-
|
|
32
23
|
default_timezone: Optional[str]
|
|
33
24
|
"""Default timezone"""
|
|
34
25
|
|
|
@@ -41,6 +32,46 @@ class ConversationMetaCreateParams(TypedDict, total=False):
|
|
|
41
32
|
When null/not provided, represents default settings for this scene.
|
|
42
33
|
"""
|
|
43
34
|
|
|
35
|
+
llm_custom_setting: Optional[LlmCustomSetting]
|
|
36
|
+
"""LLM custom settings for algorithm control.
|
|
37
|
+
|
|
38
|
+
**Only for global config (group_id=null), not allowed for group config (group_id
|
|
39
|
+
provided).**
|
|
40
|
+
|
|
41
|
+
Allows configuring different LLM providers/models for different tasks like
|
|
42
|
+
boundary detection and memory extraction.
|
|
43
|
+
"""
|
|
44
|
+
|
|
45
|
+
name: Optional[str]
|
|
46
|
+
"""Group/conversation name.
|
|
47
|
+
|
|
48
|
+
**Required for group config (group_id provided), not allowed for global config
|
|
49
|
+
(group_id=null).**
|
|
50
|
+
"""
|
|
51
|
+
|
|
52
|
+
scene: Optional[str]
|
|
53
|
+
"""Scene identifier.
|
|
54
|
+
|
|
55
|
+
**Required for global config (group_id=null), not allowed for group config
|
|
56
|
+
(group_id provided).**
|
|
57
|
+
|
|
58
|
+
Enum values from ScenarioType:
|
|
59
|
+
|
|
60
|
+
- group_chat: work/group chat scenario, suitable for group conversations such as
|
|
61
|
+
multi-person collaboration and project discussions
|
|
62
|
+
- assistant: assistant scenario, suitable for one-on-one AI assistant
|
|
63
|
+
conversations
|
|
64
|
+
"""
|
|
65
|
+
|
|
66
|
+
scene_desc: Optional[Dict[str, object]]
|
|
67
|
+
"""Scene description object.
|
|
68
|
+
|
|
69
|
+
**Required for global config (group_id=null), not allowed for group config
|
|
70
|
+
(group_id provided).**
|
|
71
|
+
|
|
72
|
+
Can include fields like description, type, etc.
|
|
73
|
+
"""
|
|
74
|
+
|
|
44
75
|
tags: Optional[SequenceNotStr[str]]
|
|
45
76
|
"""Tag list"""
|
|
46
77
|
|
|
@@ -48,12 +79,52 @@ class ConversationMetaCreateParams(TypedDict, total=False):
|
|
|
48
79
|
"""Participant details, key is user ID, value is user detail object"""
|
|
49
80
|
|
|
50
81
|
|
|
51
|
-
class
|
|
52
|
-
"""
|
|
82
|
+
class LlmCustomSettingBoundary(TypedDict, total=False):
|
|
83
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
84
|
+
|
|
85
|
+
model: Required[str]
|
|
86
|
+
"""Model name"""
|
|
87
|
+
|
|
88
|
+
provider: Required[str]
|
|
89
|
+
"""LLM provider name"""
|
|
90
|
+
|
|
91
|
+
extra: Optional[Dict[str, object]]
|
|
92
|
+
"""Additional provider-specific configuration"""
|
|
93
|
+
|
|
94
|
+
|
|
95
|
+
class LlmCustomSettingExtraction(TypedDict, total=False):
|
|
96
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
97
|
+
|
|
98
|
+
model: Required[str]
|
|
99
|
+
"""Model name"""
|
|
53
100
|
|
|
54
|
-
|
|
101
|
+
provider: Required[str]
|
|
102
|
+
"""LLM provider name"""
|
|
103
|
+
|
|
104
|
+
extra: Optional[Dict[str, object]]
|
|
105
|
+
"""Additional provider-specific configuration"""
|
|
106
|
+
|
|
107
|
+
|
|
108
|
+
class LlmCustomSetting(TypedDict, total=False):
|
|
109
|
+
"""LLM custom settings for algorithm control.
|
|
110
|
+
|
|
111
|
+
**Only for global config (group_id=null),
|
|
112
|
+
not allowed for group config (group_id provided).**
|
|
113
|
+
|
|
114
|
+
Allows configuring different LLM providers/models for different tasks like boundary detection and memory extraction.
|
|
55
115
|
"""
|
|
56
116
|
|
|
117
|
+
boundary: Optional[LlmCustomSettingBoundary]
|
|
118
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
119
|
+
|
|
120
|
+
extra: Optional[Dict[str, object]]
|
|
121
|
+
"""Additional task-specific LLM configurations"""
|
|
122
|
+
|
|
123
|
+
extraction: Optional[LlmCustomSettingExtraction]
|
|
124
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
125
|
+
|
|
126
|
+
|
|
127
|
+
class UserDetails(TypedDict, total=False):
|
|
57
128
|
custom_role: Optional[str]
|
|
58
129
|
"""User's job/position role (e.g. developer, designer, manager)"""
|
|
59
130
|
|
|
@@ -4,7 +4,52 @@ from typing import Dict, List, Optional
|
|
|
4
4
|
|
|
5
5
|
from ...._models import BaseModel
|
|
6
6
|
|
|
7
|
-
__all__ = [
|
|
7
|
+
__all__ = [
|
|
8
|
+
"ConversationMetaCreateResponse",
|
|
9
|
+
"Result",
|
|
10
|
+
"ResultLlmCustomSetting",
|
|
11
|
+
"ResultLlmCustomSettingBoundary",
|
|
12
|
+
"ResultLlmCustomSettingExtraction",
|
|
13
|
+
]
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
class ResultLlmCustomSettingBoundary(BaseModel):
|
|
17
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
18
|
+
|
|
19
|
+
model: str
|
|
20
|
+
"""Model name"""
|
|
21
|
+
|
|
22
|
+
provider: str
|
|
23
|
+
"""LLM provider name"""
|
|
24
|
+
|
|
25
|
+
extra: Optional[Dict[str, object]] = None
|
|
26
|
+
"""Additional provider-specific configuration"""
|
|
27
|
+
|
|
28
|
+
|
|
29
|
+
class ResultLlmCustomSettingExtraction(BaseModel):
|
|
30
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
31
|
+
|
|
32
|
+
model: str
|
|
33
|
+
"""Model name"""
|
|
34
|
+
|
|
35
|
+
provider: str
|
|
36
|
+
"""LLM provider name"""
|
|
37
|
+
|
|
38
|
+
extra: Optional[Dict[str, object]] = None
|
|
39
|
+
"""Additional provider-specific configuration"""
|
|
40
|
+
|
|
41
|
+
|
|
42
|
+
class ResultLlmCustomSetting(BaseModel):
|
|
43
|
+
"""LLM custom settings (only for global config)"""
|
|
44
|
+
|
|
45
|
+
boundary: Optional[ResultLlmCustomSettingBoundary] = None
|
|
46
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
47
|
+
|
|
48
|
+
extra: Optional[Dict[str, object]] = None
|
|
49
|
+
"""Additional task-specific LLM configurations"""
|
|
50
|
+
|
|
51
|
+
extraction: Optional[ResultLlmCustomSettingExtraction] = None
|
|
52
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
8
53
|
|
|
9
54
|
|
|
10
55
|
class Result(BaseModel):
|
|
@@ -13,15 +58,9 @@ class Result(BaseModel):
|
|
|
13
58
|
id: str
|
|
14
59
|
"""Document ID"""
|
|
15
60
|
|
|
16
|
-
conversation_created_at: str
|
|
61
|
+
conversation_created_at: Optional[str] = None
|
|
17
62
|
"""Conversation creation time"""
|
|
18
63
|
|
|
19
|
-
name: str
|
|
20
|
-
"""Conversation name"""
|
|
21
|
-
|
|
22
|
-
scene: str
|
|
23
|
-
"""Scene identifier"""
|
|
24
|
-
|
|
25
64
|
created_at: Optional[str] = None
|
|
26
65
|
"""Record creation time"""
|
|
27
66
|
|
|
@@ -29,16 +68,25 @@ class Result(BaseModel):
|
|
|
29
68
|
"""Default timezone"""
|
|
30
69
|
|
|
31
70
|
description: Optional[str] = None
|
|
32
|
-
"""
|
|
71
|
+
"""Description"""
|
|
33
72
|
|
|
34
73
|
group_id: Optional[str] = None
|
|
35
|
-
"""Group ID (null for
|
|
74
|
+
"""Group ID (null for global config)"""
|
|
36
75
|
|
|
37
76
|
is_default: Optional[bool] = None
|
|
38
|
-
"""Whether this is the default config"""
|
|
77
|
+
"""Whether this is the global (default) config"""
|
|
78
|
+
|
|
79
|
+
llm_custom_setting: Optional[ResultLlmCustomSetting] = None
|
|
80
|
+
"""LLM custom settings (only for global config)"""
|
|
81
|
+
|
|
82
|
+
name: Optional[str] = None
|
|
83
|
+
"""Group/conversation name (only for group config)"""
|
|
84
|
+
|
|
85
|
+
scene: Optional[str] = None
|
|
86
|
+
"""Scene identifier (only for global config)"""
|
|
39
87
|
|
|
40
88
|
scene_desc: Optional[Dict[str, object]] = None
|
|
41
|
-
"""Scene description"""
|
|
89
|
+
"""Scene description (only for global config)"""
|
|
42
90
|
|
|
43
91
|
tags: Optional[List[str]] = None
|
|
44
92
|
"""Tags"""
|
|
@@ -51,11 +99,6 @@ class Result(BaseModel):
|
|
|
51
99
|
|
|
52
100
|
|
|
53
101
|
class ConversationMetaCreateResponse(BaseModel):
|
|
54
|
-
"""Save conversation metadata API response
|
|
55
|
-
|
|
56
|
-
Response for POST /api/v1/memories/conversation-meta endpoint.
|
|
57
|
-
"""
|
|
58
|
-
|
|
59
102
|
result: Result
|
|
60
103
|
"""Saved conversation metadata"""
|
|
61
104
|
|
|
@@ -4,7 +4,52 @@ from typing import Dict, List, Optional
|
|
|
4
4
|
|
|
5
5
|
from ...._models import BaseModel
|
|
6
6
|
|
|
7
|
-
__all__ = [
|
|
7
|
+
__all__ = [
|
|
8
|
+
"ConversationMetaGetResponse",
|
|
9
|
+
"Result",
|
|
10
|
+
"ResultLlmCustomSetting",
|
|
11
|
+
"ResultLlmCustomSettingBoundary",
|
|
12
|
+
"ResultLlmCustomSettingExtraction",
|
|
13
|
+
]
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
class ResultLlmCustomSettingBoundary(BaseModel):
|
|
17
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
18
|
+
|
|
19
|
+
model: str
|
|
20
|
+
"""Model name"""
|
|
21
|
+
|
|
22
|
+
provider: str
|
|
23
|
+
"""LLM provider name"""
|
|
24
|
+
|
|
25
|
+
extra: Optional[Dict[str, object]] = None
|
|
26
|
+
"""Additional provider-specific configuration"""
|
|
27
|
+
|
|
28
|
+
|
|
29
|
+
class ResultLlmCustomSettingExtraction(BaseModel):
|
|
30
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
31
|
+
|
|
32
|
+
model: str
|
|
33
|
+
"""Model name"""
|
|
34
|
+
|
|
35
|
+
provider: str
|
|
36
|
+
"""LLM provider name"""
|
|
37
|
+
|
|
38
|
+
extra: Optional[Dict[str, object]] = None
|
|
39
|
+
"""Additional provider-specific configuration"""
|
|
40
|
+
|
|
41
|
+
|
|
42
|
+
class ResultLlmCustomSetting(BaseModel):
|
|
43
|
+
"""LLM custom settings (only for global config)"""
|
|
44
|
+
|
|
45
|
+
boundary: Optional[ResultLlmCustomSettingBoundary] = None
|
|
46
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
47
|
+
|
|
48
|
+
extra: Optional[Dict[str, object]] = None
|
|
49
|
+
"""Additional task-specific LLM configurations"""
|
|
50
|
+
|
|
51
|
+
extraction: Optional[ResultLlmCustomSettingExtraction] = None
|
|
52
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
8
53
|
|
|
9
54
|
|
|
10
55
|
class Result(BaseModel):
|
|
@@ -13,15 +58,9 @@ class Result(BaseModel):
|
|
|
13
58
|
id: str
|
|
14
59
|
"""Document ID"""
|
|
15
60
|
|
|
16
|
-
conversation_created_at: str
|
|
61
|
+
conversation_created_at: Optional[str] = None
|
|
17
62
|
"""Conversation creation time"""
|
|
18
63
|
|
|
19
|
-
name: str
|
|
20
|
-
"""Conversation name"""
|
|
21
|
-
|
|
22
|
-
scene: str
|
|
23
|
-
"""Scene identifier"""
|
|
24
|
-
|
|
25
64
|
created_at: Optional[str] = None
|
|
26
65
|
"""Record creation time"""
|
|
27
66
|
|
|
@@ -29,16 +68,25 @@ class Result(BaseModel):
|
|
|
29
68
|
"""Default timezone"""
|
|
30
69
|
|
|
31
70
|
description: Optional[str] = None
|
|
32
|
-
"""
|
|
71
|
+
"""Description"""
|
|
33
72
|
|
|
34
73
|
group_id: Optional[str] = None
|
|
35
|
-
"""Group ID (null for
|
|
74
|
+
"""Group ID (null for global config)"""
|
|
36
75
|
|
|
37
76
|
is_default: Optional[bool] = None
|
|
38
|
-
"""Whether this is the default config"""
|
|
77
|
+
"""Whether this is the global (default) config"""
|
|
78
|
+
|
|
79
|
+
llm_custom_setting: Optional[ResultLlmCustomSetting] = None
|
|
80
|
+
"""LLM custom settings (only for global config)"""
|
|
81
|
+
|
|
82
|
+
name: Optional[str] = None
|
|
83
|
+
"""Group/conversation name (only for group config)"""
|
|
84
|
+
|
|
85
|
+
scene: Optional[str] = None
|
|
86
|
+
"""Scene identifier (only for global config)"""
|
|
39
87
|
|
|
40
88
|
scene_desc: Optional[Dict[str, object]] = None
|
|
41
|
-
"""Scene description"""
|
|
89
|
+
"""Scene description (only for global config)"""
|
|
42
90
|
|
|
43
91
|
tags: Optional[List[str]] = None
|
|
44
92
|
"""Tags"""
|
|
@@ -51,11 +99,6 @@ class Result(BaseModel):
|
|
|
51
99
|
|
|
52
100
|
|
|
53
101
|
class ConversationMetaGetResponse(BaseModel):
|
|
54
|
-
"""Get conversation metadata API response
|
|
55
|
-
|
|
56
|
-
Response for GET /api/v1/memories/conversation-meta endpoint.
|
|
57
|
-
"""
|
|
58
|
-
|
|
59
102
|
result: Result
|
|
60
103
|
"""Conversation metadata"""
|
|
61
104
|
|
|
@@ -3,11 +3,17 @@
|
|
|
3
3
|
from __future__ import annotations
|
|
4
4
|
|
|
5
5
|
from typing import Dict, Optional
|
|
6
|
-
from typing_extensions import TypedDict
|
|
6
|
+
from typing_extensions import Required, TypedDict
|
|
7
7
|
|
|
8
8
|
from ...._types import SequenceNotStr
|
|
9
9
|
|
|
10
|
-
__all__ = [
|
|
10
|
+
__all__ = [
|
|
11
|
+
"ConversationMetaUpdateParams",
|
|
12
|
+
"LlmCustomSetting",
|
|
13
|
+
"LlmCustomSettingBoundary",
|
|
14
|
+
"LlmCustomSettingExtraction",
|
|
15
|
+
"UserDetails",
|
|
16
|
+
]
|
|
11
17
|
|
|
12
18
|
|
|
13
19
|
class ConversationMetaUpdateParams(TypedDict, total=False):
|
|
@@ -15,16 +21,31 @@ class ConversationMetaUpdateParams(TypedDict, total=False):
|
|
|
15
21
|
"""New default timezone"""
|
|
16
22
|
|
|
17
23
|
description: Optional[str]
|
|
18
|
-
"""New
|
|
24
|
+
"""New description"""
|
|
19
25
|
|
|
20
26
|
group_id: Optional[str]
|
|
21
|
-
"""Group ID to update. When null, updates the default config."""
|
|
27
|
+
"""Group ID to update. When null, updates the global (default) config."""
|
|
28
|
+
|
|
29
|
+
llm_custom_setting: Optional[LlmCustomSetting]
|
|
30
|
+
"""New LLM custom settings.
|
|
31
|
+
|
|
32
|
+
**Only allowed for global config (group_id=null). Not allowed for group config
|
|
33
|
+
(inherited from global config).**
|
|
34
|
+
"""
|
|
22
35
|
|
|
23
36
|
name: Optional[str]
|
|
24
|
-
"""New conversation name
|
|
37
|
+
"""New group/conversation name.
|
|
38
|
+
|
|
39
|
+
**Only allowed for group config (group_id provided). Not allowed for global
|
|
40
|
+
config.**
|
|
41
|
+
"""
|
|
25
42
|
|
|
26
43
|
scene_desc: Optional[Dict[str, object]]
|
|
27
|
-
"""New scene description
|
|
44
|
+
"""New scene description.
|
|
45
|
+
|
|
46
|
+
**Only allowed for global config (group_id=null). Not allowed for group config
|
|
47
|
+
(inherited from global config).**
|
|
48
|
+
"""
|
|
28
49
|
|
|
29
50
|
tags: Optional[SequenceNotStr[str]]
|
|
30
51
|
"""New tag list"""
|
|
@@ -33,12 +54,50 @@ class ConversationMetaUpdateParams(TypedDict, total=False):
|
|
|
33
54
|
"""New user details (will completely replace existing user_details)"""
|
|
34
55
|
|
|
35
56
|
|
|
36
|
-
class
|
|
37
|
-
"""
|
|
57
|
+
class LlmCustomSettingBoundary(TypedDict, total=False):
|
|
58
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
59
|
+
|
|
60
|
+
model: Required[str]
|
|
61
|
+
"""Model name"""
|
|
62
|
+
|
|
63
|
+
provider: Required[str]
|
|
64
|
+
"""LLM provider name"""
|
|
65
|
+
|
|
66
|
+
extra: Optional[Dict[str, object]]
|
|
67
|
+
"""Additional provider-specific configuration"""
|
|
68
|
+
|
|
38
69
|
|
|
39
|
-
|
|
70
|
+
class LlmCustomSettingExtraction(TypedDict, total=False):
|
|
71
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
72
|
+
|
|
73
|
+
model: Required[str]
|
|
74
|
+
"""Model name"""
|
|
75
|
+
|
|
76
|
+
provider: Required[str]
|
|
77
|
+
"""LLM provider name"""
|
|
78
|
+
|
|
79
|
+
extra: Optional[Dict[str, object]]
|
|
80
|
+
"""Additional provider-specific configuration"""
|
|
81
|
+
|
|
82
|
+
|
|
83
|
+
class LlmCustomSetting(TypedDict, total=False):
|
|
84
|
+
"""New LLM custom settings.
|
|
85
|
+
|
|
86
|
+
**Only allowed for global config (group_id=null).
|
|
87
|
+
Not allowed for group config (inherited from global config).**
|
|
40
88
|
"""
|
|
41
89
|
|
|
90
|
+
boundary: Optional[LlmCustomSettingBoundary]
|
|
91
|
+
"""LLM config for boundary detection (fast, cheap model recommended)"""
|
|
92
|
+
|
|
93
|
+
extra: Optional[Dict[str, object]]
|
|
94
|
+
"""Additional task-specific LLM configurations"""
|
|
95
|
+
|
|
96
|
+
extraction: Optional[LlmCustomSettingExtraction]
|
|
97
|
+
"""LLM config for memory extraction (high quality model recommended)"""
|
|
98
|
+
|
|
99
|
+
|
|
100
|
+
class UserDetails(TypedDict, total=False):
|
|
42
101
|
custom_role: Optional[str]
|
|
43
102
|
"""User's job/position role (e.g. developer, designer, manager)"""
|
|
44
103
|
|
|
@@ -30,11 +30,6 @@ class Result(BaseModel):
|
|
|
30
30
|
|
|
31
31
|
|
|
32
32
|
class ConversationMetaUpdateResponse(BaseModel):
|
|
33
|
-
"""Patch conversation metadata API response
|
|
34
|
-
|
|
35
|
-
Response for PATCH /api/v1/memories/conversation-meta endpoint.
|
|
36
|
-
"""
|
|
37
|
-
|
|
38
33
|
result: Result
|
|
39
34
|
"""Patch result with updated fields"""
|
|
40
35
|
|
|
@@ -7,10 +7,10 @@ from typing_extensions import Required, TypedDict
|
|
|
7
7
|
|
|
8
8
|
from ..._types import SequenceNotStr
|
|
9
9
|
|
|
10
|
-
__all__ = ["
|
|
10
|
+
__all__ = ["MemoryAddParams"]
|
|
11
11
|
|
|
12
12
|
|
|
13
|
-
class
|
|
13
|
+
class MemoryAddParams(TypedDict, total=False):
|
|
14
14
|
content: Required[str]
|
|
15
15
|
"""Message content"""
|
|
16
16
|
|
|
@@ -9,6 +9,12 @@ __all__ = ["MemoryDeleteParams"]
|
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
class MemoryDeleteParams(TypedDict, total=False):
|
|
12
|
+
id: Optional[str]
|
|
13
|
+
"""Alias for memory_id (backward compatibility)"""
|
|
14
|
+
|
|
15
|
+
event_id: Optional[str]
|
|
16
|
+
"""Alias for memory_id (backward compatibility)"""
|
|
17
|
+
|
|
12
18
|
group_id: Optional[str]
|
|
13
19
|
"""Group ID (filter condition)"""
|
|
14
20
|
|