evermemos 0.3.8__py3-none-any.whl → 0.3.10__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (33) hide show
  1. evermemos/_base_client.py +5 -2
  2. evermemos/_compat.py +3 -3
  3. evermemos/_utils/_json.py +35 -0
  4. evermemos/_version.py +1 -1
  5. evermemos/resources/v1/__init__.py +13 -13
  6. evermemos/resources/v1/memories/conversation_meta.py +82 -40
  7. evermemos/resources/v1/memories/memories.py +133 -228
  8. evermemos/resources/v1/{stats → status}/__init__.py +13 -13
  9. evermemos/resources/v1/{stats → status}/request.py +6 -6
  10. evermemos/resources/v1/{stats/stats.py → status/status.py} +27 -27
  11. evermemos/resources/v1/v1.py +19 -19
  12. evermemos/types/v1/__init__.py +2 -4
  13. evermemos/types/v1/memories/conversation_meta_create_params.py +90 -19
  14. evermemos/types/v1/memories/conversation_meta_create_response.py +60 -17
  15. evermemos/types/v1/memories/conversation_meta_get_response.py +60 -17
  16. evermemos/types/v1/memories/conversation_meta_update_params.py +68 -9
  17. evermemos/types/v1/memories/conversation_meta_update_response.py +0 -5
  18. evermemos/types/v1/{memory_create_params.py → memory_add_params.py} +2 -2
  19. evermemos/types/v1/{memory_create_response.py → memory_add_response.py} +2 -2
  20. evermemos/types/v1/memory_delete_params.py +6 -0
  21. evermemos/types/v1/memory_delete_response.py +0 -5
  22. evermemos/types/v1/memory_get_response.py +7 -10
  23. evermemos/types/v1/memory_search_response.py +5 -18
  24. evermemos/types/v1/{stats → status}/request_get_params.py +1 -1
  25. evermemos/types/v1/{stats → status}/request_get_response.py +0 -5
  26. {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/METADATA +24 -27
  27. {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/RECORD +30 -32
  28. evermemos/types/v1/global_user_profile/__init__.py +0 -3
  29. evermemos/types/v1/memory_load_params.py +0 -83
  30. evermemos/types/v1/memory_load_response.py +0 -26
  31. /evermemos/types/v1/{stats → status}/__init__.py +0 -0
  32. {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/WHEEL +0 -0
  33. {evermemos-0.3.8.dist-info → evermemos-0.3.10.dist-info}/licenses/LICENSE +0 -0
@@ -4,13 +4,13 @@ from __future__ import annotations
4
4
 
5
5
  from ..._compat import cached_property
6
6
  from ..._resource import SyncAPIResource, AsyncAPIResource
7
- from .stats.stats import (
8
- StatsResource,
9
- AsyncStatsResource,
10
- StatsResourceWithRawResponse,
11
- AsyncStatsResourceWithRawResponse,
12
- StatsResourceWithStreamingResponse,
13
- AsyncStatsResourceWithStreamingResponse,
7
+ from .status.status import (
8
+ StatusResource,
9
+ AsyncStatusResource,
10
+ StatusResourceWithRawResponse,
11
+ AsyncStatusResourceWithRawResponse,
12
+ StatusResourceWithStreamingResponse,
13
+ AsyncStatusResourceWithStreamingResponse,
14
14
  )
15
15
  from .memories.memories import (
16
16
  MemoriesResource,
@@ -30,8 +30,8 @@ class V1Resource(SyncAPIResource):
30
30
  return MemoriesResource(self._client)
31
31
 
32
32
  @cached_property
33
- def stats(self) -> StatsResource:
34
- return StatsResource(self._client)
33
+ def status(self) -> StatusResource:
34
+ return StatusResource(self._client)
35
35
 
36
36
  @cached_property
37
37
  def with_raw_response(self) -> V1ResourceWithRawResponse:
@@ -59,8 +59,8 @@ class AsyncV1Resource(AsyncAPIResource):
59
59
  return AsyncMemoriesResource(self._client)
60
60
 
61
61
  @cached_property
62
- def stats(self) -> AsyncStatsResource:
63
- return AsyncStatsResource(self._client)
62
+ def status(self) -> AsyncStatusResource:
63
+ return AsyncStatusResource(self._client)
64
64
 
65
65
  @cached_property
66
66
  def with_raw_response(self) -> AsyncV1ResourceWithRawResponse:
@@ -91,8 +91,8 @@ class V1ResourceWithRawResponse:
91
91
  return MemoriesResourceWithRawResponse(self._v1.memories)
92
92
 
93
93
  @cached_property
94
- def stats(self) -> StatsResourceWithRawResponse:
95
- return StatsResourceWithRawResponse(self._v1.stats)
94
+ def status(self) -> StatusResourceWithRawResponse:
95
+ return StatusResourceWithRawResponse(self._v1.status)
96
96
 
97
97
 
98
98
  class AsyncV1ResourceWithRawResponse:
@@ -104,8 +104,8 @@ class AsyncV1ResourceWithRawResponse:
104
104
  return AsyncMemoriesResourceWithRawResponse(self._v1.memories)
105
105
 
106
106
  @cached_property
107
- def stats(self) -> AsyncStatsResourceWithRawResponse:
108
- return AsyncStatsResourceWithRawResponse(self._v1.stats)
107
+ def status(self) -> AsyncStatusResourceWithRawResponse:
108
+ return AsyncStatusResourceWithRawResponse(self._v1.status)
109
109
 
110
110
 
111
111
  class V1ResourceWithStreamingResponse:
@@ -117,8 +117,8 @@ class V1ResourceWithStreamingResponse:
117
117
  return MemoriesResourceWithStreamingResponse(self._v1.memories)
118
118
 
119
119
  @cached_property
120
- def stats(self) -> StatsResourceWithStreamingResponse:
121
- return StatsResourceWithStreamingResponse(self._v1.stats)
120
+ def status(self) -> StatusResourceWithStreamingResponse:
121
+ return StatusResourceWithStreamingResponse(self._v1.status)
122
122
 
123
123
 
124
124
  class AsyncV1ResourceWithStreamingResponse:
@@ -130,5 +130,5 @@ class AsyncV1ResourceWithStreamingResponse:
130
130
  return AsyncMemoriesResourceWithStreamingResponse(self._v1.memories)
131
131
 
132
132
  @cached_property
133
- def stats(self) -> AsyncStatsResourceWithStreamingResponse:
134
- return AsyncStatsResourceWithStreamingResponse(self._v1.stats)
133
+ def status(self) -> AsyncStatusResourceWithStreamingResponse:
134
+ return AsyncStatusResourceWithStreamingResponse(self._v1.status)
@@ -4,11 +4,9 @@ from __future__ import annotations
4
4
 
5
5
  from .metadata import Metadata as Metadata
6
6
  from .memory_type import MemoryType as MemoryType
7
- from .memory_load_params import MemoryLoadParams as MemoryLoadParams
7
+ from .memory_add_params import MemoryAddParams as MemoryAddParams
8
+ from .memory_add_response import MemoryAddResponse as MemoryAddResponse
8
9
  from .memory_get_response import MemoryGetResponse as MemoryGetResponse
9
- from .memory_create_params import MemoryCreateParams as MemoryCreateParams
10
10
  from .memory_delete_params import MemoryDeleteParams as MemoryDeleteParams
11
- from .memory_load_response import MemoryLoadResponse as MemoryLoadResponse
12
- from .memory_create_response import MemoryCreateResponse as MemoryCreateResponse
13
11
  from .memory_delete_response import MemoryDeleteResponse as MemoryDeleteResponse
14
12
  from .memory_search_response import MemorySearchResponse as MemorySearchResponse
@@ -7,28 +7,19 @@ from typing_extensions import Required, TypedDict
7
7
 
8
8
  from ...._types import SequenceNotStr
9
9
 
10
- __all__ = ["ConversationMetaCreateParams", "UserDetails"]
10
+ __all__ = [
11
+ "ConversationMetaCreateParams",
12
+ "LlmCustomSetting",
13
+ "LlmCustomSettingBoundary",
14
+ "LlmCustomSettingExtraction",
15
+ "UserDetails",
16
+ ]
11
17
 
12
18
 
13
19
  class ConversationMetaCreateParams(TypedDict, total=False):
14
20
  created_at: Required[str]
15
21
  """Conversation creation time (ISO 8601 format)"""
16
22
 
17
- name: Required[str]
18
- """Conversation name"""
19
-
20
- scene: Required[str]
21
- """Scene identifier, enum values from ScenarioType:
22
-
23
- - group_chat: work/group chat scenario, suitable for group conversations such as
24
- multi-person collaboration and project discussions
25
- - assistant: assistant scenario, suitable for one-on-one AI assistant
26
- conversations
27
- """
28
-
29
- scene_desc: Required[Dict[str, object]]
30
- """Scene description object, can include fields like description"""
31
-
32
23
  default_timezone: Optional[str]
33
24
  """Default timezone"""
34
25
 
@@ -41,6 +32,46 @@ class ConversationMetaCreateParams(TypedDict, total=False):
41
32
  When null/not provided, represents default settings for this scene.
42
33
  """
43
34
 
35
+ llm_custom_setting: Optional[LlmCustomSetting]
36
+ """LLM custom settings for algorithm control.
37
+
38
+ **Only for global config (group_id=null), not allowed for group config (group_id
39
+ provided).**
40
+
41
+ Allows configuring different LLM providers/models for different tasks like
42
+ boundary detection and memory extraction.
43
+ """
44
+
45
+ name: Optional[str]
46
+ """Group/conversation name.
47
+
48
+ **Required for group config (group_id provided), not allowed for global config
49
+ (group_id=null).**
50
+ """
51
+
52
+ scene: Optional[str]
53
+ """Scene identifier.
54
+
55
+ **Required for global config (group_id=null), not allowed for group config
56
+ (group_id provided).**
57
+
58
+ Enum values from ScenarioType:
59
+
60
+ - group_chat: work/group chat scenario, suitable for group conversations such as
61
+ multi-person collaboration and project discussions
62
+ - assistant: assistant scenario, suitable for one-on-one AI assistant
63
+ conversations
64
+ """
65
+
66
+ scene_desc: Optional[Dict[str, object]]
67
+ """Scene description object.
68
+
69
+ **Required for global config (group_id=null), not allowed for group config
70
+ (group_id provided).**
71
+
72
+ Can include fields like description, type, etc.
73
+ """
74
+
44
75
  tags: Optional[SequenceNotStr[str]]
45
76
  """Tag list"""
46
77
 
@@ -48,12 +79,52 @@ class ConversationMetaCreateParams(TypedDict, total=False):
48
79
  """Participant details, key is user ID, value is user detail object"""
49
80
 
50
81
 
51
- class UserDetails(TypedDict, total=False):
52
- """User details
82
+ class LlmCustomSettingBoundary(TypedDict, total=False):
83
+ """LLM config for boundary detection (fast, cheap model recommended)"""
84
+
85
+ model: Required[str]
86
+ """Model name"""
87
+
88
+ provider: Required[str]
89
+ """LLM provider name"""
90
+
91
+ extra: Optional[Dict[str, object]]
92
+ """Additional provider-specific configuration"""
93
+
94
+
95
+ class LlmCustomSettingExtraction(TypedDict, total=False):
96
+ """LLM config for memory extraction (high quality model recommended)"""
97
+
98
+ model: Required[str]
99
+ """Model name"""
53
100
 
54
- Structure for the value of ConversationMetaRequest.user_details
101
+ provider: Required[str]
102
+ """LLM provider name"""
103
+
104
+ extra: Optional[Dict[str, object]]
105
+ """Additional provider-specific configuration"""
106
+
107
+
108
+ class LlmCustomSetting(TypedDict, total=False):
109
+ """LLM custom settings for algorithm control.
110
+
111
+ **Only for global config (group_id=null),
112
+ not allowed for group config (group_id provided).**
113
+
114
+ Allows configuring different LLM providers/models for different tasks like boundary detection and memory extraction.
55
115
  """
56
116
 
117
+ boundary: Optional[LlmCustomSettingBoundary]
118
+ """LLM config for boundary detection (fast, cheap model recommended)"""
119
+
120
+ extra: Optional[Dict[str, object]]
121
+ """Additional task-specific LLM configurations"""
122
+
123
+ extraction: Optional[LlmCustomSettingExtraction]
124
+ """LLM config for memory extraction (high quality model recommended)"""
125
+
126
+
127
+ class UserDetails(TypedDict, total=False):
57
128
  custom_role: Optional[str]
58
129
  """User's job/position role (e.g. developer, designer, manager)"""
59
130
 
@@ -4,7 +4,52 @@ from typing import Dict, List, Optional
4
4
 
5
5
  from ...._models import BaseModel
6
6
 
7
- __all__ = ["ConversationMetaCreateResponse", "Result"]
7
+ __all__ = [
8
+ "ConversationMetaCreateResponse",
9
+ "Result",
10
+ "ResultLlmCustomSetting",
11
+ "ResultLlmCustomSettingBoundary",
12
+ "ResultLlmCustomSettingExtraction",
13
+ ]
14
+
15
+
16
+ class ResultLlmCustomSettingBoundary(BaseModel):
17
+ """LLM config for boundary detection (fast, cheap model recommended)"""
18
+
19
+ model: str
20
+ """Model name"""
21
+
22
+ provider: str
23
+ """LLM provider name"""
24
+
25
+ extra: Optional[Dict[str, object]] = None
26
+ """Additional provider-specific configuration"""
27
+
28
+
29
+ class ResultLlmCustomSettingExtraction(BaseModel):
30
+ """LLM config for memory extraction (high quality model recommended)"""
31
+
32
+ model: str
33
+ """Model name"""
34
+
35
+ provider: str
36
+ """LLM provider name"""
37
+
38
+ extra: Optional[Dict[str, object]] = None
39
+ """Additional provider-specific configuration"""
40
+
41
+
42
+ class ResultLlmCustomSetting(BaseModel):
43
+ """LLM custom settings (only for global config)"""
44
+
45
+ boundary: Optional[ResultLlmCustomSettingBoundary] = None
46
+ """LLM config for boundary detection (fast, cheap model recommended)"""
47
+
48
+ extra: Optional[Dict[str, object]] = None
49
+ """Additional task-specific LLM configurations"""
50
+
51
+ extraction: Optional[ResultLlmCustomSettingExtraction] = None
52
+ """LLM config for memory extraction (high quality model recommended)"""
8
53
 
9
54
 
10
55
  class Result(BaseModel):
@@ -13,15 +58,9 @@ class Result(BaseModel):
13
58
  id: str
14
59
  """Document ID"""
15
60
 
16
- conversation_created_at: str
61
+ conversation_created_at: Optional[str] = None
17
62
  """Conversation creation time"""
18
63
 
19
- name: str
20
- """Conversation name"""
21
-
22
- scene: str
23
- """Scene identifier"""
24
-
25
64
  created_at: Optional[str] = None
26
65
  """Record creation time"""
27
66
 
@@ -29,16 +68,25 @@ class Result(BaseModel):
29
68
  """Default timezone"""
30
69
 
31
70
  description: Optional[str] = None
32
- """Conversation description"""
71
+ """Description"""
33
72
 
34
73
  group_id: Optional[str] = None
35
- """Group ID (null for default config)"""
74
+ """Group ID (null for global config)"""
36
75
 
37
76
  is_default: Optional[bool] = None
38
- """Whether this is the default config"""
77
+ """Whether this is the global (default) config"""
78
+
79
+ llm_custom_setting: Optional[ResultLlmCustomSetting] = None
80
+ """LLM custom settings (only for global config)"""
81
+
82
+ name: Optional[str] = None
83
+ """Group/conversation name (only for group config)"""
84
+
85
+ scene: Optional[str] = None
86
+ """Scene identifier (only for global config)"""
39
87
 
40
88
  scene_desc: Optional[Dict[str, object]] = None
41
- """Scene description"""
89
+ """Scene description (only for global config)"""
42
90
 
43
91
  tags: Optional[List[str]] = None
44
92
  """Tags"""
@@ -51,11 +99,6 @@ class Result(BaseModel):
51
99
 
52
100
 
53
101
  class ConversationMetaCreateResponse(BaseModel):
54
- """Save conversation metadata API response
55
-
56
- Response for POST /api/v1/memories/conversation-meta endpoint.
57
- """
58
-
59
102
  result: Result
60
103
  """Saved conversation metadata"""
61
104
 
@@ -4,7 +4,52 @@ from typing import Dict, List, Optional
4
4
 
5
5
  from ...._models import BaseModel
6
6
 
7
- __all__ = ["ConversationMetaGetResponse", "Result"]
7
+ __all__ = [
8
+ "ConversationMetaGetResponse",
9
+ "Result",
10
+ "ResultLlmCustomSetting",
11
+ "ResultLlmCustomSettingBoundary",
12
+ "ResultLlmCustomSettingExtraction",
13
+ ]
14
+
15
+
16
+ class ResultLlmCustomSettingBoundary(BaseModel):
17
+ """LLM config for boundary detection (fast, cheap model recommended)"""
18
+
19
+ model: str
20
+ """Model name"""
21
+
22
+ provider: str
23
+ """LLM provider name"""
24
+
25
+ extra: Optional[Dict[str, object]] = None
26
+ """Additional provider-specific configuration"""
27
+
28
+
29
+ class ResultLlmCustomSettingExtraction(BaseModel):
30
+ """LLM config for memory extraction (high quality model recommended)"""
31
+
32
+ model: str
33
+ """Model name"""
34
+
35
+ provider: str
36
+ """LLM provider name"""
37
+
38
+ extra: Optional[Dict[str, object]] = None
39
+ """Additional provider-specific configuration"""
40
+
41
+
42
+ class ResultLlmCustomSetting(BaseModel):
43
+ """LLM custom settings (only for global config)"""
44
+
45
+ boundary: Optional[ResultLlmCustomSettingBoundary] = None
46
+ """LLM config for boundary detection (fast, cheap model recommended)"""
47
+
48
+ extra: Optional[Dict[str, object]] = None
49
+ """Additional task-specific LLM configurations"""
50
+
51
+ extraction: Optional[ResultLlmCustomSettingExtraction] = None
52
+ """LLM config for memory extraction (high quality model recommended)"""
8
53
 
9
54
 
10
55
  class Result(BaseModel):
@@ -13,15 +58,9 @@ class Result(BaseModel):
13
58
  id: str
14
59
  """Document ID"""
15
60
 
16
- conversation_created_at: str
61
+ conversation_created_at: Optional[str] = None
17
62
  """Conversation creation time"""
18
63
 
19
- name: str
20
- """Conversation name"""
21
-
22
- scene: str
23
- """Scene identifier"""
24
-
25
64
  created_at: Optional[str] = None
26
65
  """Record creation time"""
27
66
 
@@ -29,16 +68,25 @@ class Result(BaseModel):
29
68
  """Default timezone"""
30
69
 
31
70
  description: Optional[str] = None
32
- """Conversation description"""
71
+ """Description"""
33
72
 
34
73
  group_id: Optional[str] = None
35
- """Group ID (null for default config)"""
74
+ """Group ID (null for global config)"""
36
75
 
37
76
  is_default: Optional[bool] = None
38
- """Whether this is the default config"""
77
+ """Whether this is the global (default) config"""
78
+
79
+ llm_custom_setting: Optional[ResultLlmCustomSetting] = None
80
+ """LLM custom settings (only for global config)"""
81
+
82
+ name: Optional[str] = None
83
+ """Group/conversation name (only for group config)"""
84
+
85
+ scene: Optional[str] = None
86
+ """Scene identifier (only for global config)"""
39
87
 
40
88
  scene_desc: Optional[Dict[str, object]] = None
41
- """Scene description"""
89
+ """Scene description (only for global config)"""
42
90
 
43
91
  tags: Optional[List[str]] = None
44
92
  """Tags"""
@@ -51,11 +99,6 @@ class Result(BaseModel):
51
99
 
52
100
 
53
101
  class ConversationMetaGetResponse(BaseModel):
54
- """Get conversation metadata API response
55
-
56
- Response for GET /api/v1/memories/conversation-meta endpoint.
57
- """
58
-
59
102
  result: Result
60
103
  """Conversation metadata"""
61
104
 
@@ -3,11 +3,17 @@
3
3
  from __future__ import annotations
4
4
 
5
5
  from typing import Dict, Optional
6
- from typing_extensions import TypedDict
6
+ from typing_extensions import Required, TypedDict
7
7
 
8
8
  from ...._types import SequenceNotStr
9
9
 
10
- __all__ = ["ConversationMetaUpdateParams", "UserDetails"]
10
+ __all__ = [
11
+ "ConversationMetaUpdateParams",
12
+ "LlmCustomSetting",
13
+ "LlmCustomSettingBoundary",
14
+ "LlmCustomSettingExtraction",
15
+ "UserDetails",
16
+ ]
11
17
 
12
18
 
13
19
  class ConversationMetaUpdateParams(TypedDict, total=False):
@@ -15,16 +21,31 @@ class ConversationMetaUpdateParams(TypedDict, total=False):
15
21
  """New default timezone"""
16
22
 
17
23
  description: Optional[str]
18
- """New conversation description"""
24
+ """New description"""
19
25
 
20
26
  group_id: Optional[str]
21
- """Group ID to update. When null, updates the default config."""
27
+ """Group ID to update. When null, updates the global (default) config."""
28
+
29
+ llm_custom_setting: Optional[LlmCustomSetting]
30
+ """New LLM custom settings.
31
+
32
+ **Only allowed for global config (group_id=null). Not allowed for group config
33
+ (inherited from global config).**
34
+ """
22
35
 
23
36
  name: Optional[str]
24
- """New conversation name"""
37
+ """New group/conversation name.
38
+
39
+ **Only allowed for group config (group_id provided). Not allowed for global
40
+ config.**
41
+ """
25
42
 
26
43
  scene_desc: Optional[Dict[str, object]]
27
- """New scene description"""
44
+ """New scene description.
45
+
46
+ **Only allowed for global config (group_id=null). Not allowed for group config
47
+ (inherited from global config).**
48
+ """
28
49
 
29
50
  tags: Optional[SequenceNotStr[str]]
30
51
  """New tag list"""
@@ -33,12 +54,50 @@ class ConversationMetaUpdateParams(TypedDict, total=False):
33
54
  """New user details (will completely replace existing user_details)"""
34
55
 
35
56
 
36
- class UserDetails(TypedDict, total=False):
37
- """User details
57
+ class LlmCustomSettingBoundary(TypedDict, total=False):
58
+ """LLM config for boundary detection (fast, cheap model recommended)"""
59
+
60
+ model: Required[str]
61
+ """Model name"""
62
+
63
+ provider: Required[str]
64
+ """LLM provider name"""
65
+
66
+ extra: Optional[Dict[str, object]]
67
+ """Additional provider-specific configuration"""
68
+
38
69
 
39
- Structure for the value of ConversationMetaRequest.user_details
70
+ class LlmCustomSettingExtraction(TypedDict, total=False):
71
+ """LLM config for memory extraction (high quality model recommended)"""
72
+
73
+ model: Required[str]
74
+ """Model name"""
75
+
76
+ provider: Required[str]
77
+ """LLM provider name"""
78
+
79
+ extra: Optional[Dict[str, object]]
80
+ """Additional provider-specific configuration"""
81
+
82
+
83
+ class LlmCustomSetting(TypedDict, total=False):
84
+ """New LLM custom settings.
85
+
86
+ **Only allowed for global config (group_id=null).
87
+ Not allowed for group config (inherited from global config).**
40
88
  """
41
89
 
90
+ boundary: Optional[LlmCustomSettingBoundary]
91
+ """LLM config for boundary detection (fast, cheap model recommended)"""
92
+
93
+ extra: Optional[Dict[str, object]]
94
+ """Additional task-specific LLM configurations"""
95
+
96
+ extraction: Optional[LlmCustomSettingExtraction]
97
+ """LLM config for memory extraction (high quality model recommended)"""
98
+
99
+
100
+ class UserDetails(TypedDict, total=False):
42
101
  custom_role: Optional[str]
43
102
  """User's job/position role (e.g. developer, designer, manager)"""
44
103
 
@@ -30,11 +30,6 @@ class Result(BaseModel):
30
30
 
31
31
 
32
32
  class ConversationMetaUpdateResponse(BaseModel):
33
- """Patch conversation metadata API response
34
-
35
- Response for PATCH /api/v1/memories/conversation-meta endpoint.
36
- """
37
-
38
33
  result: Result
39
34
  """Patch result with updated fields"""
40
35
 
@@ -7,10 +7,10 @@ from typing_extensions import Required, TypedDict
7
7
 
8
8
  from ..._types import SequenceNotStr
9
9
 
10
- __all__ = ["MemoryCreateParams"]
10
+ __all__ = ["MemoryAddParams"]
11
11
 
12
12
 
13
- class MemoryCreateParams(TypedDict, total=False):
13
+ class MemoryAddParams(TypedDict, total=False):
14
14
  content: Required[str]
15
15
  """Message content"""
16
16
 
@@ -2,10 +2,10 @@
2
2
 
3
3
  from ..._models import BaseModel
4
4
 
5
- __all__ = ["MemoryCreateResponse"]
5
+ __all__ = ["MemoryAddResponse"]
6
6
 
7
7
 
8
- class MemoryCreateResponse(BaseModel):
8
+ class MemoryAddResponse(BaseModel):
9
9
  message: str
10
10
 
11
11
  request_id: str
@@ -9,6 +9,12 @@ __all__ = ["MemoryDeleteParams"]
9
9
 
10
10
 
11
11
  class MemoryDeleteParams(TypedDict, total=False):
12
+ id: Optional[str]
13
+ """Alias for memory_id (backward compatibility)"""
14
+
15
+ event_id: Optional[str]
16
+ """Alias for memory_id (backward compatibility)"""
17
+
12
18
  group_id: Optional[str]
13
19
  """Group ID (filter condition)"""
14
20
 
@@ -18,11 +18,6 @@ class Result(BaseModel):
18
18
 
19
19
 
20
20
  class MemoryDeleteResponse(BaseModel):
21
- """Delete memories API response
22
-
23
- Response for DELETE /api/v1/memories endpoint.
24
- """
25
-
26
21
  result: Result
27
22
  """Delete operation result"""
28
23