llama-cloud 0.1.36__py3-none-any.whl → 0.1.38__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +20 -4
- llama_cloud/client.py +3 -0
- llama_cloud/resources/__init__.py +3 -3
- llama_cloud/resources/admin/client.py +57 -0
- llama_cloud/resources/alpha/__init__.py +2 -0
- llama_cloud/resources/alpha/client.py +118 -0
- llama_cloud/resources/beta/client.py +576 -20
- llama_cloud/resources/chat_apps/client.py +32 -8
- llama_cloud/resources/classifier/client.py +139 -11
- llama_cloud/resources/data_sinks/client.py +32 -8
- llama_cloud/resources/data_sources/client.py +32 -8
- llama_cloud/resources/data_sources/types/data_source_update_component.py +2 -0
- llama_cloud/resources/embedding_model_configs/client.py +48 -12
- llama_cloud/resources/files/__init__.py +2 -2
- llama_cloud/resources/files/client.py +189 -113
- llama_cloud/resources/files/types/__init__.py +1 -3
- llama_cloud/resources/jobs/client.py +12 -6
- llama_cloud/resources/llama_extract/client.py +138 -32
- llama_cloud/resources/organizations/client.py +18 -4
- llama_cloud/resources/parsing/client.py +16 -4
- llama_cloud/resources/pipelines/client.py +32 -8
- llama_cloud/resources/projects/client.py +78 -18
- llama_cloud/resources/reports/client.py +126 -30
- llama_cloud/resources/retrievers/client.py +48 -12
- llama_cloud/types/__init__.py +20 -2
- llama_cloud/types/agent_deployment_summary.py +1 -0
- llama_cloud/types/classify_job.py +2 -0
- llama_cloud/types/cloud_jira_data_source_v_2.py +52 -0
- llama_cloud/types/cloud_jira_data_source_v_2_api_version.py +21 -0
- llama_cloud/types/configurable_data_source_names.py +4 -0
- llama_cloud/types/data_source_component.py +2 -0
- llama_cloud/types/data_source_create_component.py +2 -0
- llama_cloud/types/data_source_reader_version_metadata_reader_version.py +9 -1
- llama_cloud/types/file_create.py +41 -0
- llama_cloud/types/{classify_job_with_status.py → file_filter.py} +8 -15
- llama_cloud/types/file_query_response.py +38 -0
- llama_cloud/types/llama_extract_mode_availability.py +37 -0
- llama_cloud/types/llama_extract_mode_availability_status.py +17 -0
- llama_cloud/types/paginated_response_classify_job.py +34 -0
- llama_cloud/types/pipeline_data_source_component.py +2 -0
- llama_cloud/types/usage_response_active_alerts_item.py +4 -0
- {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/METADATA +2 -1
- {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/RECORD +47 -38
- {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/WHEEL +1 -1
- /llama_cloud/{resources/files/types → types}/file_create_permission_info_value.py +0 -0
- /llama_cloud/{resources/files/types → types}/file_create_resource_info_value.py +0 -0
- {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/LICENSE +0 -0
|
@@ -33,7 +33,11 @@ class DataSourcesClient:
|
|
|
33
33
|
self._client_wrapper = client_wrapper
|
|
34
34
|
|
|
35
35
|
def list_data_sources(
|
|
36
|
-
self,
|
|
36
|
+
self,
|
|
37
|
+
*,
|
|
38
|
+
project_id: typing.Optional[str] = None,
|
|
39
|
+
organization_id: typing.Optional[str] = None,
|
|
40
|
+
project_id: typing.Optional[str] = None,
|
|
37
41
|
) -> typing.List[DataSource]:
|
|
38
42
|
"""
|
|
39
43
|
List data sources for a given project.
|
|
@@ -43,6 +47,8 @@ class DataSourcesClient:
|
|
|
43
47
|
- project_id: typing.Optional[str].
|
|
44
48
|
|
|
45
49
|
- organization_id: typing.Optional[str].
|
|
50
|
+
|
|
51
|
+
- project_id: typing.Optional[str].
|
|
46
52
|
---
|
|
47
53
|
from llama_cloud.client import LlamaCloud
|
|
48
54
|
|
|
@@ -55,7 +61,7 @@ class DataSourcesClient:
|
|
|
55
61
|
"GET",
|
|
56
62
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/data-sources"),
|
|
57
63
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
58
|
-
headers=self._client_wrapper.get_headers(),
|
|
64
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
59
65
|
timeout=60,
|
|
60
66
|
)
|
|
61
67
|
if 200 <= _response.status_code < 300:
|
|
@@ -74,6 +80,7 @@ class DataSourcesClient:
|
|
|
74
80
|
project_id: typing.Optional[str] = None,
|
|
75
81
|
organization_id: typing.Optional[str] = None,
|
|
76
82
|
request: DataSourceCreate,
|
|
83
|
+
project_id: typing.Optional[str] = None,
|
|
77
84
|
) -> DataSource:
|
|
78
85
|
"""
|
|
79
86
|
Create a new data source.
|
|
@@ -84,6 +91,8 @@ class DataSourcesClient:
|
|
|
84
91
|
- organization_id: typing.Optional[str].
|
|
85
92
|
|
|
86
93
|
- request: DataSourceCreate.
|
|
94
|
+
|
|
95
|
+
- project_id: typing.Optional[str].
|
|
87
96
|
---
|
|
88
97
|
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
89
98
|
from llama_cloud.client import LlamaCloud
|
|
@@ -103,7 +112,7 @@ class DataSourcesClient:
|
|
|
103
112
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/data-sources"),
|
|
104
113
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
105
114
|
json=jsonable_encoder(request),
|
|
106
|
-
headers=self._client_wrapper.get_headers(),
|
|
115
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
107
116
|
timeout=60,
|
|
108
117
|
)
|
|
109
118
|
if 200 <= _response.status_code < 300:
|
|
@@ -122,6 +131,7 @@ class DataSourcesClient:
|
|
|
122
131
|
project_id: typing.Optional[str] = None,
|
|
123
132
|
organization_id: typing.Optional[str] = None,
|
|
124
133
|
request: DataSourceCreate,
|
|
134
|
+
project_id: typing.Optional[str] = None,
|
|
125
135
|
) -> DataSource:
|
|
126
136
|
"""
|
|
127
137
|
Upserts a data source.
|
|
@@ -133,6 +143,8 @@ class DataSourcesClient:
|
|
|
133
143
|
- organization_id: typing.Optional[str].
|
|
134
144
|
|
|
135
145
|
- request: DataSourceCreate.
|
|
146
|
+
|
|
147
|
+
- project_id: typing.Optional[str].
|
|
136
148
|
---
|
|
137
149
|
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
138
150
|
from llama_cloud.client import LlamaCloud
|
|
@@ -152,7 +164,7 @@ class DataSourcesClient:
|
|
|
152
164
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/data-sources"),
|
|
153
165
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
154
166
|
json=jsonable_encoder(request),
|
|
155
|
-
headers=self._client_wrapper.get_headers(),
|
|
167
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
156
168
|
timeout=60,
|
|
157
169
|
)
|
|
158
170
|
if 200 <= _response.status_code < 300:
|
|
@@ -293,7 +305,11 @@ class AsyncDataSourcesClient:
|
|
|
293
305
|
self._client_wrapper = client_wrapper
|
|
294
306
|
|
|
295
307
|
async def list_data_sources(
|
|
296
|
-
self,
|
|
308
|
+
self,
|
|
309
|
+
*,
|
|
310
|
+
project_id: typing.Optional[str] = None,
|
|
311
|
+
organization_id: typing.Optional[str] = None,
|
|
312
|
+
project_id: typing.Optional[str] = None,
|
|
297
313
|
) -> typing.List[DataSource]:
|
|
298
314
|
"""
|
|
299
315
|
List data sources for a given project.
|
|
@@ -303,6 +319,8 @@ class AsyncDataSourcesClient:
|
|
|
303
319
|
- project_id: typing.Optional[str].
|
|
304
320
|
|
|
305
321
|
- organization_id: typing.Optional[str].
|
|
322
|
+
|
|
323
|
+
- project_id: typing.Optional[str].
|
|
306
324
|
---
|
|
307
325
|
from llama_cloud.client import AsyncLlamaCloud
|
|
308
326
|
|
|
@@ -315,7 +333,7 @@ class AsyncDataSourcesClient:
|
|
|
315
333
|
"GET",
|
|
316
334
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/data-sources"),
|
|
317
335
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
318
|
-
headers=self._client_wrapper.get_headers(),
|
|
336
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
319
337
|
timeout=60,
|
|
320
338
|
)
|
|
321
339
|
if 200 <= _response.status_code < 300:
|
|
@@ -334,6 +352,7 @@ class AsyncDataSourcesClient:
|
|
|
334
352
|
project_id: typing.Optional[str] = None,
|
|
335
353
|
organization_id: typing.Optional[str] = None,
|
|
336
354
|
request: DataSourceCreate,
|
|
355
|
+
project_id: typing.Optional[str] = None,
|
|
337
356
|
) -> DataSource:
|
|
338
357
|
"""
|
|
339
358
|
Create a new data source.
|
|
@@ -344,6 +363,8 @@ class AsyncDataSourcesClient:
|
|
|
344
363
|
- organization_id: typing.Optional[str].
|
|
345
364
|
|
|
346
365
|
- request: DataSourceCreate.
|
|
366
|
+
|
|
367
|
+
- project_id: typing.Optional[str].
|
|
347
368
|
---
|
|
348
369
|
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
349
370
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -363,7 +384,7 @@ class AsyncDataSourcesClient:
|
|
|
363
384
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/data-sources"),
|
|
364
385
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
365
386
|
json=jsonable_encoder(request),
|
|
366
|
-
headers=self._client_wrapper.get_headers(),
|
|
387
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
367
388
|
timeout=60,
|
|
368
389
|
)
|
|
369
390
|
if 200 <= _response.status_code < 300:
|
|
@@ -382,6 +403,7 @@ class AsyncDataSourcesClient:
|
|
|
382
403
|
project_id: typing.Optional[str] = None,
|
|
383
404
|
organization_id: typing.Optional[str] = None,
|
|
384
405
|
request: DataSourceCreate,
|
|
406
|
+
project_id: typing.Optional[str] = None,
|
|
385
407
|
) -> DataSource:
|
|
386
408
|
"""
|
|
387
409
|
Upserts a data source.
|
|
@@ -393,6 +415,8 @@ class AsyncDataSourcesClient:
|
|
|
393
415
|
- organization_id: typing.Optional[str].
|
|
394
416
|
|
|
395
417
|
- request: DataSourceCreate.
|
|
418
|
+
|
|
419
|
+
- project_id: typing.Optional[str].
|
|
396
420
|
---
|
|
397
421
|
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
398
422
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -412,7 +436,7 @@ class AsyncDataSourcesClient:
|
|
|
412
436
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/data-sources"),
|
|
413
437
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
414
438
|
json=jsonable_encoder(request),
|
|
415
|
-
headers=self._client_wrapper.get_headers(),
|
|
439
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
416
440
|
timeout=60,
|
|
417
441
|
)
|
|
418
442
|
if 200 <= _response.status_code < 300:
|
|
@@ -6,6 +6,7 @@ from ....types.cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSo
|
|
|
6
6
|
from ....types.cloud_box_data_source import CloudBoxDataSource
|
|
7
7
|
from ....types.cloud_confluence_data_source import CloudConfluenceDataSource
|
|
8
8
|
from ....types.cloud_jira_data_source import CloudJiraDataSource
|
|
9
|
+
from ....types.cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
|
|
9
10
|
from ....types.cloud_notion_page_data_source import CloudNotionPageDataSource
|
|
10
11
|
from ....types.cloud_one_drive_data_source import CloudOneDriveDataSource
|
|
11
12
|
from ....types.cloud_s_3_data_source import CloudS3DataSource
|
|
@@ -22,5 +23,6 @@ DataSourceUpdateComponent = typing.Union[
|
|
|
22
23
|
CloudNotionPageDataSource,
|
|
23
24
|
CloudConfluenceDataSource,
|
|
24
25
|
CloudJiraDataSource,
|
|
26
|
+
CloudJiraDataSourceV2,
|
|
25
27
|
CloudBoxDataSource,
|
|
26
28
|
]
|
|
@@ -31,13 +31,19 @@ class EmbeddingModelConfigsClient:
|
|
|
31
31
|
self._client_wrapper = client_wrapper
|
|
32
32
|
|
|
33
33
|
def list_embedding_model_configs(
|
|
34
|
-
self,
|
|
34
|
+
self,
|
|
35
|
+
*,
|
|
36
|
+
project_id: typing.Optional[str] = None,
|
|
37
|
+
organization_id: typing.Optional[str] = None,
|
|
38
|
+
project_id: typing.Optional[str] = None,
|
|
35
39
|
) -> typing.List[EmbeddingModelConfig]:
|
|
36
40
|
"""
|
|
37
41
|
Parameters:
|
|
38
42
|
- project_id: typing.Optional[str].
|
|
39
43
|
|
|
40
44
|
- organization_id: typing.Optional[str].
|
|
45
|
+
|
|
46
|
+
- project_id: typing.Optional[str].
|
|
41
47
|
---
|
|
42
48
|
from llama_cloud.client import LlamaCloud
|
|
43
49
|
|
|
@@ -50,7 +56,7 @@ class EmbeddingModelConfigsClient:
|
|
|
50
56
|
"GET",
|
|
51
57
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
52
58
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
53
|
-
headers=self._client_wrapper.get_headers(),
|
|
59
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
54
60
|
timeout=60,
|
|
55
61
|
)
|
|
56
62
|
if 200 <= _response.status_code < 300:
|
|
@@ -70,6 +76,7 @@ class EmbeddingModelConfigsClient:
|
|
|
70
76
|
organization_id: typing.Optional[str] = None,
|
|
71
77
|
name: str,
|
|
72
78
|
embedding_config: EmbeddingModelConfigCreateEmbeddingConfig,
|
|
79
|
+
project_id: typing.Optional[str] = None,
|
|
73
80
|
) -> EmbeddingModelConfig:
|
|
74
81
|
"""
|
|
75
82
|
Create a new embedding model configuration within a specified project.
|
|
@@ -82,13 +89,15 @@ class EmbeddingModelConfigsClient:
|
|
|
82
89
|
- name: str. The name of the embedding model config.
|
|
83
90
|
|
|
84
91
|
- embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
|
|
92
|
+
|
|
93
|
+
- project_id: typing.Optional[str].
|
|
85
94
|
"""
|
|
86
95
|
_response = self._client_wrapper.httpx_client.request(
|
|
87
96
|
"POST",
|
|
88
97
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
89
98
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
90
99
|
json=jsonable_encoder({"name": name, "embedding_config": embedding_config}),
|
|
91
|
-
headers=self._client_wrapper.get_headers(),
|
|
100
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
92
101
|
timeout=60,
|
|
93
102
|
)
|
|
94
103
|
if 200 <= _response.status_code < 300:
|
|
@@ -107,6 +116,7 @@ class EmbeddingModelConfigsClient:
|
|
|
107
116
|
project_id: typing.Optional[str] = None,
|
|
108
117
|
organization_id: typing.Optional[str] = None,
|
|
109
118
|
request: EmbeddingModelConfigUpdate,
|
|
119
|
+
project_id: typing.Optional[str] = None,
|
|
110
120
|
) -> EmbeddingModelConfig:
|
|
111
121
|
"""
|
|
112
122
|
Upserts an embedding model config.
|
|
@@ -118,13 +128,15 @@ class EmbeddingModelConfigsClient:
|
|
|
118
128
|
- organization_id: typing.Optional[str].
|
|
119
129
|
|
|
120
130
|
- request: EmbeddingModelConfigUpdate.
|
|
131
|
+
|
|
132
|
+
- project_id: typing.Optional[str].
|
|
121
133
|
"""
|
|
122
134
|
_response = self._client_wrapper.httpx_client.request(
|
|
123
135
|
"PUT",
|
|
124
136
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
125
137
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
126
138
|
json=jsonable_encoder(request),
|
|
127
|
-
headers=self._client_wrapper.get_headers(),
|
|
139
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
128
140
|
timeout=60,
|
|
129
141
|
)
|
|
130
142
|
if 200 <= _response.status_code < 300:
|
|
@@ -144,6 +156,7 @@ class EmbeddingModelConfigsClient:
|
|
|
144
156
|
project_id: typing.Optional[str] = None,
|
|
145
157
|
organization_id: typing.Optional[str] = None,
|
|
146
158
|
request: EmbeddingModelConfigUpdate,
|
|
159
|
+
project_id: typing.Optional[str] = None,
|
|
147
160
|
) -> EmbeddingModelConfig:
|
|
148
161
|
"""
|
|
149
162
|
Update an embedding model config by ID.
|
|
@@ -156,6 +169,8 @@ class EmbeddingModelConfigsClient:
|
|
|
156
169
|
- organization_id: typing.Optional[str].
|
|
157
170
|
|
|
158
171
|
- request: EmbeddingModelConfigUpdate.
|
|
172
|
+
|
|
173
|
+
- project_id: typing.Optional[str].
|
|
159
174
|
"""
|
|
160
175
|
_response = self._client_wrapper.httpx_client.request(
|
|
161
176
|
"PUT",
|
|
@@ -164,7 +179,7 @@ class EmbeddingModelConfigsClient:
|
|
|
164
179
|
),
|
|
165
180
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
166
181
|
json=jsonable_encoder(request),
|
|
167
|
-
headers=self._client_wrapper.get_headers(),
|
|
182
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
168
183
|
timeout=60,
|
|
169
184
|
)
|
|
170
185
|
if 200 <= _response.status_code < 300:
|
|
@@ -183,6 +198,7 @@ class EmbeddingModelConfigsClient:
|
|
|
183
198
|
*,
|
|
184
199
|
project_id: typing.Optional[str] = None,
|
|
185
200
|
organization_id: typing.Optional[str] = None,
|
|
201
|
+
project_id: typing.Optional[str] = None,
|
|
186
202
|
) -> None:
|
|
187
203
|
"""
|
|
188
204
|
Delete an embedding model config by ID.
|
|
@@ -193,6 +209,8 @@ class EmbeddingModelConfigsClient:
|
|
|
193
209
|
- project_id: typing.Optional[str].
|
|
194
210
|
|
|
195
211
|
- organization_id: typing.Optional[str].
|
|
212
|
+
|
|
213
|
+
- project_id: typing.Optional[str].
|
|
196
214
|
---
|
|
197
215
|
from llama_cloud.client import LlamaCloud
|
|
198
216
|
|
|
@@ -209,7 +227,7 @@ class EmbeddingModelConfigsClient:
|
|
|
209
227
|
f"{self._client_wrapper.get_base_url()}/", f"api/v1/embedding-model-configs/{embedding_model_config_id}"
|
|
210
228
|
),
|
|
211
229
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
212
|
-
headers=self._client_wrapper.get_headers(),
|
|
230
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
213
231
|
timeout=60,
|
|
214
232
|
)
|
|
215
233
|
if 200 <= _response.status_code < 300:
|
|
@@ -228,13 +246,19 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
228
246
|
self._client_wrapper = client_wrapper
|
|
229
247
|
|
|
230
248
|
async def list_embedding_model_configs(
|
|
231
|
-
self,
|
|
249
|
+
self,
|
|
250
|
+
*,
|
|
251
|
+
project_id: typing.Optional[str] = None,
|
|
252
|
+
organization_id: typing.Optional[str] = None,
|
|
253
|
+
project_id: typing.Optional[str] = None,
|
|
232
254
|
) -> typing.List[EmbeddingModelConfig]:
|
|
233
255
|
"""
|
|
234
256
|
Parameters:
|
|
235
257
|
- project_id: typing.Optional[str].
|
|
236
258
|
|
|
237
259
|
- organization_id: typing.Optional[str].
|
|
260
|
+
|
|
261
|
+
- project_id: typing.Optional[str].
|
|
238
262
|
---
|
|
239
263
|
from llama_cloud.client import AsyncLlamaCloud
|
|
240
264
|
|
|
@@ -247,7 +271,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
247
271
|
"GET",
|
|
248
272
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
249
273
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
250
|
-
headers=self._client_wrapper.get_headers(),
|
|
274
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
251
275
|
timeout=60,
|
|
252
276
|
)
|
|
253
277
|
if 200 <= _response.status_code < 300:
|
|
@@ -267,6 +291,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
267
291
|
organization_id: typing.Optional[str] = None,
|
|
268
292
|
name: str,
|
|
269
293
|
embedding_config: EmbeddingModelConfigCreateEmbeddingConfig,
|
|
294
|
+
project_id: typing.Optional[str] = None,
|
|
270
295
|
) -> EmbeddingModelConfig:
|
|
271
296
|
"""
|
|
272
297
|
Create a new embedding model configuration within a specified project.
|
|
@@ -279,13 +304,15 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
279
304
|
- name: str. The name of the embedding model config.
|
|
280
305
|
|
|
281
306
|
- embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
|
|
307
|
+
|
|
308
|
+
- project_id: typing.Optional[str].
|
|
282
309
|
"""
|
|
283
310
|
_response = await self._client_wrapper.httpx_client.request(
|
|
284
311
|
"POST",
|
|
285
312
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
286
313
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
287
314
|
json=jsonable_encoder({"name": name, "embedding_config": embedding_config}),
|
|
288
|
-
headers=self._client_wrapper.get_headers(),
|
|
315
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
289
316
|
timeout=60,
|
|
290
317
|
)
|
|
291
318
|
if 200 <= _response.status_code < 300:
|
|
@@ -304,6 +331,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
304
331
|
project_id: typing.Optional[str] = None,
|
|
305
332
|
organization_id: typing.Optional[str] = None,
|
|
306
333
|
request: EmbeddingModelConfigUpdate,
|
|
334
|
+
project_id: typing.Optional[str] = None,
|
|
307
335
|
) -> EmbeddingModelConfig:
|
|
308
336
|
"""
|
|
309
337
|
Upserts an embedding model config.
|
|
@@ -315,13 +343,15 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
315
343
|
- organization_id: typing.Optional[str].
|
|
316
344
|
|
|
317
345
|
- request: EmbeddingModelConfigUpdate.
|
|
346
|
+
|
|
347
|
+
- project_id: typing.Optional[str].
|
|
318
348
|
"""
|
|
319
349
|
_response = await self._client_wrapper.httpx_client.request(
|
|
320
350
|
"PUT",
|
|
321
351
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
322
352
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
323
353
|
json=jsonable_encoder(request),
|
|
324
|
-
headers=self._client_wrapper.get_headers(),
|
|
354
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
325
355
|
timeout=60,
|
|
326
356
|
)
|
|
327
357
|
if 200 <= _response.status_code < 300:
|
|
@@ -341,6 +371,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
341
371
|
project_id: typing.Optional[str] = None,
|
|
342
372
|
organization_id: typing.Optional[str] = None,
|
|
343
373
|
request: EmbeddingModelConfigUpdate,
|
|
374
|
+
project_id: typing.Optional[str] = None,
|
|
344
375
|
) -> EmbeddingModelConfig:
|
|
345
376
|
"""
|
|
346
377
|
Update an embedding model config by ID.
|
|
@@ -353,6 +384,8 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
353
384
|
- organization_id: typing.Optional[str].
|
|
354
385
|
|
|
355
386
|
- request: EmbeddingModelConfigUpdate.
|
|
387
|
+
|
|
388
|
+
- project_id: typing.Optional[str].
|
|
356
389
|
"""
|
|
357
390
|
_response = await self._client_wrapper.httpx_client.request(
|
|
358
391
|
"PUT",
|
|
@@ -361,7 +394,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
361
394
|
),
|
|
362
395
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
363
396
|
json=jsonable_encoder(request),
|
|
364
|
-
headers=self._client_wrapper.get_headers(),
|
|
397
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
365
398
|
timeout=60,
|
|
366
399
|
)
|
|
367
400
|
if 200 <= _response.status_code < 300:
|
|
@@ -380,6 +413,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
380
413
|
*,
|
|
381
414
|
project_id: typing.Optional[str] = None,
|
|
382
415
|
organization_id: typing.Optional[str] = None,
|
|
416
|
+
project_id: typing.Optional[str] = None,
|
|
383
417
|
) -> None:
|
|
384
418
|
"""
|
|
385
419
|
Delete an embedding model config by ID.
|
|
@@ -390,6 +424,8 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
390
424
|
- project_id: typing.Optional[str].
|
|
391
425
|
|
|
392
426
|
- organization_id: typing.Optional[str].
|
|
427
|
+
|
|
428
|
+
- project_id: typing.Optional[str].
|
|
393
429
|
---
|
|
394
430
|
from llama_cloud.client import AsyncLlamaCloud
|
|
395
431
|
|
|
@@ -406,7 +442,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
406
442
|
f"{self._client_wrapper.get_base_url()}/", f"api/v1/embedding-model-configs/{embedding_model_config_id}"
|
|
407
443
|
),
|
|
408
444
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
409
|
-
headers=self._client_wrapper.get_headers(),
|
|
445
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
410
446
|
timeout=60,
|
|
411
447
|
)
|
|
412
448
|
if 200 <= _response.status_code < 300:
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
|
2
2
|
|
|
3
|
-
from .types import FileCreateFromUrlResourceInfoValue
|
|
3
|
+
from .types import FileCreateFromUrlResourceInfoValue
|
|
4
4
|
|
|
5
|
-
__all__ = ["FileCreateFromUrlResourceInfoValue"
|
|
5
|
+
__all__ = ["FileCreateFromUrlResourceInfoValue"]
|