llama-cloud 0.1.38__py3-none-any.whl → 0.1.39__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/resources/alpha/client.py +2 -8
- llama_cloud/resources/beta/client.py +30 -126
- llama_cloud/resources/chat_apps/client.py +8 -32
- llama_cloud/resources/classifier/client.py +8 -32
- llama_cloud/resources/data_sinks/client.py +8 -32
- llama_cloud/resources/data_sources/client.py +8 -32
- llama_cloud/resources/embedding_model_configs/client.py +12 -48
- llama_cloud/resources/files/client.py +42 -176
- llama_cloud/resources/jobs/client.py +2 -8
- llama_cloud/resources/llama_extract/client.py +32 -138
- llama_cloud/resources/organizations/client.py +4 -18
- llama_cloud/resources/parsing/client.py +4 -16
- llama_cloud/resources/pipelines/client.py +8 -32
- llama_cloud/resources/projects/client.py +18 -78
- llama_cloud/resources/reports/client.py +30 -126
- llama_cloud/resources/retrievers/client.py +12 -48
- llama_cloud/types/file.py +1 -1
- llama_cloud/types/pipeline_file.py +4 -4
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/METADATA +1 -1
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/RECORD +22 -22
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/LICENSE +0 -0
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/WHEEL +0 -0
|
@@ -31,19 +31,13 @@ class EmbeddingModelConfigsClient:
|
|
|
31
31
|
self._client_wrapper = client_wrapper
|
|
32
32
|
|
|
33
33
|
def list_embedding_model_configs(
|
|
34
|
-
self,
|
|
35
|
-
*,
|
|
36
|
-
project_id: typing.Optional[str] = None,
|
|
37
|
-
organization_id: typing.Optional[str] = None,
|
|
38
|
-
project_id: typing.Optional[str] = None,
|
|
34
|
+
self, *, project_id: typing.Optional[str] = None, organization_id: typing.Optional[str] = None
|
|
39
35
|
) -> typing.List[EmbeddingModelConfig]:
|
|
40
36
|
"""
|
|
41
37
|
Parameters:
|
|
42
38
|
- project_id: typing.Optional[str].
|
|
43
39
|
|
|
44
40
|
- organization_id: typing.Optional[str].
|
|
45
|
-
|
|
46
|
-
- project_id: typing.Optional[str].
|
|
47
41
|
---
|
|
48
42
|
from llama_cloud.client import LlamaCloud
|
|
49
43
|
|
|
@@ -56,7 +50,7 @@ class EmbeddingModelConfigsClient:
|
|
|
56
50
|
"GET",
|
|
57
51
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
58
52
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
59
|
-
headers=
|
|
53
|
+
headers=self._client_wrapper.get_headers(),
|
|
60
54
|
timeout=60,
|
|
61
55
|
)
|
|
62
56
|
if 200 <= _response.status_code < 300:
|
|
@@ -76,7 +70,6 @@ class EmbeddingModelConfigsClient:
|
|
|
76
70
|
organization_id: typing.Optional[str] = None,
|
|
77
71
|
name: str,
|
|
78
72
|
embedding_config: EmbeddingModelConfigCreateEmbeddingConfig,
|
|
79
|
-
project_id: typing.Optional[str] = None,
|
|
80
73
|
) -> EmbeddingModelConfig:
|
|
81
74
|
"""
|
|
82
75
|
Create a new embedding model configuration within a specified project.
|
|
@@ -89,15 +82,13 @@ class EmbeddingModelConfigsClient:
|
|
|
89
82
|
- name: str. The name of the embedding model config.
|
|
90
83
|
|
|
91
84
|
- embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
|
|
92
|
-
|
|
93
|
-
- project_id: typing.Optional[str].
|
|
94
85
|
"""
|
|
95
86
|
_response = self._client_wrapper.httpx_client.request(
|
|
96
87
|
"POST",
|
|
97
88
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
98
89
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
99
90
|
json=jsonable_encoder({"name": name, "embedding_config": embedding_config}),
|
|
100
|
-
headers=
|
|
91
|
+
headers=self._client_wrapper.get_headers(),
|
|
101
92
|
timeout=60,
|
|
102
93
|
)
|
|
103
94
|
if 200 <= _response.status_code < 300:
|
|
@@ -116,7 +107,6 @@ class EmbeddingModelConfigsClient:
|
|
|
116
107
|
project_id: typing.Optional[str] = None,
|
|
117
108
|
organization_id: typing.Optional[str] = None,
|
|
118
109
|
request: EmbeddingModelConfigUpdate,
|
|
119
|
-
project_id: typing.Optional[str] = None,
|
|
120
110
|
) -> EmbeddingModelConfig:
|
|
121
111
|
"""
|
|
122
112
|
Upserts an embedding model config.
|
|
@@ -128,15 +118,13 @@ class EmbeddingModelConfigsClient:
|
|
|
128
118
|
- organization_id: typing.Optional[str].
|
|
129
119
|
|
|
130
120
|
- request: EmbeddingModelConfigUpdate.
|
|
131
|
-
|
|
132
|
-
- project_id: typing.Optional[str].
|
|
133
121
|
"""
|
|
134
122
|
_response = self._client_wrapper.httpx_client.request(
|
|
135
123
|
"PUT",
|
|
136
124
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
137
125
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
138
126
|
json=jsonable_encoder(request),
|
|
139
|
-
headers=
|
|
127
|
+
headers=self._client_wrapper.get_headers(),
|
|
140
128
|
timeout=60,
|
|
141
129
|
)
|
|
142
130
|
if 200 <= _response.status_code < 300:
|
|
@@ -156,7 +144,6 @@ class EmbeddingModelConfigsClient:
|
|
|
156
144
|
project_id: typing.Optional[str] = None,
|
|
157
145
|
organization_id: typing.Optional[str] = None,
|
|
158
146
|
request: EmbeddingModelConfigUpdate,
|
|
159
|
-
project_id: typing.Optional[str] = None,
|
|
160
147
|
) -> EmbeddingModelConfig:
|
|
161
148
|
"""
|
|
162
149
|
Update an embedding model config by ID.
|
|
@@ -169,8 +156,6 @@ class EmbeddingModelConfigsClient:
|
|
|
169
156
|
- organization_id: typing.Optional[str].
|
|
170
157
|
|
|
171
158
|
- request: EmbeddingModelConfigUpdate.
|
|
172
|
-
|
|
173
|
-
- project_id: typing.Optional[str].
|
|
174
159
|
"""
|
|
175
160
|
_response = self._client_wrapper.httpx_client.request(
|
|
176
161
|
"PUT",
|
|
@@ -179,7 +164,7 @@ class EmbeddingModelConfigsClient:
|
|
|
179
164
|
),
|
|
180
165
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
181
166
|
json=jsonable_encoder(request),
|
|
182
|
-
headers=
|
|
167
|
+
headers=self._client_wrapper.get_headers(),
|
|
183
168
|
timeout=60,
|
|
184
169
|
)
|
|
185
170
|
if 200 <= _response.status_code < 300:
|
|
@@ -198,7 +183,6 @@ class EmbeddingModelConfigsClient:
|
|
|
198
183
|
*,
|
|
199
184
|
project_id: typing.Optional[str] = None,
|
|
200
185
|
organization_id: typing.Optional[str] = None,
|
|
201
|
-
project_id: typing.Optional[str] = None,
|
|
202
186
|
) -> None:
|
|
203
187
|
"""
|
|
204
188
|
Delete an embedding model config by ID.
|
|
@@ -209,8 +193,6 @@ class EmbeddingModelConfigsClient:
|
|
|
209
193
|
- project_id: typing.Optional[str].
|
|
210
194
|
|
|
211
195
|
- organization_id: typing.Optional[str].
|
|
212
|
-
|
|
213
|
-
- project_id: typing.Optional[str].
|
|
214
196
|
---
|
|
215
197
|
from llama_cloud.client import LlamaCloud
|
|
216
198
|
|
|
@@ -227,7 +209,7 @@ class EmbeddingModelConfigsClient:
|
|
|
227
209
|
f"{self._client_wrapper.get_base_url()}/", f"api/v1/embedding-model-configs/{embedding_model_config_id}"
|
|
228
210
|
),
|
|
229
211
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
230
|
-
headers=
|
|
212
|
+
headers=self._client_wrapper.get_headers(),
|
|
231
213
|
timeout=60,
|
|
232
214
|
)
|
|
233
215
|
if 200 <= _response.status_code < 300:
|
|
@@ -246,19 +228,13 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
246
228
|
self._client_wrapper = client_wrapper
|
|
247
229
|
|
|
248
230
|
async def list_embedding_model_configs(
|
|
249
|
-
self,
|
|
250
|
-
*,
|
|
251
|
-
project_id: typing.Optional[str] = None,
|
|
252
|
-
organization_id: typing.Optional[str] = None,
|
|
253
|
-
project_id: typing.Optional[str] = None,
|
|
231
|
+
self, *, project_id: typing.Optional[str] = None, organization_id: typing.Optional[str] = None
|
|
254
232
|
) -> typing.List[EmbeddingModelConfig]:
|
|
255
233
|
"""
|
|
256
234
|
Parameters:
|
|
257
235
|
- project_id: typing.Optional[str].
|
|
258
236
|
|
|
259
237
|
- organization_id: typing.Optional[str].
|
|
260
|
-
|
|
261
|
-
- project_id: typing.Optional[str].
|
|
262
238
|
---
|
|
263
239
|
from llama_cloud.client import AsyncLlamaCloud
|
|
264
240
|
|
|
@@ -271,7 +247,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
271
247
|
"GET",
|
|
272
248
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
273
249
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
274
|
-
headers=
|
|
250
|
+
headers=self._client_wrapper.get_headers(),
|
|
275
251
|
timeout=60,
|
|
276
252
|
)
|
|
277
253
|
if 200 <= _response.status_code < 300:
|
|
@@ -291,7 +267,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
291
267
|
organization_id: typing.Optional[str] = None,
|
|
292
268
|
name: str,
|
|
293
269
|
embedding_config: EmbeddingModelConfigCreateEmbeddingConfig,
|
|
294
|
-
project_id: typing.Optional[str] = None,
|
|
295
270
|
) -> EmbeddingModelConfig:
|
|
296
271
|
"""
|
|
297
272
|
Create a new embedding model configuration within a specified project.
|
|
@@ -304,15 +279,13 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
304
279
|
- name: str. The name of the embedding model config.
|
|
305
280
|
|
|
306
281
|
- embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
|
|
307
|
-
|
|
308
|
-
- project_id: typing.Optional[str].
|
|
309
282
|
"""
|
|
310
283
|
_response = await self._client_wrapper.httpx_client.request(
|
|
311
284
|
"POST",
|
|
312
285
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
313
286
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
314
287
|
json=jsonable_encoder({"name": name, "embedding_config": embedding_config}),
|
|
315
|
-
headers=
|
|
288
|
+
headers=self._client_wrapper.get_headers(),
|
|
316
289
|
timeout=60,
|
|
317
290
|
)
|
|
318
291
|
if 200 <= _response.status_code < 300:
|
|
@@ -331,7 +304,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
331
304
|
project_id: typing.Optional[str] = None,
|
|
332
305
|
organization_id: typing.Optional[str] = None,
|
|
333
306
|
request: EmbeddingModelConfigUpdate,
|
|
334
|
-
project_id: typing.Optional[str] = None,
|
|
335
307
|
) -> EmbeddingModelConfig:
|
|
336
308
|
"""
|
|
337
309
|
Upserts an embedding model config.
|
|
@@ -343,15 +315,13 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
343
315
|
- organization_id: typing.Optional[str].
|
|
344
316
|
|
|
345
317
|
- request: EmbeddingModelConfigUpdate.
|
|
346
|
-
|
|
347
|
-
- project_id: typing.Optional[str].
|
|
348
318
|
"""
|
|
349
319
|
_response = await self._client_wrapper.httpx_client.request(
|
|
350
320
|
"PUT",
|
|
351
321
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/embedding-model-configs"),
|
|
352
322
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
353
323
|
json=jsonable_encoder(request),
|
|
354
|
-
headers=
|
|
324
|
+
headers=self._client_wrapper.get_headers(),
|
|
355
325
|
timeout=60,
|
|
356
326
|
)
|
|
357
327
|
if 200 <= _response.status_code < 300:
|
|
@@ -371,7 +341,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
371
341
|
project_id: typing.Optional[str] = None,
|
|
372
342
|
organization_id: typing.Optional[str] = None,
|
|
373
343
|
request: EmbeddingModelConfigUpdate,
|
|
374
|
-
project_id: typing.Optional[str] = None,
|
|
375
344
|
) -> EmbeddingModelConfig:
|
|
376
345
|
"""
|
|
377
346
|
Update an embedding model config by ID.
|
|
@@ -384,8 +353,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
384
353
|
- organization_id: typing.Optional[str].
|
|
385
354
|
|
|
386
355
|
- request: EmbeddingModelConfigUpdate.
|
|
387
|
-
|
|
388
|
-
- project_id: typing.Optional[str].
|
|
389
356
|
"""
|
|
390
357
|
_response = await self._client_wrapper.httpx_client.request(
|
|
391
358
|
"PUT",
|
|
@@ -394,7 +361,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
394
361
|
),
|
|
395
362
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
396
363
|
json=jsonable_encoder(request),
|
|
397
|
-
headers=
|
|
364
|
+
headers=self._client_wrapper.get_headers(),
|
|
398
365
|
timeout=60,
|
|
399
366
|
)
|
|
400
367
|
if 200 <= _response.status_code < 300:
|
|
@@ -413,7 +380,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
413
380
|
*,
|
|
414
381
|
project_id: typing.Optional[str] = None,
|
|
415
382
|
organization_id: typing.Optional[str] = None,
|
|
416
|
-
project_id: typing.Optional[str] = None,
|
|
417
383
|
) -> None:
|
|
418
384
|
"""
|
|
419
385
|
Delete an embedding model config by ID.
|
|
@@ -424,8 +390,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
424
390
|
- project_id: typing.Optional[str].
|
|
425
391
|
|
|
426
392
|
- organization_id: typing.Optional[str].
|
|
427
|
-
|
|
428
|
-
- project_id: typing.Optional[str].
|
|
429
393
|
---
|
|
430
394
|
from llama_cloud.client import AsyncLlamaCloud
|
|
431
395
|
|
|
@@ -442,7 +406,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
442
406
|
f"{self._client_wrapper.get_base_url()}/", f"api/v1/embedding-model-configs/{embedding_model_config_id}"
|
|
443
407
|
),
|
|
444
408
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
445
|
-
headers=
|
|
409
|
+
headers=self._client_wrapper.get_headers(),
|
|
446
410
|
timeout=60,
|
|
447
411
|
)
|
|
448
412
|
if 200 <= _response.status_code < 300:
|