llama-cloud 0.0.2__py3-none-any.whl → 0.0.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +2 -0
- llama_cloud/client.py +16 -4
- llama_cloud/core/jsonable_encoder.py +3 -0
- llama_cloud/environment.py +7 -0
- llama_cloud/resources/api_keys/client.py +11 -16
- llama_cloud/resources/billing/client.py +9 -12
- llama_cloud/resources/component_definitions/client.py +9 -12
- llama_cloud/resources/data_sinks/client.py +21 -30
- llama_cloud/resources/data_sources/client.py +21 -30
- llama_cloud/resources/deprecated/client.py +27 -48
- llama_cloud/resources/evals/client.py +27 -44
- llama_cloud/resources/files/client.py +15 -24
- llama_cloud/resources/parsing/client.py +27 -48
- llama_cloud/resources/pipelines/client.py +73 -128
- llama_cloud/resources/projects/client.py +43 -72
- llama_cloud/types/api_key.py +3 -0
- llama_cloud/types/azure_open_ai_embedding.py +3 -0
- llama_cloud/types/base.py +3 -0
- llama_cloud/types/base_prompt_template.py +3 -0
- llama_cloud/types/bedrock_embedding.py +3 -0
- llama_cloud/types/chat_message.py +3 -0
- llama_cloud/types/cloud_az_storage_blob_data_source.py +3 -0
- llama_cloud/types/cloud_chroma_vector_store.py +3 -0
- llama_cloud/types/cloud_document.py +3 -0
- llama_cloud/types/cloud_document_create.py +3 -0
- llama_cloud/types/cloud_gcs_data_source.py +3 -0
- llama_cloud/types/cloud_google_drive_data_source.py +3 -0
- llama_cloud/types/cloud_one_drive_data_source.py +3 -0
- llama_cloud/types/cloud_pinecone_vector_store.py +3 -0
- llama_cloud/types/cloud_postgres_vector_store.py +3 -0
- llama_cloud/types/cloud_qdrant_vector_store.py +3 -0
- llama_cloud/types/cloud_s_3_data_source.py +3 -0
- llama_cloud/types/cloud_sharepoint_data_source.py +3 -0
- llama_cloud/types/cloud_weaviate_vector_store.py +3 -0
- llama_cloud/types/code_splitter.py +3 -0
- llama_cloud/types/cohere_embedding.py +3 -0
- llama_cloud/types/configurable_transformation_definition.py +3 -0
- llama_cloud/types/configured_transformation_item.py +3 -0
- llama_cloud/types/data_sink.py +3 -0
- llama_cloud/types/data_sink_create.py +3 -0
- llama_cloud/types/data_sink_definition.py +3 -0
- llama_cloud/types/data_source.py +3 -0
- llama_cloud/types/data_source_create.py +3 -0
- llama_cloud/types/data_source_definition.py +3 -0
- llama_cloud/types/eval_dataset.py +3 -0
- llama_cloud/types/eval_dataset_job_params.py +3 -0
- llama_cloud/types/eval_dataset_job_record.py +3 -0
- llama_cloud/types/eval_execution_params.py +3 -0
- llama_cloud/types/eval_execution_params_override.py +3 -0
- llama_cloud/types/eval_llm_model_data.py +3 -0
- llama_cloud/types/eval_question.py +3 -0
- llama_cloud/types/eval_question_create.py +3 -0
- llama_cloud/types/eval_question_result.py +3 -0
- llama_cloud/types/file.py +3 -0
- llama_cloud/types/gemini_embedding.py +3 -0
- llama_cloud/types/html_node_parser.py +3 -0
- llama_cloud/types/http_validation_error.py +3 -0
- llama_cloud/types/hugging_face_inference_api_embedding.py +3 -0
- llama_cloud/types/json_node_parser.py +3 -0
- llama_cloud/types/llm.py +3 -0
- llama_cloud/types/local_eval.py +3 -0
- llama_cloud/types/local_eval_results.py +3 -0
- llama_cloud/types/local_eval_sets.py +3 -0
- llama_cloud/types/markdown_element_node_parser.py +3 -0
- llama_cloud/types/markdown_node_parser.py +3 -0
- llama_cloud/types/metadata_filter.py +3 -0
- llama_cloud/types/metadata_filters.py +3 -0
- llama_cloud/types/metric_result.py +3 -0
- llama_cloud/types/node_parser.py +3 -0
- llama_cloud/types/open_ai_embedding.py +3 -0
- llama_cloud/types/parsing_history_item.py +3 -0
- llama_cloud/types/parsing_job.py +3 -0
- llama_cloud/types/parsing_job_json_result.py +3 -0
- llama_cloud/types/parsing_job_markdown_result.py +3 -0
- llama_cloud/types/parsing_job_text_result.py +3 -0
- llama_cloud/types/parsing_usage.py +3 -0
- llama_cloud/types/pipeline.py +3 -0
- llama_cloud/types/pipeline_create.py +3 -0
- llama_cloud/types/pipeline_data_source.py +3 -0
- llama_cloud/types/pipeline_data_source_create.py +3 -0
- llama_cloud/types/pipeline_deployment.py +3 -0
- llama_cloud/types/pipeline_file.py +3 -0
- llama_cloud/types/pipeline_file_create.py +3 -0
- llama_cloud/types/pipeline_file_status_response.py +3 -0
- llama_cloud/types/preset_retrieval_params.py +3 -0
- llama_cloud/types/presigned_url.py +3 -0
- llama_cloud/types/project.py +3 -0
- llama_cloud/types/project_create.py +3 -0
- llama_cloud/types/prompt_mixin_prompts.py +3 -0
- llama_cloud/types/prompt_spec.py +3 -0
- llama_cloud/types/related_node_info.py +3 -0
- llama_cloud/types/retrieve_results.py +3 -0
- llama_cloud/types/sentence_splitter.py +3 -0
- llama_cloud/types/simple_file_node_parser.py +3 -0
- llama_cloud/types/supported_eval_llm_model.py +3 -0
- llama_cloud/types/text_node.py +3 -0
- llama_cloud/types/text_node_with_score.py +3 -0
- llama_cloud/types/token_text_splitter.py +3 -0
- llama_cloud/types/validation_error.py +3 -0
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.4.dist-info}/METADATA +1 -1
- llama_cloud-0.0.4.dist-info/RECORD +174 -0
- llama_cloud-0.0.2.dist-info/RECORD +0 -173
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.4.dist-info}/LICENSE +0 -0
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.4.dist-info}/WHEEL +0 -0
|
@@ -17,6 +17,9 @@ from .types.data_source_update_component import DataSourceUpdateComponent
|
|
|
17
17
|
from .types.data_source_update_custom_metadata_value import DataSourceUpdateCustomMetadataValue
|
|
18
18
|
|
|
19
19
|
try:
|
|
20
|
+
import pydantic
|
|
21
|
+
if pydantic.__version__.startswith("1."):
|
|
22
|
+
raise ImportError
|
|
20
23
|
import pydantic.v1 as pydantic # type: ignore
|
|
21
24
|
except ImportError:
|
|
22
25
|
import pydantic # type: ignore
|
|
@@ -37,11 +40,10 @@ class DataSourcesClient:
|
|
|
37
40
|
Parameters:
|
|
38
41
|
- project_id: typing.Optional[str].
|
|
39
42
|
---
|
|
40
|
-
from
|
|
43
|
+
from llama_cloud.client import LlamaCloud
|
|
41
44
|
|
|
42
45
|
client = LlamaCloud(
|
|
43
46
|
token="YOUR_TOKEN",
|
|
44
|
-
base_url="https://yourhost.com/path/to/api",
|
|
45
47
|
)
|
|
46
48
|
client.data_sources.list_data_sources()
|
|
47
49
|
"""
|
|
@@ -71,12 +73,11 @@ class DataSourcesClient:
|
|
|
71
73
|
|
|
72
74
|
- request: DataSourceCreate.
|
|
73
75
|
---
|
|
74
|
-
from
|
|
75
|
-
from
|
|
76
|
+
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
77
|
+
from llama_cloud.client import LlamaCloud
|
|
76
78
|
|
|
77
79
|
client = LlamaCloud(
|
|
78
80
|
token="YOUR_TOKEN",
|
|
79
|
-
base_url="https://yourhost.com/path/to/api",
|
|
80
81
|
)
|
|
81
82
|
client.data_sources.create_data_source(
|
|
82
83
|
request=DataSourceCreate(
|
|
@@ -113,12 +114,11 @@ class DataSourcesClient:
|
|
|
113
114
|
|
|
114
115
|
- request: DataSourceCreate.
|
|
115
116
|
---
|
|
116
|
-
from
|
|
117
|
-
from
|
|
117
|
+
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
118
|
+
from llama_cloud.client import LlamaCloud
|
|
118
119
|
|
|
119
120
|
client = LlamaCloud(
|
|
120
121
|
token="YOUR_TOKEN",
|
|
121
|
-
base_url="https://yourhost.com/path/to/api",
|
|
122
122
|
)
|
|
123
123
|
client.data_sources.upsert_data_source(
|
|
124
124
|
request=DataSourceCreate(
|
|
@@ -152,11 +152,10 @@ class DataSourcesClient:
|
|
|
152
152
|
Parameters:
|
|
153
153
|
- data_source_id: str.
|
|
154
154
|
---
|
|
155
|
-
from
|
|
155
|
+
from llama_cloud.client import LlamaCloud
|
|
156
156
|
|
|
157
157
|
client = LlamaCloud(
|
|
158
158
|
token="YOUR_TOKEN",
|
|
159
|
-
base_url="https://yourhost.com/path/to/api",
|
|
160
159
|
)
|
|
161
160
|
client.data_sources.get_data_source(
|
|
162
161
|
data_source_id="string",
|
|
@@ -201,12 +200,11 @@ class DataSourcesClient:
|
|
|
201
200
|
|
|
202
201
|
- component: typing.Optional[DataSourceUpdateComponent].
|
|
203
202
|
---
|
|
204
|
-
from
|
|
205
|
-
from
|
|
203
|
+
from llama_cloud import ConfigurableDataSourceNames
|
|
204
|
+
from llama_cloud.client import LlamaCloud
|
|
206
205
|
|
|
207
206
|
client = LlamaCloud(
|
|
208
207
|
token="YOUR_TOKEN",
|
|
209
|
-
base_url="https://yourhost.com/path/to/api",
|
|
210
208
|
)
|
|
211
209
|
client.data_sources.update_data_source(
|
|
212
210
|
data_source_id="string",
|
|
@@ -244,11 +242,10 @@ class DataSourcesClient:
|
|
|
244
242
|
Parameters:
|
|
245
243
|
- data_source_id: str.
|
|
246
244
|
---
|
|
247
|
-
from
|
|
245
|
+
from llama_cloud.client import LlamaCloud
|
|
248
246
|
|
|
249
247
|
client = LlamaCloud(
|
|
250
248
|
token="YOUR_TOKEN",
|
|
251
|
-
base_url="https://yourhost.com/path/to/api",
|
|
252
249
|
)
|
|
253
250
|
client.data_sources.delete_data_source(
|
|
254
251
|
data_source_id="string",
|
|
@@ -283,11 +280,10 @@ class AsyncDataSourcesClient:
|
|
|
283
280
|
Parameters:
|
|
284
281
|
- project_id: typing.Optional[str].
|
|
285
282
|
---
|
|
286
|
-
from
|
|
283
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
287
284
|
|
|
288
285
|
client = AsyncLlamaCloud(
|
|
289
286
|
token="YOUR_TOKEN",
|
|
290
|
-
base_url="https://yourhost.com/path/to/api",
|
|
291
287
|
)
|
|
292
288
|
await client.data_sources.list_data_sources()
|
|
293
289
|
"""
|
|
@@ -319,12 +315,11 @@ class AsyncDataSourcesClient:
|
|
|
319
315
|
|
|
320
316
|
- request: DataSourceCreate.
|
|
321
317
|
---
|
|
322
|
-
from
|
|
323
|
-
from
|
|
318
|
+
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
319
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
324
320
|
|
|
325
321
|
client = AsyncLlamaCloud(
|
|
326
322
|
token="YOUR_TOKEN",
|
|
327
|
-
base_url="https://yourhost.com/path/to/api",
|
|
328
323
|
)
|
|
329
324
|
await client.data_sources.create_data_source(
|
|
330
325
|
request=DataSourceCreate(
|
|
@@ -363,12 +358,11 @@ class AsyncDataSourcesClient:
|
|
|
363
358
|
|
|
364
359
|
- request: DataSourceCreate.
|
|
365
360
|
---
|
|
366
|
-
from
|
|
367
|
-
from
|
|
361
|
+
from llama_cloud import ConfigurableDataSourceNames, DataSourceCreate
|
|
362
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
368
363
|
|
|
369
364
|
client = AsyncLlamaCloud(
|
|
370
365
|
token="YOUR_TOKEN",
|
|
371
|
-
base_url="https://yourhost.com/path/to/api",
|
|
372
366
|
)
|
|
373
367
|
await client.data_sources.upsert_data_source(
|
|
374
368
|
request=DataSourceCreate(
|
|
@@ -402,11 +396,10 @@ class AsyncDataSourcesClient:
|
|
|
402
396
|
Parameters:
|
|
403
397
|
- data_source_id: str.
|
|
404
398
|
---
|
|
405
|
-
from
|
|
399
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
406
400
|
|
|
407
401
|
client = AsyncLlamaCloud(
|
|
408
402
|
token="YOUR_TOKEN",
|
|
409
|
-
base_url="https://yourhost.com/path/to/api",
|
|
410
403
|
)
|
|
411
404
|
await client.data_sources.get_data_source(
|
|
412
405
|
data_source_id="string",
|
|
@@ -451,12 +444,11 @@ class AsyncDataSourcesClient:
|
|
|
451
444
|
|
|
452
445
|
- component: typing.Optional[DataSourceUpdateComponent].
|
|
453
446
|
---
|
|
454
|
-
from
|
|
455
|
-
from
|
|
447
|
+
from llama_cloud import ConfigurableDataSourceNames
|
|
448
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
456
449
|
|
|
457
450
|
client = AsyncLlamaCloud(
|
|
458
451
|
token="YOUR_TOKEN",
|
|
459
|
-
base_url="https://yourhost.com/path/to/api",
|
|
460
452
|
)
|
|
461
453
|
await client.data_sources.update_data_source(
|
|
462
454
|
data_source_id="string",
|
|
@@ -494,11 +486,10 @@ class AsyncDataSourcesClient:
|
|
|
494
486
|
Parameters:
|
|
495
487
|
- data_source_id: str.
|
|
496
488
|
---
|
|
497
|
-
from
|
|
489
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
498
490
|
|
|
499
491
|
client = AsyncLlamaCloud(
|
|
500
492
|
token="YOUR_TOKEN",
|
|
501
|
-
base_url="https://yourhost.com/path/to/api",
|
|
502
493
|
)
|
|
503
494
|
await client.data_sources.delete_data_source(
|
|
504
495
|
data_source_id="string",
|
|
@@ -20,6 +20,9 @@ from ...types.parsing_usage import ParsingUsage
|
|
|
20
20
|
from ...types.presigned_url import PresignedUrl
|
|
21
21
|
|
|
22
22
|
try:
|
|
23
|
+
import pydantic
|
|
24
|
+
if pydantic.__version__.startswith("1."):
|
|
25
|
+
raise ImportError
|
|
23
26
|
import pydantic.v1 as pydantic # type: ignore
|
|
24
27
|
except ImportError:
|
|
25
28
|
import pydantic # type: ignore
|
|
@@ -41,11 +44,10 @@ class DeprecatedClient:
|
|
|
41
44
|
|
|
42
45
|
- name: str.
|
|
43
46
|
---
|
|
44
|
-
from
|
|
47
|
+
from llama_cloud.client import LlamaCloud
|
|
45
48
|
|
|
46
49
|
client = LlamaCloud(
|
|
47
50
|
token="YOUR_TOKEN",
|
|
48
|
-
base_url="https://yourhost.com/path/to/api",
|
|
49
51
|
)
|
|
50
52
|
client.deprecated.get_job_image_result(
|
|
51
53
|
job_id="string",
|
|
@@ -75,11 +77,10 @@ class DeprecatedClient:
|
|
|
75
77
|
Get a list of supported file extensions
|
|
76
78
|
|
|
77
79
|
---
|
|
78
|
-
from
|
|
80
|
+
from llama_cloud.client import LlamaCloud
|
|
79
81
|
|
|
80
82
|
client = LlamaCloud(
|
|
81
83
|
token="YOUR_TOKEN",
|
|
82
|
-
base_url="https://yourhost.com/path/to/api",
|
|
83
84
|
)
|
|
84
85
|
client.deprecated.get_supported_file_extensions()
|
|
85
86
|
"""
|
|
@@ -174,11 +175,10 @@ class DeprecatedClient:
|
|
|
174
175
|
Get parsing usage for user
|
|
175
176
|
|
|
176
177
|
---
|
|
177
|
-
from
|
|
178
|
+
from llama_cloud.client import LlamaCloud
|
|
178
179
|
|
|
179
180
|
client = LlamaCloud(
|
|
180
181
|
token="YOUR_TOKEN",
|
|
181
|
-
base_url="https://yourhost.com/path/to/api",
|
|
182
182
|
)
|
|
183
183
|
client.deprecated.usage()
|
|
184
184
|
"""
|
|
@@ -205,11 +205,10 @@ class DeprecatedClient:
|
|
|
205
205
|
Parameters:
|
|
206
206
|
- job_id: str.
|
|
207
207
|
---
|
|
208
|
-
from
|
|
208
|
+
from llama_cloud.client import LlamaCloud
|
|
209
209
|
|
|
210
210
|
client = LlamaCloud(
|
|
211
211
|
token="YOUR_TOKEN",
|
|
212
|
-
base_url="https://yourhost.com/path/to/api",
|
|
213
212
|
)
|
|
214
213
|
client.deprecated.get_job(
|
|
215
214
|
job_id="string",
|
|
@@ -238,11 +237,10 @@ class DeprecatedClient:
|
|
|
238
237
|
Parameters:
|
|
239
238
|
- job_id: str.
|
|
240
239
|
---
|
|
241
|
-
from
|
|
240
|
+
from llama_cloud.client import LlamaCloud
|
|
242
241
|
|
|
243
242
|
client = LlamaCloud(
|
|
244
243
|
token="YOUR_TOKEN",
|
|
245
|
-
base_url="https://yourhost.com/path/to/api",
|
|
246
244
|
)
|
|
247
245
|
client.deprecated.get_job_text_result(
|
|
248
246
|
job_id="string",
|
|
@@ -271,11 +269,10 @@ class DeprecatedClient:
|
|
|
271
269
|
Parameters:
|
|
272
270
|
- job_id: str.
|
|
273
271
|
---
|
|
274
|
-
from
|
|
272
|
+
from llama_cloud.client import LlamaCloud
|
|
275
273
|
|
|
276
274
|
client = LlamaCloud(
|
|
277
275
|
token="YOUR_TOKEN",
|
|
278
|
-
base_url="https://yourhost.com/path/to/api",
|
|
279
276
|
)
|
|
280
277
|
client.deprecated.get_job_raw_text_result(
|
|
281
278
|
job_id="string",
|
|
@@ -306,11 +303,10 @@ class DeprecatedClient:
|
|
|
306
303
|
Parameters:
|
|
307
304
|
- job_id: str.
|
|
308
305
|
---
|
|
309
|
-
from
|
|
306
|
+
from llama_cloud.client import LlamaCloud
|
|
310
307
|
|
|
311
308
|
client = LlamaCloud(
|
|
312
309
|
token="YOUR_TOKEN",
|
|
313
|
-
base_url="https://yourhost.com/path/to/api",
|
|
314
310
|
)
|
|
315
311
|
client.deprecated.get_job_result(
|
|
316
312
|
job_id="string",
|
|
@@ -341,11 +337,10 @@ class DeprecatedClient:
|
|
|
341
337
|
Parameters:
|
|
342
338
|
- job_id: str.
|
|
343
339
|
---
|
|
344
|
-
from
|
|
340
|
+
from llama_cloud.client import LlamaCloud
|
|
345
341
|
|
|
346
342
|
client = LlamaCloud(
|
|
347
343
|
token="YOUR_TOKEN",
|
|
348
|
-
base_url="https://yourhost.com/path/to/api",
|
|
349
344
|
)
|
|
350
345
|
client.deprecated.get_job_raw_md_result(
|
|
351
346
|
job_id="string",
|
|
@@ -376,11 +371,10 @@ class DeprecatedClient:
|
|
|
376
371
|
Parameters:
|
|
377
372
|
- job_id: str.
|
|
378
373
|
---
|
|
379
|
-
from
|
|
374
|
+
from llama_cloud.client import LlamaCloud
|
|
380
375
|
|
|
381
376
|
client = LlamaCloud(
|
|
382
377
|
token="YOUR_TOKEN",
|
|
383
|
-
base_url="https://yourhost.com/path/to/api",
|
|
384
378
|
)
|
|
385
379
|
client.deprecated.get_job_json_result(
|
|
386
380
|
job_id="string",
|
|
@@ -409,11 +403,10 @@ class DeprecatedClient:
|
|
|
409
403
|
Parameters:
|
|
410
404
|
- job_id: str.
|
|
411
405
|
---
|
|
412
|
-
from
|
|
406
|
+
from llama_cloud.client import LlamaCloud
|
|
413
407
|
|
|
414
408
|
client = LlamaCloud(
|
|
415
409
|
token="YOUR_TOKEN",
|
|
416
|
-
base_url="https://yourhost.com/path/to/api",
|
|
417
410
|
)
|
|
418
411
|
client.deprecated.get_job_json_raw_result(
|
|
419
412
|
job_id="string",
|
|
@@ -442,11 +435,10 @@ class DeprecatedClient:
|
|
|
442
435
|
Get parsing history for user
|
|
443
436
|
|
|
444
437
|
---
|
|
445
|
-
from
|
|
438
|
+
from llama_cloud.client import LlamaCloud
|
|
446
439
|
|
|
447
440
|
client = LlamaCloud(
|
|
448
441
|
token="YOUR_TOKEN",
|
|
449
|
-
base_url="https://yourhost.com/path/to/api",
|
|
450
442
|
)
|
|
451
443
|
client.deprecated.get_parsing_history_result()
|
|
452
444
|
"""
|
|
@@ -475,11 +467,10 @@ class DeprecatedClient:
|
|
|
475
467
|
|
|
476
468
|
- filename: str.
|
|
477
469
|
---
|
|
478
|
-
from
|
|
470
|
+
from llama_cloud.client import LlamaCloud
|
|
479
471
|
|
|
480
472
|
client = LlamaCloud(
|
|
481
473
|
token="YOUR_TOKEN",
|
|
482
|
-
base_url="https://yourhost.com/path/to/api",
|
|
483
474
|
)
|
|
484
475
|
client.deprecated.generate_presigned_url(
|
|
485
476
|
job_id="string",
|
|
@@ -518,11 +509,10 @@ class AsyncDeprecatedClient:
|
|
|
518
509
|
|
|
519
510
|
- name: str.
|
|
520
511
|
---
|
|
521
|
-
from
|
|
512
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
522
513
|
|
|
523
514
|
client = AsyncLlamaCloud(
|
|
524
515
|
token="YOUR_TOKEN",
|
|
525
|
-
base_url="https://yourhost.com/path/to/api",
|
|
526
516
|
)
|
|
527
517
|
await client.deprecated.get_job_image_result(
|
|
528
518
|
job_id="string",
|
|
@@ -552,11 +542,10 @@ class AsyncDeprecatedClient:
|
|
|
552
542
|
Get a list of supported file extensions
|
|
553
543
|
|
|
554
544
|
---
|
|
555
|
-
from
|
|
545
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
556
546
|
|
|
557
547
|
client = AsyncLlamaCloud(
|
|
558
548
|
token="YOUR_TOKEN",
|
|
559
|
-
base_url="https://yourhost.com/path/to/api",
|
|
560
549
|
)
|
|
561
550
|
await client.deprecated.get_supported_file_extensions()
|
|
562
551
|
"""
|
|
@@ -651,11 +640,10 @@ class AsyncDeprecatedClient:
|
|
|
651
640
|
Get parsing usage for user
|
|
652
641
|
|
|
653
642
|
---
|
|
654
|
-
from
|
|
643
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
655
644
|
|
|
656
645
|
client = AsyncLlamaCloud(
|
|
657
646
|
token="YOUR_TOKEN",
|
|
658
|
-
base_url="https://yourhost.com/path/to/api",
|
|
659
647
|
)
|
|
660
648
|
await client.deprecated.usage()
|
|
661
649
|
"""
|
|
@@ -682,11 +670,10 @@ class AsyncDeprecatedClient:
|
|
|
682
670
|
Parameters:
|
|
683
671
|
- job_id: str.
|
|
684
672
|
---
|
|
685
|
-
from
|
|
673
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
686
674
|
|
|
687
675
|
client = AsyncLlamaCloud(
|
|
688
676
|
token="YOUR_TOKEN",
|
|
689
|
-
base_url="https://yourhost.com/path/to/api",
|
|
690
677
|
)
|
|
691
678
|
await client.deprecated.get_job(
|
|
692
679
|
job_id="string",
|
|
@@ -715,11 +702,10 @@ class AsyncDeprecatedClient:
|
|
|
715
702
|
Parameters:
|
|
716
703
|
- job_id: str.
|
|
717
704
|
---
|
|
718
|
-
from
|
|
705
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
719
706
|
|
|
720
707
|
client = AsyncLlamaCloud(
|
|
721
708
|
token="YOUR_TOKEN",
|
|
722
|
-
base_url="https://yourhost.com/path/to/api",
|
|
723
709
|
)
|
|
724
710
|
await client.deprecated.get_job_text_result(
|
|
725
711
|
job_id="string",
|
|
@@ -748,11 +734,10 @@ class AsyncDeprecatedClient:
|
|
|
748
734
|
Parameters:
|
|
749
735
|
- job_id: str.
|
|
750
736
|
---
|
|
751
|
-
from
|
|
737
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
752
738
|
|
|
753
739
|
client = AsyncLlamaCloud(
|
|
754
740
|
token="YOUR_TOKEN",
|
|
755
|
-
base_url="https://yourhost.com/path/to/api",
|
|
756
741
|
)
|
|
757
742
|
await client.deprecated.get_job_raw_text_result(
|
|
758
743
|
job_id="string",
|
|
@@ -783,11 +768,10 @@ class AsyncDeprecatedClient:
|
|
|
783
768
|
Parameters:
|
|
784
769
|
- job_id: str.
|
|
785
770
|
---
|
|
786
|
-
from
|
|
771
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
787
772
|
|
|
788
773
|
client = AsyncLlamaCloud(
|
|
789
774
|
token="YOUR_TOKEN",
|
|
790
|
-
base_url="https://yourhost.com/path/to/api",
|
|
791
775
|
)
|
|
792
776
|
await client.deprecated.get_job_result(
|
|
793
777
|
job_id="string",
|
|
@@ -818,11 +802,10 @@ class AsyncDeprecatedClient:
|
|
|
818
802
|
Parameters:
|
|
819
803
|
- job_id: str.
|
|
820
804
|
---
|
|
821
|
-
from
|
|
805
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
822
806
|
|
|
823
807
|
client = AsyncLlamaCloud(
|
|
824
808
|
token="YOUR_TOKEN",
|
|
825
|
-
base_url="https://yourhost.com/path/to/api",
|
|
826
809
|
)
|
|
827
810
|
await client.deprecated.get_job_raw_md_result(
|
|
828
811
|
job_id="string",
|
|
@@ -853,11 +836,10 @@ class AsyncDeprecatedClient:
|
|
|
853
836
|
Parameters:
|
|
854
837
|
- job_id: str.
|
|
855
838
|
---
|
|
856
|
-
from
|
|
839
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
857
840
|
|
|
858
841
|
client = AsyncLlamaCloud(
|
|
859
842
|
token="YOUR_TOKEN",
|
|
860
|
-
base_url="https://yourhost.com/path/to/api",
|
|
861
843
|
)
|
|
862
844
|
await client.deprecated.get_job_json_result(
|
|
863
845
|
job_id="string",
|
|
@@ -886,11 +868,10 @@ class AsyncDeprecatedClient:
|
|
|
886
868
|
Parameters:
|
|
887
869
|
- job_id: str.
|
|
888
870
|
---
|
|
889
|
-
from
|
|
871
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
890
872
|
|
|
891
873
|
client = AsyncLlamaCloud(
|
|
892
874
|
token="YOUR_TOKEN",
|
|
893
|
-
base_url="https://yourhost.com/path/to/api",
|
|
894
875
|
)
|
|
895
876
|
await client.deprecated.get_job_json_raw_result(
|
|
896
877
|
job_id="string",
|
|
@@ -919,11 +900,10 @@ class AsyncDeprecatedClient:
|
|
|
919
900
|
Get parsing history for user
|
|
920
901
|
|
|
921
902
|
---
|
|
922
|
-
from
|
|
903
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
923
904
|
|
|
924
905
|
client = AsyncLlamaCloud(
|
|
925
906
|
token="YOUR_TOKEN",
|
|
926
|
-
base_url="https://yourhost.com/path/to/api",
|
|
927
907
|
)
|
|
928
908
|
await client.deprecated.get_parsing_history_result()
|
|
929
909
|
"""
|
|
@@ -952,11 +932,10 @@ class AsyncDeprecatedClient:
|
|
|
952
932
|
|
|
953
933
|
- filename: str.
|
|
954
934
|
---
|
|
955
|
-
from
|
|
935
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
956
936
|
|
|
957
937
|
client = AsyncLlamaCloud(
|
|
958
938
|
token="YOUR_TOKEN",
|
|
959
|
-
base_url="https://yourhost.com/path/to/api",
|
|
960
939
|
)
|
|
961
940
|
await client.deprecated.generate_presigned_url(
|
|
962
941
|
job_id="string",
|