llama-cloud 0.0.2__py3-none-any.whl → 0.0.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/core/jsonable_encoder.py +3 -0
- llama_cloud/resources/api_keys/client.py +11 -8
- llama_cloud/resources/billing/client.py +9 -6
- llama_cloud/resources/component_definitions/client.py +9 -6
- llama_cloud/resources/data_sinks/client.py +21 -18
- llama_cloud/resources/data_sources/client.py +21 -18
- llama_cloud/resources/deprecated/client.py +27 -24
- llama_cloud/resources/evals/client.py +27 -24
- llama_cloud/resources/files/client.py +15 -12
- llama_cloud/resources/parsing/client.py +27 -24
- llama_cloud/resources/pipelines/client.py +73 -70
- llama_cloud/resources/projects/client.py +43 -40
- llama_cloud/types/api_key.py +3 -0
- llama_cloud/types/azure_open_ai_embedding.py +3 -0
- llama_cloud/types/base.py +3 -0
- llama_cloud/types/base_prompt_template.py +3 -0
- llama_cloud/types/bedrock_embedding.py +3 -0
- llama_cloud/types/chat_message.py +3 -0
- llama_cloud/types/cloud_az_storage_blob_data_source.py +3 -0
- llama_cloud/types/cloud_chroma_vector_store.py +3 -0
- llama_cloud/types/cloud_document.py +3 -0
- llama_cloud/types/cloud_document_create.py +3 -0
- llama_cloud/types/cloud_gcs_data_source.py +3 -0
- llama_cloud/types/cloud_google_drive_data_source.py +3 -0
- llama_cloud/types/cloud_one_drive_data_source.py +3 -0
- llama_cloud/types/cloud_pinecone_vector_store.py +3 -0
- llama_cloud/types/cloud_postgres_vector_store.py +3 -0
- llama_cloud/types/cloud_qdrant_vector_store.py +3 -0
- llama_cloud/types/cloud_s_3_data_source.py +3 -0
- llama_cloud/types/cloud_sharepoint_data_source.py +3 -0
- llama_cloud/types/cloud_weaviate_vector_store.py +3 -0
- llama_cloud/types/code_splitter.py +3 -0
- llama_cloud/types/cohere_embedding.py +3 -0
- llama_cloud/types/configurable_transformation_definition.py +3 -0
- llama_cloud/types/configured_transformation_item.py +3 -0
- llama_cloud/types/data_sink.py +3 -0
- llama_cloud/types/data_sink_create.py +3 -0
- llama_cloud/types/data_sink_definition.py +3 -0
- llama_cloud/types/data_source.py +3 -0
- llama_cloud/types/data_source_create.py +3 -0
- llama_cloud/types/data_source_definition.py +3 -0
- llama_cloud/types/eval_dataset.py +3 -0
- llama_cloud/types/eval_dataset_job_params.py +3 -0
- llama_cloud/types/eval_dataset_job_record.py +3 -0
- llama_cloud/types/eval_execution_params.py +3 -0
- llama_cloud/types/eval_execution_params_override.py +3 -0
- llama_cloud/types/eval_llm_model_data.py +3 -0
- llama_cloud/types/eval_question.py +3 -0
- llama_cloud/types/eval_question_create.py +3 -0
- llama_cloud/types/eval_question_result.py +3 -0
- llama_cloud/types/file.py +3 -0
- llama_cloud/types/gemini_embedding.py +3 -0
- llama_cloud/types/html_node_parser.py +3 -0
- llama_cloud/types/http_validation_error.py +3 -0
- llama_cloud/types/hugging_face_inference_api_embedding.py +3 -0
- llama_cloud/types/json_node_parser.py +3 -0
- llama_cloud/types/llm.py +3 -0
- llama_cloud/types/local_eval.py +3 -0
- llama_cloud/types/local_eval_results.py +3 -0
- llama_cloud/types/local_eval_sets.py +3 -0
- llama_cloud/types/markdown_element_node_parser.py +3 -0
- llama_cloud/types/markdown_node_parser.py +3 -0
- llama_cloud/types/metadata_filter.py +3 -0
- llama_cloud/types/metadata_filters.py +3 -0
- llama_cloud/types/metric_result.py +3 -0
- llama_cloud/types/node_parser.py +3 -0
- llama_cloud/types/open_ai_embedding.py +3 -0
- llama_cloud/types/parsing_history_item.py +3 -0
- llama_cloud/types/parsing_job.py +3 -0
- llama_cloud/types/parsing_job_json_result.py +3 -0
- llama_cloud/types/parsing_job_markdown_result.py +3 -0
- llama_cloud/types/parsing_job_text_result.py +3 -0
- llama_cloud/types/parsing_usage.py +3 -0
- llama_cloud/types/pipeline.py +3 -0
- llama_cloud/types/pipeline_create.py +3 -0
- llama_cloud/types/pipeline_data_source.py +3 -0
- llama_cloud/types/pipeline_data_source_create.py +3 -0
- llama_cloud/types/pipeline_deployment.py +3 -0
- llama_cloud/types/pipeline_file.py +3 -0
- llama_cloud/types/pipeline_file_create.py +3 -0
- llama_cloud/types/pipeline_file_status_response.py +3 -0
- llama_cloud/types/preset_retrieval_params.py +3 -0
- llama_cloud/types/presigned_url.py +3 -0
- llama_cloud/types/project.py +3 -0
- llama_cloud/types/project_create.py +3 -0
- llama_cloud/types/prompt_mixin_prompts.py +3 -0
- llama_cloud/types/prompt_spec.py +3 -0
- llama_cloud/types/related_node_info.py +3 -0
- llama_cloud/types/retrieve_results.py +3 -0
- llama_cloud/types/sentence_splitter.py +3 -0
- llama_cloud/types/simple_file_node_parser.py +3 -0
- llama_cloud/types/supported_eval_llm_model.py +3 -0
- llama_cloud/types/text_node.py +3 -0
- llama_cloud/types/text_node_with_score.py +3 -0
- llama_cloud/types/token_text_splitter.py +3 -0
- llama_cloud/types/validation_error.py +3 -0
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.3.dist-info}/METADATA +1 -1
- llama_cloud-0.0.3.dist-info/RECORD +173 -0
- llama_cloud-0.0.2.dist-info/RECORD +0 -173
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.3.dist-info}/LICENSE +0 -0
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.3.dist-info}/WHEEL +0 -0
|
@@ -15,6 +15,9 @@ from ...types.http_validation_error import HttpValidationError
|
|
|
15
15
|
from ...types.supported_eval_llm_model import SupportedEvalLlmModel
|
|
16
16
|
|
|
17
17
|
try:
|
|
18
|
+
import pydantic
|
|
19
|
+
if pydantic.__version__.startswith("1."):
|
|
20
|
+
raise ImportError
|
|
18
21
|
import pydantic.v1 as pydantic # type: ignore
|
|
19
22
|
except ImportError:
|
|
20
23
|
import pydantic # type: ignore
|
|
@@ -34,7 +37,7 @@ class EvalsClient:
|
|
|
34
37
|
Parameters:
|
|
35
38
|
- dataset_id: str.
|
|
36
39
|
---
|
|
37
|
-
from
|
|
40
|
+
from llama_cloud.client import LlamaCloud
|
|
38
41
|
|
|
39
42
|
client = LlamaCloud(
|
|
40
43
|
token="YOUR_TOKEN",
|
|
@@ -69,7 +72,7 @@ class EvalsClient:
|
|
|
69
72
|
|
|
70
73
|
- name: str. The name of the EvalDataset.
|
|
71
74
|
---
|
|
72
|
-
from
|
|
75
|
+
from llama_cloud.client import LlamaCloud
|
|
73
76
|
|
|
74
77
|
client = LlamaCloud(
|
|
75
78
|
token="YOUR_TOKEN",
|
|
@@ -104,7 +107,7 @@ class EvalsClient:
|
|
|
104
107
|
Parameters:
|
|
105
108
|
- dataset_id: str.
|
|
106
109
|
---
|
|
107
|
-
from
|
|
110
|
+
from llama_cloud.client import LlamaCloud
|
|
108
111
|
|
|
109
112
|
client = LlamaCloud(
|
|
110
113
|
token="YOUR_TOKEN",
|
|
@@ -137,7 +140,7 @@ class EvalsClient:
|
|
|
137
140
|
Parameters:
|
|
138
141
|
- dataset_id: str.
|
|
139
142
|
---
|
|
140
|
-
from
|
|
143
|
+
from llama_cloud.client import LlamaCloud
|
|
141
144
|
|
|
142
145
|
client = LlamaCloud(
|
|
143
146
|
token="YOUR_TOKEN",
|
|
@@ -174,8 +177,8 @@ class EvalsClient:
|
|
|
174
177
|
|
|
175
178
|
- request: EvalQuestionCreate.
|
|
176
179
|
---
|
|
177
|
-
from
|
|
178
|
-
from
|
|
180
|
+
from llama_cloud import EvalQuestionCreate
|
|
181
|
+
from llama_cloud.client import LlamaCloud
|
|
179
182
|
|
|
180
183
|
client = LlamaCloud(
|
|
181
184
|
token="YOUR_TOKEN",
|
|
@@ -218,7 +221,7 @@ class EvalsClient:
|
|
|
218
221
|
|
|
219
222
|
- request: typing.List[EvalQuestionCreate].
|
|
220
223
|
---
|
|
221
|
-
from
|
|
224
|
+
from llama_cloud.client import LlamaCloud
|
|
222
225
|
|
|
223
226
|
client = LlamaCloud(
|
|
224
227
|
token="YOUR_TOKEN",
|
|
@@ -255,7 +258,7 @@ class EvalsClient:
|
|
|
255
258
|
Parameters:
|
|
256
259
|
- question_id: str.
|
|
257
260
|
---
|
|
258
|
-
from
|
|
261
|
+
from llama_cloud.client import LlamaCloud
|
|
259
262
|
|
|
260
263
|
client = LlamaCloud(
|
|
261
264
|
token="YOUR_TOKEN",
|
|
@@ -290,8 +293,8 @@ class EvalsClient:
|
|
|
290
293
|
|
|
291
294
|
- request: EvalQuestionCreate.
|
|
292
295
|
---
|
|
293
|
-
from
|
|
294
|
-
from
|
|
296
|
+
from llama_cloud import EvalQuestionCreate
|
|
297
|
+
from llama_cloud.client import LlamaCloud
|
|
295
298
|
|
|
296
299
|
client = LlamaCloud(
|
|
297
300
|
token="YOUR_TOKEN",
|
|
@@ -328,7 +331,7 @@ class EvalsClient:
|
|
|
328
331
|
Parameters:
|
|
329
332
|
- question_id: str.
|
|
330
333
|
---
|
|
331
|
-
from
|
|
334
|
+
from llama_cloud.client import LlamaCloud
|
|
332
335
|
|
|
333
336
|
client = LlamaCloud(
|
|
334
337
|
token="YOUR_TOKEN",
|
|
@@ -359,7 +362,7 @@ class EvalsClient:
|
|
|
359
362
|
Get all supported models.
|
|
360
363
|
|
|
361
364
|
---
|
|
362
|
-
from
|
|
365
|
+
from llama_cloud.client import LlamaCloud
|
|
363
366
|
|
|
364
367
|
client = LlamaCloud(
|
|
365
368
|
token="YOUR_TOKEN",
|
|
@@ -395,7 +398,7 @@ class AsyncEvalsClient:
|
|
|
395
398
|
Parameters:
|
|
396
399
|
- dataset_id: str.
|
|
397
400
|
---
|
|
398
|
-
from
|
|
401
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
399
402
|
|
|
400
403
|
client = AsyncLlamaCloud(
|
|
401
404
|
token="YOUR_TOKEN",
|
|
@@ -430,7 +433,7 @@ class AsyncEvalsClient:
|
|
|
430
433
|
|
|
431
434
|
- name: str. The name of the EvalDataset.
|
|
432
435
|
---
|
|
433
|
-
from
|
|
436
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
434
437
|
|
|
435
438
|
client = AsyncLlamaCloud(
|
|
436
439
|
token="YOUR_TOKEN",
|
|
@@ -465,7 +468,7 @@ class AsyncEvalsClient:
|
|
|
465
468
|
Parameters:
|
|
466
469
|
- dataset_id: str.
|
|
467
470
|
---
|
|
468
|
-
from
|
|
471
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
469
472
|
|
|
470
473
|
client = AsyncLlamaCloud(
|
|
471
474
|
token="YOUR_TOKEN",
|
|
@@ -498,7 +501,7 @@ class AsyncEvalsClient:
|
|
|
498
501
|
Parameters:
|
|
499
502
|
- dataset_id: str.
|
|
500
503
|
---
|
|
501
|
-
from
|
|
504
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
502
505
|
|
|
503
506
|
client = AsyncLlamaCloud(
|
|
504
507
|
token="YOUR_TOKEN",
|
|
@@ -535,8 +538,8 @@ class AsyncEvalsClient:
|
|
|
535
538
|
|
|
536
539
|
- request: EvalQuestionCreate.
|
|
537
540
|
---
|
|
538
|
-
from
|
|
539
|
-
from
|
|
541
|
+
from llama_cloud import EvalQuestionCreate
|
|
542
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
540
543
|
|
|
541
544
|
client = AsyncLlamaCloud(
|
|
542
545
|
token="YOUR_TOKEN",
|
|
@@ -579,7 +582,7 @@ class AsyncEvalsClient:
|
|
|
579
582
|
|
|
580
583
|
- request: typing.List[EvalQuestionCreate].
|
|
581
584
|
---
|
|
582
|
-
from
|
|
585
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
583
586
|
|
|
584
587
|
client = AsyncLlamaCloud(
|
|
585
588
|
token="YOUR_TOKEN",
|
|
@@ -616,7 +619,7 @@ class AsyncEvalsClient:
|
|
|
616
619
|
Parameters:
|
|
617
620
|
- question_id: str.
|
|
618
621
|
---
|
|
619
|
-
from
|
|
622
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
620
623
|
|
|
621
624
|
client = AsyncLlamaCloud(
|
|
622
625
|
token="YOUR_TOKEN",
|
|
@@ -651,8 +654,8 @@ class AsyncEvalsClient:
|
|
|
651
654
|
|
|
652
655
|
- request: EvalQuestionCreate.
|
|
653
656
|
---
|
|
654
|
-
from
|
|
655
|
-
from
|
|
657
|
+
from llama_cloud import EvalQuestionCreate
|
|
658
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
656
659
|
|
|
657
660
|
client = AsyncLlamaCloud(
|
|
658
661
|
token="YOUR_TOKEN",
|
|
@@ -689,7 +692,7 @@ class AsyncEvalsClient:
|
|
|
689
692
|
Parameters:
|
|
690
693
|
- question_id: str.
|
|
691
694
|
---
|
|
692
|
-
from
|
|
695
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
693
696
|
|
|
694
697
|
client = AsyncLlamaCloud(
|
|
695
698
|
token="YOUR_TOKEN",
|
|
@@ -720,7 +723,7 @@ class AsyncEvalsClient:
|
|
|
720
723
|
Get all supported models.
|
|
721
724
|
|
|
722
725
|
---
|
|
723
|
-
from
|
|
726
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
724
727
|
|
|
725
728
|
client = AsyncLlamaCloud(
|
|
726
729
|
token="YOUR_TOKEN",
|
|
@@ -16,6 +16,9 @@ from ...types.presigned_url import PresignedUrl
|
|
|
16
16
|
from .types.file_create_resource_info_value import FileCreateResourceInfoValue
|
|
17
17
|
|
|
18
18
|
try:
|
|
19
|
+
import pydantic
|
|
20
|
+
if pydantic.__version__.startswith("1."):
|
|
21
|
+
raise ImportError
|
|
19
22
|
import pydantic.v1 as pydantic # type: ignore
|
|
20
23
|
except ImportError:
|
|
21
24
|
import pydantic # type: ignore
|
|
@@ -37,7 +40,7 @@ class FilesClient:
|
|
|
37
40
|
|
|
38
41
|
- project_id: typing.Optional[str].
|
|
39
42
|
---
|
|
40
|
-
from
|
|
43
|
+
from llama_cloud.client import LlamaCloud
|
|
41
44
|
|
|
42
45
|
client = LlamaCloud(
|
|
43
46
|
token="YOUR_TOKEN",
|
|
@@ -73,7 +76,7 @@ class FilesClient:
|
|
|
73
76
|
|
|
74
77
|
- project_id: typing.Optional[str].
|
|
75
78
|
---
|
|
76
|
-
from
|
|
79
|
+
from llama_cloud.client import LlamaCloud
|
|
77
80
|
|
|
78
81
|
client = LlamaCloud(
|
|
79
82
|
token="YOUR_TOKEN",
|
|
@@ -107,7 +110,7 @@ class FilesClient:
|
|
|
107
110
|
Parameters:
|
|
108
111
|
- project_id: typing.Optional[str].
|
|
109
112
|
---
|
|
110
|
-
from
|
|
113
|
+
from llama_cloud.client import LlamaCloud
|
|
111
114
|
|
|
112
115
|
client = LlamaCloud(
|
|
113
116
|
token="YOUR_TOKEN",
|
|
@@ -186,7 +189,7 @@ class FilesClient:
|
|
|
186
189
|
|
|
187
190
|
- data_source_id: typing.Optional[str]. The ID of the data source that the file belongs to
|
|
188
191
|
---
|
|
189
|
-
from
|
|
192
|
+
from llama_cloud.client import LlamaCloud
|
|
190
193
|
|
|
191
194
|
client = LlamaCloud(
|
|
192
195
|
token="YOUR_TOKEN",
|
|
@@ -230,7 +233,7 @@ class FilesClient:
|
|
|
230
233
|
Parameters:
|
|
231
234
|
- project_id: typing.Optional[str].
|
|
232
235
|
---
|
|
233
|
-
from
|
|
236
|
+
from llama_cloud.client import LlamaCloud
|
|
234
237
|
|
|
235
238
|
client = LlamaCloud(
|
|
236
239
|
token="YOUR_TOKEN",
|
|
@@ -264,7 +267,7 @@ class FilesClient:
|
|
|
264
267
|
|
|
265
268
|
- project_id: typing.Optional[str].
|
|
266
269
|
---
|
|
267
|
-
from
|
|
270
|
+
from llama_cloud.client import LlamaCloud
|
|
268
271
|
|
|
269
272
|
client = LlamaCloud(
|
|
270
273
|
token="YOUR_TOKEN",
|
|
@@ -305,7 +308,7 @@ class AsyncFilesClient:
|
|
|
305
308
|
|
|
306
309
|
- project_id: typing.Optional[str].
|
|
307
310
|
---
|
|
308
|
-
from
|
|
311
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
309
312
|
|
|
310
313
|
client = AsyncLlamaCloud(
|
|
311
314
|
token="YOUR_TOKEN",
|
|
@@ -341,7 +344,7 @@ class AsyncFilesClient:
|
|
|
341
344
|
|
|
342
345
|
- project_id: typing.Optional[str].
|
|
343
346
|
---
|
|
344
|
-
from
|
|
347
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
345
348
|
|
|
346
349
|
client = AsyncLlamaCloud(
|
|
347
350
|
token="YOUR_TOKEN",
|
|
@@ -375,7 +378,7 @@ class AsyncFilesClient:
|
|
|
375
378
|
Parameters:
|
|
376
379
|
- project_id: typing.Optional[str].
|
|
377
380
|
---
|
|
378
|
-
from
|
|
381
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
379
382
|
|
|
380
383
|
client = AsyncLlamaCloud(
|
|
381
384
|
token="YOUR_TOKEN",
|
|
@@ -454,7 +457,7 @@ class AsyncFilesClient:
|
|
|
454
457
|
|
|
455
458
|
- data_source_id: typing.Optional[str]. The ID of the data source that the file belongs to
|
|
456
459
|
---
|
|
457
|
-
from
|
|
460
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
458
461
|
|
|
459
462
|
client = AsyncLlamaCloud(
|
|
460
463
|
token="YOUR_TOKEN",
|
|
@@ -498,7 +501,7 @@ class AsyncFilesClient:
|
|
|
498
501
|
Parameters:
|
|
499
502
|
- project_id: typing.Optional[str].
|
|
500
503
|
---
|
|
501
|
-
from
|
|
504
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
502
505
|
|
|
503
506
|
client = AsyncLlamaCloud(
|
|
504
507
|
token="YOUR_TOKEN",
|
|
@@ -532,7 +535,7 @@ class AsyncFilesClient:
|
|
|
532
535
|
|
|
533
536
|
- project_id: typing.Optional[str].
|
|
534
537
|
---
|
|
535
|
-
from
|
|
538
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
536
539
|
|
|
537
540
|
client = AsyncLlamaCloud(
|
|
538
541
|
token="YOUR_TOKEN",
|
|
@@ -20,6 +20,9 @@ from ...types.parsing_usage import ParsingUsage
|
|
|
20
20
|
from ...types.presigned_url import PresignedUrl
|
|
21
21
|
|
|
22
22
|
try:
|
|
23
|
+
import pydantic
|
|
24
|
+
if pydantic.__version__.startswith("1."):
|
|
25
|
+
raise ImportError
|
|
23
26
|
import pydantic.v1 as pydantic # type: ignore
|
|
24
27
|
except ImportError:
|
|
25
28
|
import pydantic # type: ignore
|
|
@@ -41,7 +44,7 @@ class ParsingClient:
|
|
|
41
44
|
|
|
42
45
|
- name: str.
|
|
43
46
|
---
|
|
44
|
-
from
|
|
47
|
+
from llama_cloud.client import LlamaCloud
|
|
45
48
|
|
|
46
49
|
client = LlamaCloud(
|
|
47
50
|
token="YOUR_TOKEN",
|
|
@@ -75,7 +78,7 @@ class ParsingClient:
|
|
|
75
78
|
Get a list of supported file extensions
|
|
76
79
|
|
|
77
80
|
---
|
|
78
|
-
from
|
|
81
|
+
from llama_cloud.client import LlamaCloud
|
|
79
82
|
|
|
80
83
|
client = LlamaCloud(
|
|
81
84
|
token="YOUR_TOKEN",
|
|
@@ -174,7 +177,7 @@ class ParsingClient:
|
|
|
174
177
|
Get parsing usage for user
|
|
175
178
|
|
|
176
179
|
---
|
|
177
|
-
from
|
|
180
|
+
from llama_cloud.client import LlamaCloud
|
|
178
181
|
|
|
179
182
|
client = LlamaCloud(
|
|
180
183
|
token="YOUR_TOKEN",
|
|
@@ -205,7 +208,7 @@ class ParsingClient:
|
|
|
205
208
|
Parameters:
|
|
206
209
|
- job_id: str.
|
|
207
210
|
---
|
|
208
|
-
from
|
|
211
|
+
from llama_cloud.client import LlamaCloud
|
|
209
212
|
|
|
210
213
|
client = LlamaCloud(
|
|
211
214
|
token="YOUR_TOKEN",
|
|
@@ -238,7 +241,7 @@ class ParsingClient:
|
|
|
238
241
|
Parameters:
|
|
239
242
|
- job_id: str.
|
|
240
243
|
---
|
|
241
|
-
from
|
|
244
|
+
from llama_cloud.client import LlamaCloud
|
|
242
245
|
|
|
243
246
|
client = LlamaCloud(
|
|
244
247
|
token="YOUR_TOKEN",
|
|
@@ -271,7 +274,7 @@ class ParsingClient:
|
|
|
271
274
|
Parameters:
|
|
272
275
|
- job_id: str.
|
|
273
276
|
---
|
|
274
|
-
from
|
|
277
|
+
from llama_cloud.client import LlamaCloud
|
|
275
278
|
|
|
276
279
|
client = LlamaCloud(
|
|
277
280
|
token="YOUR_TOKEN",
|
|
@@ -306,7 +309,7 @@ class ParsingClient:
|
|
|
306
309
|
Parameters:
|
|
307
310
|
- job_id: str.
|
|
308
311
|
---
|
|
309
|
-
from
|
|
312
|
+
from llama_cloud.client import LlamaCloud
|
|
310
313
|
|
|
311
314
|
client = LlamaCloud(
|
|
312
315
|
token="YOUR_TOKEN",
|
|
@@ -341,7 +344,7 @@ class ParsingClient:
|
|
|
341
344
|
Parameters:
|
|
342
345
|
- job_id: str.
|
|
343
346
|
---
|
|
344
|
-
from
|
|
347
|
+
from llama_cloud.client import LlamaCloud
|
|
345
348
|
|
|
346
349
|
client = LlamaCloud(
|
|
347
350
|
token="YOUR_TOKEN",
|
|
@@ -376,7 +379,7 @@ class ParsingClient:
|
|
|
376
379
|
Parameters:
|
|
377
380
|
- job_id: str.
|
|
378
381
|
---
|
|
379
|
-
from
|
|
382
|
+
from llama_cloud.client import LlamaCloud
|
|
380
383
|
|
|
381
384
|
client = LlamaCloud(
|
|
382
385
|
token="YOUR_TOKEN",
|
|
@@ -409,7 +412,7 @@ class ParsingClient:
|
|
|
409
412
|
Parameters:
|
|
410
413
|
- job_id: str.
|
|
411
414
|
---
|
|
412
|
-
from
|
|
415
|
+
from llama_cloud.client import LlamaCloud
|
|
413
416
|
|
|
414
417
|
client = LlamaCloud(
|
|
415
418
|
token="YOUR_TOKEN",
|
|
@@ -442,7 +445,7 @@ class ParsingClient:
|
|
|
442
445
|
Get parsing history for user
|
|
443
446
|
|
|
444
447
|
---
|
|
445
|
-
from
|
|
448
|
+
from llama_cloud.client import LlamaCloud
|
|
446
449
|
|
|
447
450
|
client = LlamaCloud(
|
|
448
451
|
token="YOUR_TOKEN",
|
|
@@ -475,7 +478,7 @@ class ParsingClient:
|
|
|
475
478
|
|
|
476
479
|
- filename: str.
|
|
477
480
|
---
|
|
478
|
-
from
|
|
481
|
+
from llama_cloud.client import LlamaCloud
|
|
479
482
|
|
|
480
483
|
client = LlamaCloud(
|
|
481
484
|
token="YOUR_TOKEN",
|
|
@@ -518,7 +521,7 @@ class AsyncParsingClient:
|
|
|
518
521
|
|
|
519
522
|
- name: str.
|
|
520
523
|
---
|
|
521
|
-
from
|
|
524
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
522
525
|
|
|
523
526
|
client = AsyncLlamaCloud(
|
|
524
527
|
token="YOUR_TOKEN",
|
|
@@ -552,7 +555,7 @@ class AsyncParsingClient:
|
|
|
552
555
|
Get a list of supported file extensions
|
|
553
556
|
|
|
554
557
|
---
|
|
555
|
-
from
|
|
558
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
556
559
|
|
|
557
560
|
client = AsyncLlamaCloud(
|
|
558
561
|
token="YOUR_TOKEN",
|
|
@@ -651,7 +654,7 @@ class AsyncParsingClient:
|
|
|
651
654
|
Get parsing usage for user
|
|
652
655
|
|
|
653
656
|
---
|
|
654
|
-
from
|
|
657
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
655
658
|
|
|
656
659
|
client = AsyncLlamaCloud(
|
|
657
660
|
token="YOUR_TOKEN",
|
|
@@ -682,7 +685,7 @@ class AsyncParsingClient:
|
|
|
682
685
|
Parameters:
|
|
683
686
|
- job_id: str.
|
|
684
687
|
---
|
|
685
|
-
from
|
|
688
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
686
689
|
|
|
687
690
|
client = AsyncLlamaCloud(
|
|
688
691
|
token="YOUR_TOKEN",
|
|
@@ -715,7 +718,7 @@ class AsyncParsingClient:
|
|
|
715
718
|
Parameters:
|
|
716
719
|
- job_id: str.
|
|
717
720
|
---
|
|
718
|
-
from
|
|
721
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
719
722
|
|
|
720
723
|
client = AsyncLlamaCloud(
|
|
721
724
|
token="YOUR_TOKEN",
|
|
@@ -748,7 +751,7 @@ class AsyncParsingClient:
|
|
|
748
751
|
Parameters:
|
|
749
752
|
- job_id: str.
|
|
750
753
|
---
|
|
751
|
-
from
|
|
754
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
752
755
|
|
|
753
756
|
client = AsyncLlamaCloud(
|
|
754
757
|
token="YOUR_TOKEN",
|
|
@@ -783,7 +786,7 @@ class AsyncParsingClient:
|
|
|
783
786
|
Parameters:
|
|
784
787
|
- job_id: str.
|
|
785
788
|
---
|
|
786
|
-
from
|
|
789
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
787
790
|
|
|
788
791
|
client = AsyncLlamaCloud(
|
|
789
792
|
token="YOUR_TOKEN",
|
|
@@ -818,7 +821,7 @@ class AsyncParsingClient:
|
|
|
818
821
|
Parameters:
|
|
819
822
|
- job_id: str.
|
|
820
823
|
---
|
|
821
|
-
from
|
|
824
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
822
825
|
|
|
823
826
|
client = AsyncLlamaCloud(
|
|
824
827
|
token="YOUR_TOKEN",
|
|
@@ -853,7 +856,7 @@ class AsyncParsingClient:
|
|
|
853
856
|
Parameters:
|
|
854
857
|
- job_id: str.
|
|
855
858
|
---
|
|
856
|
-
from
|
|
859
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
857
860
|
|
|
858
861
|
client = AsyncLlamaCloud(
|
|
859
862
|
token="YOUR_TOKEN",
|
|
@@ -886,7 +889,7 @@ class AsyncParsingClient:
|
|
|
886
889
|
Parameters:
|
|
887
890
|
- job_id: str.
|
|
888
891
|
---
|
|
889
|
-
from
|
|
892
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
890
893
|
|
|
891
894
|
client = AsyncLlamaCloud(
|
|
892
895
|
token="YOUR_TOKEN",
|
|
@@ -919,7 +922,7 @@ class AsyncParsingClient:
|
|
|
919
922
|
Get parsing history for user
|
|
920
923
|
|
|
921
924
|
---
|
|
922
|
-
from
|
|
925
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
923
926
|
|
|
924
927
|
client = AsyncLlamaCloud(
|
|
925
928
|
token="YOUR_TOKEN",
|
|
@@ -952,7 +955,7 @@ class AsyncParsingClient:
|
|
|
952
955
|
|
|
953
956
|
- filename: str.
|
|
954
957
|
---
|
|
955
|
-
from
|
|
958
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
956
959
|
|
|
957
960
|
client = AsyncLlamaCloud(
|
|
958
961
|
token="YOUR_TOKEN",
|