llama-cloud 0.0.2__py3-none-any.whl → 0.0.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +2 -0
- llama_cloud/client.py +16 -4
- llama_cloud/core/jsonable_encoder.py +3 -0
- llama_cloud/environment.py +7 -0
- llama_cloud/resources/api_keys/client.py +11 -16
- llama_cloud/resources/billing/client.py +9 -12
- llama_cloud/resources/component_definitions/client.py +9 -12
- llama_cloud/resources/data_sinks/client.py +21 -30
- llama_cloud/resources/data_sources/client.py +21 -30
- llama_cloud/resources/deprecated/client.py +27 -48
- llama_cloud/resources/evals/client.py +27 -44
- llama_cloud/resources/files/client.py +15 -24
- llama_cloud/resources/parsing/client.py +27 -48
- llama_cloud/resources/pipelines/client.py +73 -128
- llama_cloud/resources/projects/client.py +43 -72
- llama_cloud/types/api_key.py +3 -0
- llama_cloud/types/azure_open_ai_embedding.py +3 -0
- llama_cloud/types/base.py +3 -0
- llama_cloud/types/base_prompt_template.py +3 -0
- llama_cloud/types/bedrock_embedding.py +3 -0
- llama_cloud/types/chat_message.py +3 -0
- llama_cloud/types/cloud_az_storage_blob_data_source.py +3 -0
- llama_cloud/types/cloud_chroma_vector_store.py +3 -0
- llama_cloud/types/cloud_document.py +3 -0
- llama_cloud/types/cloud_document_create.py +3 -0
- llama_cloud/types/cloud_gcs_data_source.py +3 -0
- llama_cloud/types/cloud_google_drive_data_source.py +3 -0
- llama_cloud/types/cloud_one_drive_data_source.py +3 -0
- llama_cloud/types/cloud_pinecone_vector_store.py +3 -0
- llama_cloud/types/cloud_postgres_vector_store.py +3 -0
- llama_cloud/types/cloud_qdrant_vector_store.py +3 -0
- llama_cloud/types/cloud_s_3_data_source.py +3 -0
- llama_cloud/types/cloud_sharepoint_data_source.py +3 -0
- llama_cloud/types/cloud_weaviate_vector_store.py +3 -0
- llama_cloud/types/code_splitter.py +3 -0
- llama_cloud/types/cohere_embedding.py +3 -0
- llama_cloud/types/configurable_transformation_definition.py +3 -0
- llama_cloud/types/configured_transformation_item.py +3 -0
- llama_cloud/types/data_sink.py +3 -0
- llama_cloud/types/data_sink_create.py +3 -0
- llama_cloud/types/data_sink_definition.py +3 -0
- llama_cloud/types/data_source.py +3 -0
- llama_cloud/types/data_source_create.py +3 -0
- llama_cloud/types/data_source_definition.py +3 -0
- llama_cloud/types/eval_dataset.py +3 -0
- llama_cloud/types/eval_dataset_job_params.py +3 -0
- llama_cloud/types/eval_dataset_job_record.py +3 -0
- llama_cloud/types/eval_execution_params.py +3 -0
- llama_cloud/types/eval_execution_params_override.py +3 -0
- llama_cloud/types/eval_llm_model_data.py +3 -0
- llama_cloud/types/eval_question.py +3 -0
- llama_cloud/types/eval_question_create.py +3 -0
- llama_cloud/types/eval_question_result.py +3 -0
- llama_cloud/types/file.py +3 -0
- llama_cloud/types/gemini_embedding.py +3 -0
- llama_cloud/types/html_node_parser.py +3 -0
- llama_cloud/types/http_validation_error.py +3 -0
- llama_cloud/types/hugging_face_inference_api_embedding.py +3 -0
- llama_cloud/types/json_node_parser.py +3 -0
- llama_cloud/types/llm.py +3 -0
- llama_cloud/types/local_eval.py +3 -0
- llama_cloud/types/local_eval_results.py +3 -0
- llama_cloud/types/local_eval_sets.py +3 -0
- llama_cloud/types/markdown_element_node_parser.py +3 -0
- llama_cloud/types/markdown_node_parser.py +3 -0
- llama_cloud/types/metadata_filter.py +3 -0
- llama_cloud/types/metadata_filters.py +3 -0
- llama_cloud/types/metric_result.py +3 -0
- llama_cloud/types/node_parser.py +3 -0
- llama_cloud/types/open_ai_embedding.py +3 -0
- llama_cloud/types/parsing_history_item.py +3 -0
- llama_cloud/types/parsing_job.py +3 -0
- llama_cloud/types/parsing_job_json_result.py +3 -0
- llama_cloud/types/parsing_job_markdown_result.py +3 -0
- llama_cloud/types/parsing_job_text_result.py +3 -0
- llama_cloud/types/parsing_usage.py +3 -0
- llama_cloud/types/pipeline.py +3 -0
- llama_cloud/types/pipeline_create.py +3 -0
- llama_cloud/types/pipeline_data_source.py +3 -0
- llama_cloud/types/pipeline_data_source_create.py +3 -0
- llama_cloud/types/pipeline_deployment.py +3 -0
- llama_cloud/types/pipeline_file.py +3 -0
- llama_cloud/types/pipeline_file_create.py +3 -0
- llama_cloud/types/pipeline_file_status_response.py +3 -0
- llama_cloud/types/preset_retrieval_params.py +3 -0
- llama_cloud/types/presigned_url.py +3 -0
- llama_cloud/types/project.py +3 -0
- llama_cloud/types/project_create.py +3 -0
- llama_cloud/types/prompt_mixin_prompts.py +3 -0
- llama_cloud/types/prompt_spec.py +3 -0
- llama_cloud/types/related_node_info.py +3 -0
- llama_cloud/types/retrieve_results.py +3 -0
- llama_cloud/types/sentence_splitter.py +3 -0
- llama_cloud/types/simple_file_node_parser.py +3 -0
- llama_cloud/types/supported_eval_llm_model.py +3 -0
- llama_cloud/types/text_node.py +3 -0
- llama_cloud/types/text_node_with_score.py +3 -0
- llama_cloud/types/token_text_splitter.py +3 -0
- llama_cloud/types/validation_error.py +3 -0
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.4.dist-info}/METADATA +1 -1
- llama_cloud-0.0.4.dist-info/RECORD +174 -0
- llama_cloud-0.0.2.dist-info/RECORD +0 -173
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.4.dist-info}/LICENSE +0 -0
- {llama_cloud-0.0.2.dist-info → llama_cloud-0.0.4.dist-info}/WHEEL +0 -0
|
@@ -20,6 +20,9 @@ from ...types.parsing_usage import ParsingUsage
|
|
|
20
20
|
from ...types.presigned_url import PresignedUrl
|
|
21
21
|
|
|
22
22
|
try:
|
|
23
|
+
import pydantic
|
|
24
|
+
if pydantic.__version__.startswith("1."):
|
|
25
|
+
raise ImportError
|
|
23
26
|
import pydantic.v1 as pydantic # type: ignore
|
|
24
27
|
except ImportError:
|
|
25
28
|
import pydantic # type: ignore
|
|
@@ -41,11 +44,10 @@ class ParsingClient:
|
|
|
41
44
|
|
|
42
45
|
- name: str.
|
|
43
46
|
---
|
|
44
|
-
from
|
|
47
|
+
from llama_cloud.client import LlamaCloud
|
|
45
48
|
|
|
46
49
|
client = LlamaCloud(
|
|
47
50
|
token="YOUR_TOKEN",
|
|
48
|
-
base_url="https://yourhost.com/path/to/api",
|
|
49
51
|
)
|
|
50
52
|
client.parsing.get_job_image_result(
|
|
51
53
|
job_id="string",
|
|
@@ -75,11 +77,10 @@ class ParsingClient:
|
|
|
75
77
|
Get a list of supported file extensions
|
|
76
78
|
|
|
77
79
|
---
|
|
78
|
-
from
|
|
80
|
+
from llama_cloud.client import LlamaCloud
|
|
79
81
|
|
|
80
82
|
client = LlamaCloud(
|
|
81
83
|
token="YOUR_TOKEN",
|
|
82
|
-
base_url="https://yourhost.com/path/to/api",
|
|
83
84
|
)
|
|
84
85
|
client.parsing.get_supported_file_extensions()
|
|
85
86
|
"""
|
|
@@ -174,11 +175,10 @@ class ParsingClient:
|
|
|
174
175
|
Get parsing usage for user
|
|
175
176
|
|
|
176
177
|
---
|
|
177
|
-
from
|
|
178
|
+
from llama_cloud.client import LlamaCloud
|
|
178
179
|
|
|
179
180
|
client = LlamaCloud(
|
|
180
181
|
token="YOUR_TOKEN",
|
|
181
|
-
base_url="https://yourhost.com/path/to/api",
|
|
182
182
|
)
|
|
183
183
|
client.parsing.usage()
|
|
184
184
|
"""
|
|
@@ -205,11 +205,10 @@ class ParsingClient:
|
|
|
205
205
|
Parameters:
|
|
206
206
|
- job_id: str.
|
|
207
207
|
---
|
|
208
|
-
from
|
|
208
|
+
from llama_cloud.client import LlamaCloud
|
|
209
209
|
|
|
210
210
|
client = LlamaCloud(
|
|
211
211
|
token="YOUR_TOKEN",
|
|
212
|
-
base_url="https://yourhost.com/path/to/api",
|
|
213
212
|
)
|
|
214
213
|
client.parsing.get_job(
|
|
215
214
|
job_id="string",
|
|
@@ -238,11 +237,10 @@ class ParsingClient:
|
|
|
238
237
|
Parameters:
|
|
239
238
|
- job_id: str.
|
|
240
239
|
---
|
|
241
|
-
from
|
|
240
|
+
from llama_cloud.client import LlamaCloud
|
|
242
241
|
|
|
243
242
|
client = LlamaCloud(
|
|
244
243
|
token="YOUR_TOKEN",
|
|
245
|
-
base_url="https://yourhost.com/path/to/api",
|
|
246
244
|
)
|
|
247
245
|
client.parsing.get_job_text_result(
|
|
248
246
|
job_id="string",
|
|
@@ -271,11 +269,10 @@ class ParsingClient:
|
|
|
271
269
|
Parameters:
|
|
272
270
|
- job_id: str.
|
|
273
271
|
---
|
|
274
|
-
from
|
|
272
|
+
from llama_cloud.client import LlamaCloud
|
|
275
273
|
|
|
276
274
|
client = LlamaCloud(
|
|
277
275
|
token="YOUR_TOKEN",
|
|
278
|
-
base_url="https://yourhost.com/path/to/api",
|
|
279
276
|
)
|
|
280
277
|
client.parsing.get_job_raw_text_result(
|
|
281
278
|
job_id="string",
|
|
@@ -306,11 +303,10 @@ class ParsingClient:
|
|
|
306
303
|
Parameters:
|
|
307
304
|
- job_id: str.
|
|
308
305
|
---
|
|
309
|
-
from
|
|
306
|
+
from llama_cloud.client import LlamaCloud
|
|
310
307
|
|
|
311
308
|
client = LlamaCloud(
|
|
312
309
|
token="YOUR_TOKEN",
|
|
313
|
-
base_url="https://yourhost.com/path/to/api",
|
|
314
310
|
)
|
|
315
311
|
client.parsing.get_job_result(
|
|
316
312
|
job_id="string",
|
|
@@ -341,11 +337,10 @@ class ParsingClient:
|
|
|
341
337
|
Parameters:
|
|
342
338
|
- job_id: str.
|
|
343
339
|
---
|
|
344
|
-
from
|
|
340
|
+
from llama_cloud.client import LlamaCloud
|
|
345
341
|
|
|
346
342
|
client = LlamaCloud(
|
|
347
343
|
token="YOUR_TOKEN",
|
|
348
|
-
base_url="https://yourhost.com/path/to/api",
|
|
349
344
|
)
|
|
350
345
|
client.parsing.get_job_raw_md_result(
|
|
351
346
|
job_id="string",
|
|
@@ -376,11 +371,10 @@ class ParsingClient:
|
|
|
376
371
|
Parameters:
|
|
377
372
|
- job_id: str.
|
|
378
373
|
---
|
|
379
|
-
from
|
|
374
|
+
from llama_cloud.client import LlamaCloud
|
|
380
375
|
|
|
381
376
|
client = LlamaCloud(
|
|
382
377
|
token="YOUR_TOKEN",
|
|
383
|
-
base_url="https://yourhost.com/path/to/api",
|
|
384
378
|
)
|
|
385
379
|
client.parsing.get_job_json_result(
|
|
386
380
|
job_id="string",
|
|
@@ -409,11 +403,10 @@ class ParsingClient:
|
|
|
409
403
|
Parameters:
|
|
410
404
|
- job_id: str.
|
|
411
405
|
---
|
|
412
|
-
from
|
|
406
|
+
from llama_cloud.client import LlamaCloud
|
|
413
407
|
|
|
414
408
|
client = LlamaCloud(
|
|
415
409
|
token="YOUR_TOKEN",
|
|
416
|
-
base_url="https://yourhost.com/path/to/api",
|
|
417
410
|
)
|
|
418
411
|
client.parsing.get_job_json_raw_result(
|
|
419
412
|
job_id="string",
|
|
@@ -442,11 +435,10 @@ class ParsingClient:
|
|
|
442
435
|
Get parsing history for user
|
|
443
436
|
|
|
444
437
|
---
|
|
445
|
-
from
|
|
438
|
+
from llama_cloud.client import LlamaCloud
|
|
446
439
|
|
|
447
440
|
client = LlamaCloud(
|
|
448
441
|
token="YOUR_TOKEN",
|
|
449
|
-
base_url="https://yourhost.com/path/to/api",
|
|
450
442
|
)
|
|
451
443
|
client.parsing.get_parsing_history_result()
|
|
452
444
|
"""
|
|
@@ -475,11 +467,10 @@ class ParsingClient:
|
|
|
475
467
|
|
|
476
468
|
- filename: str.
|
|
477
469
|
---
|
|
478
|
-
from
|
|
470
|
+
from llama_cloud.client import LlamaCloud
|
|
479
471
|
|
|
480
472
|
client = LlamaCloud(
|
|
481
473
|
token="YOUR_TOKEN",
|
|
482
|
-
base_url="https://yourhost.com/path/to/api",
|
|
483
474
|
)
|
|
484
475
|
client.parsing.generate_presigned_url(
|
|
485
476
|
job_id="string",
|
|
@@ -518,11 +509,10 @@ class AsyncParsingClient:
|
|
|
518
509
|
|
|
519
510
|
- name: str.
|
|
520
511
|
---
|
|
521
|
-
from
|
|
512
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
522
513
|
|
|
523
514
|
client = AsyncLlamaCloud(
|
|
524
515
|
token="YOUR_TOKEN",
|
|
525
|
-
base_url="https://yourhost.com/path/to/api",
|
|
526
516
|
)
|
|
527
517
|
await client.parsing.get_job_image_result(
|
|
528
518
|
job_id="string",
|
|
@@ -552,11 +542,10 @@ class AsyncParsingClient:
|
|
|
552
542
|
Get a list of supported file extensions
|
|
553
543
|
|
|
554
544
|
---
|
|
555
|
-
from
|
|
545
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
556
546
|
|
|
557
547
|
client = AsyncLlamaCloud(
|
|
558
548
|
token="YOUR_TOKEN",
|
|
559
|
-
base_url="https://yourhost.com/path/to/api",
|
|
560
549
|
)
|
|
561
550
|
await client.parsing.get_supported_file_extensions()
|
|
562
551
|
"""
|
|
@@ -651,11 +640,10 @@ class AsyncParsingClient:
|
|
|
651
640
|
Get parsing usage for user
|
|
652
641
|
|
|
653
642
|
---
|
|
654
|
-
from
|
|
643
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
655
644
|
|
|
656
645
|
client = AsyncLlamaCloud(
|
|
657
646
|
token="YOUR_TOKEN",
|
|
658
|
-
base_url="https://yourhost.com/path/to/api",
|
|
659
647
|
)
|
|
660
648
|
await client.parsing.usage()
|
|
661
649
|
"""
|
|
@@ -682,11 +670,10 @@ class AsyncParsingClient:
|
|
|
682
670
|
Parameters:
|
|
683
671
|
- job_id: str.
|
|
684
672
|
---
|
|
685
|
-
from
|
|
673
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
686
674
|
|
|
687
675
|
client = AsyncLlamaCloud(
|
|
688
676
|
token="YOUR_TOKEN",
|
|
689
|
-
base_url="https://yourhost.com/path/to/api",
|
|
690
677
|
)
|
|
691
678
|
await client.parsing.get_job(
|
|
692
679
|
job_id="string",
|
|
@@ -715,11 +702,10 @@ class AsyncParsingClient:
|
|
|
715
702
|
Parameters:
|
|
716
703
|
- job_id: str.
|
|
717
704
|
---
|
|
718
|
-
from
|
|
705
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
719
706
|
|
|
720
707
|
client = AsyncLlamaCloud(
|
|
721
708
|
token="YOUR_TOKEN",
|
|
722
|
-
base_url="https://yourhost.com/path/to/api",
|
|
723
709
|
)
|
|
724
710
|
await client.parsing.get_job_text_result(
|
|
725
711
|
job_id="string",
|
|
@@ -748,11 +734,10 @@ class AsyncParsingClient:
|
|
|
748
734
|
Parameters:
|
|
749
735
|
- job_id: str.
|
|
750
736
|
---
|
|
751
|
-
from
|
|
737
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
752
738
|
|
|
753
739
|
client = AsyncLlamaCloud(
|
|
754
740
|
token="YOUR_TOKEN",
|
|
755
|
-
base_url="https://yourhost.com/path/to/api",
|
|
756
741
|
)
|
|
757
742
|
await client.parsing.get_job_raw_text_result(
|
|
758
743
|
job_id="string",
|
|
@@ -783,11 +768,10 @@ class AsyncParsingClient:
|
|
|
783
768
|
Parameters:
|
|
784
769
|
- job_id: str.
|
|
785
770
|
---
|
|
786
|
-
from
|
|
771
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
787
772
|
|
|
788
773
|
client = AsyncLlamaCloud(
|
|
789
774
|
token="YOUR_TOKEN",
|
|
790
|
-
base_url="https://yourhost.com/path/to/api",
|
|
791
775
|
)
|
|
792
776
|
await client.parsing.get_job_result(
|
|
793
777
|
job_id="string",
|
|
@@ -818,11 +802,10 @@ class AsyncParsingClient:
|
|
|
818
802
|
Parameters:
|
|
819
803
|
- job_id: str.
|
|
820
804
|
---
|
|
821
|
-
from
|
|
805
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
822
806
|
|
|
823
807
|
client = AsyncLlamaCloud(
|
|
824
808
|
token="YOUR_TOKEN",
|
|
825
|
-
base_url="https://yourhost.com/path/to/api",
|
|
826
809
|
)
|
|
827
810
|
await client.parsing.get_job_raw_md_result(
|
|
828
811
|
job_id="string",
|
|
@@ -853,11 +836,10 @@ class AsyncParsingClient:
|
|
|
853
836
|
Parameters:
|
|
854
837
|
- job_id: str.
|
|
855
838
|
---
|
|
856
|
-
from
|
|
839
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
857
840
|
|
|
858
841
|
client = AsyncLlamaCloud(
|
|
859
842
|
token="YOUR_TOKEN",
|
|
860
|
-
base_url="https://yourhost.com/path/to/api",
|
|
861
843
|
)
|
|
862
844
|
await client.parsing.get_job_json_result(
|
|
863
845
|
job_id="string",
|
|
@@ -886,11 +868,10 @@ class AsyncParsingClient:
|
|
|
886
868
|
Parameters:
|
|
887
869
|
- job_id: str.
|
|
888
870
|
---
|
|
889
|
-
from
|
|
871
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
890
872
|
|
|
891
873
|
client = AsyncLlamaCloud(
|
|
892
874
|
token="YOUR_TOKEN",
|
|
893
|
-
base_url="https://yourhost.com/path/to/api",
|
|
894
875
|
)
|
|
895
876
|
await client.parsing.get_job_json_raw_result(
|
|
896
877
|
job_id="string",
|
|
@@ -919,11 +900,10 @@ class AsyncParsingClient:
|
|
|
919
900
|
Get parsing history for user
|
|
920
901
|
|
|
921
902
|
---
|
|
922
|
-
from
|
|
903
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
923
904
|
|
|
924
905
|
client = AsyncLlamaCloud(
|
|
925
906
|
token="YOUR_TOKEN",
|
|
926
|
-
base_url="https://yourhost.com/path/to/api",
|
|
927
907
|
)
|
|
928
908
|
await client.parsing.get_parsing_history_result()
|
|
929
909
|
"""
|
|
@@ -952,11 +932,10 @@ class AsyncParsingClient:
|
|
|
952
932
|
|
|
953
933
|
- filename: str.
|
|
954
934
|
---
|
|
955
|
-
from
|
|
935
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
956
936
|
|
|
957
937
|
client = AsyncLlamaCloud(
|
|
958
938
|
token="YOUR_TOKEN",
|
|
959
|
-
base_url="https://yourhost.com/path/to/api",
|
|
960
939
|
)
|
|
961
940
|
await client.parsing.generate_presigned_url(
|
|
962
941
|
job_id="string",
|