llama-cloud 0.0.3__py3-none-any.whl → 0.0.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +2 -8
- llama_cloud/client.py +16 -13
- llama_cloud/environment.py +7 -0
- llama_cloud/resources/__init__.py +1 -16
- llama_cloud/resources/component_definitions/client.py +0 -6
- llama_cloud/resources/data_sinks/client.py +0 -12
- llama_cloud/resources/data_sources/client.py +0 -12
- llama_cloud/resources/evals/client.py +0 -20
- llama_cloud/resources/files/client.py +0 -12
- llama_cloud/resources/parsing/client.py +0 -24
- llama_cloud/resources/pipelines/client.py +0 -58
- llama_cloud/resources/projects/client.py +0 -32
- llama_cloud/types/__init__.py +0 -2
- {llama_cloud-0.0.3.dist-info → llama_cloud-0.0.5.dist-info}/METADATA +1 -1
- {llama_cloud-0.0.3.dist-info → llama_cloud-0.0.5.dist-info}/RECORD +17 -23
- llama_cloud/resources/api_keys/__init__.py +0 -2
- llama_cloud/resources/api_keys/client.py +0 -305
- llama_cloud/resources/billing/__init__.py +0 -2
- llama_cloud/resources/billing/client.py +0 -237
- llama_cloud/resources/deprecated/__init__.py +0 -2
- llama_cloud/resources/deprecated/client.py +0 -985
- llama_cloud/types/api_key.py +0 -40
- {llama_cloud-0.0.3.dist-info → llama_cloud-0.0.5.dist-info}/LICENSE +0 -0
- {llama_cloud-0.0.3.dist-info → llama_cloud-0.0.5.dist-info}/WHEEL +0 -0
|
@@ -48,7 +48,6 @@ class ParsingClient:
|
|
|
48
48
|
|
|
49
49
|
client = LlamaCloud(
|
|
50
50
|
token="YOUR_TOKEN",
|
|
51
|
-
base_url="https://yourhost.com/path/to/api",
|
|
52
51
|
)
|
|
53
52
|
client.parsing.get_job_image_result(
|
|
54
53
|
job_id="string",
|
|
@@ -82,7 +81,6 @@ class ParsingClient:
|
|
|
82
81
|
|
|
83
82
|
client = LlamaCloud(
|
|
84
83
|
token="YOUR_TOKEN",
|
|
85
|
-
base_url="https://yourhost.com/path/to/api",
|
|
86
84
|
)
|
|
87
85
|
client.parsing.get_supported_file_extensions()
|
|
88
86
|
"""
|
|
@@ -181,7 +179,6 @@ class ParsingClient:
|
|
|
181
179
|
|
|
182
180
|
client = LlamaCloud(
|
|
183
181
|
token="YOUR_TOKEN",
|
|
184
|
-
base_url="https://yourhost.com/path/to/api",
|
|
185
182
|
)
|
|
186
183
|
client.parsing.usage()
|
|
187
184
|
"""
|
|
@@ -212,7 +209,6 @@ class ParsingClient:
|
|
|
212
209
|
|
|
213
210
|
client = LlamaCloud(
|
|
214
211
|
token="YOUR_TOKEN",
|
|
215
|
-
base_url="https://yourhost.com/path/to/api",
|
|
216
212
|
)
|
|
217
213
|
client.parsing.get_job(
|
|
218
214
|
job_id="string",
|
|
@@ -245,7 +241,6 @@ class ParsingClient:
|
|
|
245
241
|
|
|
246
242
|
client = LlamaCloud(
|
|
247
243
|
token="YOUR_TOKEN",
|
|
248
|
-
base_url="https://yourhost.com/path/to/api",
|
|
249
244
|
)
|
|
250
245
|
client.parsing.get_job_text_result(
|
|
251
246
|
job_id="string",
|
|
@@ -278,7 +273,6 @@ class ParsingClient:
|
|
|
278
273
|
|
|
279
274
|
client = LlamaCloud(
|
|
280
275
|
token="YOUR_TOKEN",
|
|
281
|
-
base_url="https://yourhost.com/path/to/api",
|
|
282
276
|
)
|
|
283
277
|
client.parsing.get_job_raw_text_result(
|
|
284
278
|
job_id="string",
|
|
@@ -313,7 +307,6 @@ class ParsingClient:
|
|
|
313
307
|
|
|
314
308
|
client = LlamaCloud(
|
|
315
309
|
token="YOUR_TOKEN",
|
|
316
|
-
base_url="https://yourhost.com/path/to/api",
|
|
317
310
|
)
|
|
318
311
|
client.parsing.get_job_result(
|
|
319
312
|
job_id="string",
|
|
@@ -348,7 +341,6 @@ class ParsingClient:
|
|
|
348
341
|
|
|
349
342
|
client = LlamaCloud(
|
|
350
343
|
token="YOUR_TOKEN",
|
|
351
|
-
base_url="https://yourhost.com/path/to/api",
|
|
352
344
|
)
|
|
353
345
|
client.parsing.get_job_raw_md_result(
|
|
354
346
|
job_id="string",
|
|
@@ -383,7 +375,6 @@ class ParsingClient:
|
|
|
383
375
|
|
|
384
376
|
client = LlamaCloud(
|
|
385
377
|
token="YOUR_TOKEN",
|
|
386
|
-
base_url="https://yourhost.com/path/to/api",
|
|
387
378
|
)
|
|
388
379
|
client.parsing.get_job_json_result(
|
|
389
380
|
job_id="string",
|
|
@@ -416,7 +407,6 @@ class ParsingClient:
|
|
|
416
407
|
|
|
417
408
|
client = LlamaCloud(
|
|
418
409
|
token="YOUR_TOKEN",
|
|
419
|
-
base_url="https://yourhost.com/path/to/api",
|
|
420
410
|
)
|
|
421
411
|
client.parsing.get_job_json_raw_result(
|
|
422
412
|
job_id="string",
|
|
@@ -449,7 +439,6 @@ class ParsingClient:
|
|
|
449
439
|
|
|
450
440
|
client = LlamaCloud(
|
|
451
441
|
token="YOUR_TOKEN",
|
|
452
|
-
base_url="https://yourhost.com/path/to/api",
|
|
453
442
|
)
|
|
454
443
|
client.parsing.get_parsing_history_result()
|
|
455
444
|
"""
|
|
@@ -482,7 +471,6 @@ class ParsingClient:
|
|
|
482
471
|
|
|
483
472
|
client = LlamaCloud(
|
|
484
473
|
token="YOUR_TOKEN",
|
|
485
|
-
base_url="https://yourhost.com/path/to/api",
|
|
486
474
|
)
|
|
487
475
|
client.parsing.generate_presigned_url(
|
|
488
476
|
job_id="string",
|
|
@@ -525,7 +513,6 @@ class AsyncParsingClient:
|
|
|
525
513
|
|
|
526
514
|
client = AsyncLlamaCloud(
|
|
527
515
|
token="YOUR_TOKEN",
|
|
528
|
-
base_url="https://yourhost.com/path/to/api",
|
|
529
516
|
)
|
|
530
517
|
await client.parsing.get_job_image_result(
|
|
531
518
|
job_id="string",
|
|
@@ -559,7 +546,6 @@ class AsyncParsingClient:
|
|
|
559
546
|
|
|
560
547
|
client = AsyncLlamaCloud(
|
|
561
548
|
token="YOUR_TOKEN",
|
|
562
|
-
base_url="https://yourhost.com/path/to/api",
|
|
563
549
|
)
|
|
564
550
|
await client.parsing.get_supported_file_extensions()
|
|
565
551
|
"""
|
|
@@ -658,7 +644,6 @@ class AsyncParsingClient:
|
|
|
658
644
|
|
|
659
645
|
client = AsyncLlamaCloud(
|
|
660
646
|
token="YOUR_TOKEN",
|
|
661
|
-
base_url="https://yourhost.com/path/to/api",
|
|
662
647
|
)
|
|
663
648
|
await client.parsing.usage()
|
|
664
649
|
"""
|
|
@@ -689,7 +674,6 @@ class AsyncParsingClient:
|
|
|
689
674
|
|
|
690
675
|
client = AsyncLlamaCloud(
|
|
691
676
|
token="YOUR_TOKEN",
|
|
692
|
-
base_url="https://yourhost.com/path/to/api",
|
|
693
677
|
)
|
|
694
678
|
await client.parsing.get_job(
|
|
695
679
|
job_id="string",
|
|
@@ -722,7 +706,6 @@ class AsyncParsingClient:
|
|
|
722
706
|
|
|
723
707
|
client = AsyncLlamaCloud(
|
|
724
708
|
token="YOUR_TOKEN",
|
|
725
|
-
base_url="https://yourhost.com/path/to/api",
|
|
726
709
|
)
|
|
727
710
|
await client.parsing.get_job_text_result(
|
|
728
711
|
job_id="string",
|
|
@@ -755,7 +738,6 @@ class AsyncParsingClient:
|
|
|
755
738
|
|
|
756
739
|
client = AsyncLlamaCloud(
|
|
757
740
|
token="YOUR_TOKEN",
|
|
758
|
-
base_url="https://yourhost.com/path/to/api",
|
|
759
741
|
)
|
|
760
742
|
await client.parsing.get_job_raw_text_result(
|
|
761
743
|
job_id="string",
|
|
@@ -790,7 +772,6 @@ class AsyncParsingClient:
|
|
|
790
772
|
|
|
791
773
|
client = AsyncLlamaCloud(
|
|
792
774
|
token="YOUR_TOKEN",
|
|
793
|
-
base_url="https://yourhost.com/path/to/api",
|
|
794
775
|
)
|
|
795
776
|
await client.parsing.get_job_result(
|
|
796
777
|
job_id="string",
|
|
@@ -825,7 +806,6 @@ class AsyncParsingClient:
|
|
|
825
806
|
|
|
826
807
|
client = AsyncLlamaCloud(
|
|
827
808
|
token="YOUR_TOKEN",
|
|
828
|
-
base_url="https://yourhost.com/path/to/api",
|
|
829
809
|
)
|
|
830
810
|
await client.parsing.get_job_raw_md_result(
|
|
831
811
|
job_id="string",
|
|
@@ -860,7 +840,6 @@ class AsyncParsingClient:
|
|
|
860
840
|
|
|
861
841
|
client = AsyncLlamaCloud(
|
|
862
842
|
token="YOUR_TOKEN",
|
|
863
|
-
base_url="https://yourhost.com/path/to/api",
|
|
864
843
|
)
|
|
865
844
|
await client.parsing.get_job_json_result(
|
|
866
845
|
job_id="string",
|
|
@@ -893,7 +872,6 @@ class AsyncParsingClient:
|
|
|
893
872
|
|
|
894
873
|
client = AsyncLlamaCloud(
|
|
895
874
|
token="YOUR_TOKEN",
|
|
896
|
-
base_url="https://yourhost.com/path/to/api",
|
|
897
875
|
)
|
|
898
876
|
await client.parsing.get_job_json_raw_result(
|
|
899
877
|
job_id="string",
|
|
@@ -926,7 +904,6 @@ class AsyncParsingClient:
|
|
|
926
904
|
|
|
927
905
|
client = AsyncLlamaCloud(
|
|
928
906
|
token="YOUR_TOKEN",
|
|
929
|
-
base_url="https://yourhost.com/path/to/api",
|
|
930
907
|
)
|
|
931
908
|
await client.parsing.get_parsing_history_result()
|
|
932
909
|
"""
|
|
@@ -959,7 +936,6 @@ class AsyncParsingClient:
|
|
|
959
936
|
|
|
960
937
|
client = AsyncLlamaCloud(
|
|
961
938
|
token="YOUR_TOKEN",
|
|
962
|
-
base_url="https://yourhost.com/path/to/api",
|
|
963
939
|
)
|
|
964
940
|
await client.parsing.generate_presigned_url(
|
|
965
941
|
job_id="string",
|
|
@@ -71,7 +71,6 @@ class PipelinesClient:
|
|
|
71
71
|
|
|
72
72
|
client = LlamaCloud(
|
|
73
73
|
token="YOUR_TOKEN",
|
|
74
|
-
base_url="https://yourhost.com/path/to/api",
|
|
75
74
|
)
|
|
76
75
|
client.pipelines.search_pipelines(
|
|
77
76
|
project_name="string",
|
|
@@ -121,7 +120,6 @@ class PipelinesClient:
|
|
|
121
120
|
|
|
122
121
|
client = LlamaCloud(
|
|
123
122
|
token="YOUR_TOKEN",
|
|
124
|
-
base_url="https://yourhost.com/path/to/api",
|
|
125
123
|
)
|
|
126
124
|
client.pipelines.create_pipeline(
|
|
127
125
|
request=PipelineCreate(
|
|
@@ -186,7 +184,6 @@ class PipelinesClient:
|
|
|
186
184
|
|
|
187
185
|
client = LlamaCloud(
|
|
188
186
|
token="YOUR_TOKEN",
|
|
189
|
-
base_url="https://yourhost.com/path/to/api",
|
|
190
187
|
)
|
|
191
188
|
client.pipelines.upsert_pipeline(
|
|
192
189
|
request=PipelineCreate(
|
|
@@ -241,7 +238,6 @@ class PipelinesClient:
|
|
|
241
238
|
|
|
242
239
|
client = LlamaCloud(
|
|
243
240
|
token="YOUR_TOKEN",
|
|
244
|
-
base_url="https://yourhost.com/path/to/api",
|
|
245
241
|
)
|
|
246
242
|
client.pipelines.get_pipeline(
|
|
247
243
|
pipeline_id="string",
|
|
@@ -312,7 +308,6 @@ class PipelinesClient:
|
|
|
312
308
|
|
|
313
309
|
client = LlamaCloud(
|
|
314
310
|
token="YOUR_TOKEN",
|
|
315
|
-
base_url="https://yourhost.com/path/to/api",
|
|
316
311
|
)
|
|
317
312
|
client.pipelines.update_existing_pipeline(
|
|
318
313
|
pipeline_id="string",
|
|
@@ -376,7 +371,6 @@ class PipelinesClient:
|
|
|
376
371
|
|
|
377
372
|
client = LlamaCloud(
|
|
378
373
|
token="YOUR_TOKEN",
|
|
379
|
-
base_url="https://yourhost.com/path/to/api",
|
|
380
374
|
)
|
|
381
375
|
client.pipelines.delete_pipeline(
|
|
382
376
|
pipeline_id="string",
|
|
@@ -409,7 +403,6 @@ class PipelinesClient:
|
|
|
409
403
|
|
|
410
404
|
client = LlamaCloud(
|
|
411
405
|
token="YOUR_TOKEN",
|
|
412
|
-
base_url="https://yourhost.com/path/to/api",
|
|
413
406
|
)
|
|
414
407
|
client.pipelines.sync_pipeline(
|
|
415
408
|
pipeline_id="string",
|
|
@@ -444,7 +437,6 @@ class PipelinesClient:
|
|
|
444
437
|
|
|
445
438
|
client = LlamaCloud(
|
|
446
439
|
token="YOUR_TOKEN",
|
|
447
|
-
base_url="https://yourhost.com/path/to/api",
|
|
448
440
|
)
|
|
449
441
|
client.pipelines.get_eval_dataset_executions(
|
|
450
442
|
pipeline_id="string",
|
|
@@ -495,7 +487,6 @@ class PipelinesClient:
|
|
|
495
487
|
|
|
496
488
|
client = LlamaCloud(
|
|
497
489
|
token="YOUR_TOKEN",
|
|
498
|
-
base_url="https://yourhost.com/path/to/api",
|
|
499
490
|
)
|
|
500
491
|
client.pipelines.execute_eval_dataset(
|
|
501
492
|
pipeline_id="string",
|
|
@@ -547,7 +538,6 @@ class PipelinesClient:
|
|
|
547
538
|
|
|
548
539
|
client = LlamaCloud(
|
|
549
540
|
token="YOUR_TOKEN",
|
|
550
|
-
base_url="https://yourhost.com/path/to/api",
|
|
551
541
|
)
|
|
552
542
|
client.pipelines.get_eval_dataset_execution_result(
|
|
553
543
|
pipeline_id="string",
|
|
@@ -590,7 +580,6 @@ class PipelinesClient:
|
|
|
590
580
|
|
|
591
581
|
client = LlamaCloud(
|
|
592
582
|
token="YOUR_TOKEN",
|
|
593
|
-
base_url="https://yourhost.com/path/to/api",
|
|
594
583
|
)
|
|
595
584
|
client.pipelines.get_eval_dataset_execution(
|
|
596
585
|
pipeline_id="string",
|
|
@@ -628,7 +617,6 @@ class PipelinesClient:
|
|
|
628
617
|
|
|
629
618
|
client = LlamaCloud(
|
|
630
619
|
token="YOUR_TOKEN",
|
|
631
|
-
base_url="https://yourhost.com/path/to/api",
|
|
632
620
|
)
|
|
633
621
|
client.pipelines.get_files_for_pipeline(
|
|
634
622
|
pipeline_id="string",
|
|
@@ -665,7 +653,6 @@ class PipelinesClient:
|
|
|
665
653
|
|
|
666
654
|
client = LlamaCloud(
|
|
667
655
|
token="YOUR_TOKEN",
|
|
668
|
-
base_url="https://yourhost.com/path/to/api",
|
|
669
656
|
)
|
|
670
657
|
client.pipelines.add_files_to_pipeline(
|
|
671
658
|
pipeline_id="string",
|
|
@@ -702,7 +689,6 @@ class PipelinesClient:
|
|
|
702
689
|
|
|
703
690
|
client = LlamaCloud(
|
|
704
691
|
token="YOUR_TOKEN",
|
|
705
|
-
base_url="https://yourhost.com/path/to/api",
|
|
706
692
|
)
|
|
707
693
|
client.pipelines.get_pipeline_file_status(
|
|
708
694
|
pipeline_id="string",
|
|
@@ -748,7 +734,6 @@ class PipelinesClient:
|
|
|
748
734
|
|
|
749
735
|
client = LlamaCloud(
|
|
750
736
|
token="YOUR_TOKEN",
|
|
751
|
-
base_url="https://yourhost.com/path/to/api",
|
|
752
737
|
)
|
|
753
738
|
client.pipelines.update_pipeline_file(
|
|
754
739
|
pipeline_id="string",
|
|
@@ -790,7 +775,6 @@ class PipelinesClient:
|
|
|
790
775
|
|
|
791
776
|
client = LlamaCloud(
|
|
792
777
|
token="YOUR_TOKEN",
|
|
793
|
-
base_url="https://yourhost.com/path/to/api",
|
|
794
778
|
)
|
|
795
779
|
client.pipelines.delete_pipeline_file(
|
|
796
780
|
pipeline_id="string",
|
|
@@ -826,7 +810,6 @@ class PipelinesClient:
|
|
|
826
810
|
|
|
827
811
|
client = LlamaCloud(
|
|
828
812
|
token="YOUR_TOKEN",
|
|
829
|
-
base_url="https://yourhost.com/path/to/api",
|
|
830
813
|
)
|
|
831
814
|
client.pipelines.get_pipeline_data_sources(
|
|
832
815
|
pipeline_id="string",
|
|
@@ -865,7 +848,6 @@ class PipelinesClient:
|
|
|
865
848
|
|
|
866
849
|
client = LlamaCloud(
|
|
867
850
|
token="YOUR_TOKEN",
|
|
868
|
-
base_url="https://yourhost.com/path/to/api",
|
|
869
851
|
)
|
|
870
852
|
client.pipelines.add_data_sources_to_pipeline(
|
|
871
853
|
pipeline_id="string",
|
|
@@ -904,7 +886,6 @@ class PipelinesClient:
|
|
|
904
886
|
|
|
905
887
|
client = LlamaCloud(
|
|
906
888
|
token="YOUR_TOKEN",
|
|
907
|
-
base_url="https://yourhost.com/path/to/api",
|
|
908
889
|
)
|
|
909
890
|
client.pipelines.delete_pipeline_data_source(
|
|
910
891
|
pipeline_id="string",
|
|
@@ -943,7 +924,6 @@ class PipelinesClient:
|
|
|
943
924
|
|
|
944
925
|
client = LlamaCloud(
|
|
945
926
|
token="YOUR_TOKEN",
|
|
946
|
-
base_url="https://yourhost.com/path/to/api",
|
|
947
927
|
)
|
|
948
928
|
client.pipelines.sync_pipeline_data_source(
|
|
949
929
|
pipeline_id="string",
|
|
@@ -1006,7 +986,6 @@ class PipelinesClient:
|
|
|
1006
986
|
|
|
1007
987
|
client = LlamaCloud(
|
|
1008
988
|
token="YOUR_TOKEN",
|
|
1009
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1010
989
|
)
|
|
1011
990
|
client.pipelines.run_search(
|
|
1012
991
|
pipeline_id="string",
|
|
@@ -1058,7 +1037,6 @@ class PipelinesClient:
|
|
|
1058
1037
|
|
|
1059
1038
|
client = LlamaCloud(
|
|
1060
1039
|
token="YOUR_TOKEN",
|
|
1061
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1062
1040
|
)
|
|
1063
1041
|
client.pipelines.get_pipeline_jobs(
|
|
1064
1042
|
pipeline_id="string",
|
|
@@ -1093,7 +1071,6 @@ class PipelinesClient:
|
|
|
1093
1071
|
|
|
1094
1072
|
client = LlamaCloud(
|
|
1095
1073
|
token="YOUR_TOKEN",
|
|
1096
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1097
1074
|
)
|
|
1098
1075
|
client.pipelines.get_pipeline_job(
|
|
1099
1076
|
pipeline_id="string",
|
|
@@ -1135,7 +1112,6 @@ class PipelinesClient:
|
|
|
1135
1112
|
|
|
1136
1113
|
client = LlamaCloud(
|
|
1137
1114
|
token="YOUR_TOKEN",
|
|
1138
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1139
1115
|
)
|
|
1140
1116
|
client.pipelines.list_pipeline_documents(
|
|
1141
1117
|
pipeline_id="string",
|
|
@@ -1175,7 +1151,6 @@ class PipelinesClient:
|
|
|
1175
1151
|
|
|
1176
1152
|
client = LlamaCloud(
|
|
1177
1153
|
token="YOUR_TOKEN",
|
|
1178
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1179
1154
|
)
|
|
1180
1155
|
client.pipelines.create_batch_pipeline_documents(
|
|
1181
1156
|
pipeline_id="string",
|
|
@@ -1216,7 +1191,6 @@ class PipelinesClient:
|
|
|
1216
1191
|
|
|
1217
1192
|
client = LlamaCloud(
|
|
1218
1193
|
token="YOUR_TOKEN",
|
|
1219
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1220
1194
|
)
|
|
1221
1195
|
client.pipelines.upsert_batch_pipeline_documents(
|
|
1222
1196
|
pipeline_id="string",
|
|
@@ -1255,7 +1229,6 @@ class PipelinesClient:
|
|
|
1255
1229
|
|
|
1256
1230
|
client = LlamaCloud(
|
|
1257
1231
|
token="YOUR_TOKEN",
|
|
1258
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1259
1232
|
)
|
|
1260
1233
|
client.pipelines.get_pipeline_document(
|
|
1261
1234
|
pipeline_id="string",
|
|
@@ -1293,7 +1266,6 @@ class PipelinesClient:
|
|
|
1293
1266
|
|
|
1294
1267
|
client = LlamaCloud(
|
|
1295
1268
|
token="YOUR_TOKEN",
|
|
1296
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1297
1269
|
)
|
|
1298
1270
|
client.pipelines.delete_pipeline_document(
|
|
1299
1271
|
pipeline_id="string",
|
|
@@ -1331,7 +1303,6 @@ class PipelinesClient:
|
|
|
1331
1303
|
|
|
1332
1304
|
client = LlamaCloud(
|
|
1333
1305
|
token="YOUR_TOKEN",
|
|
1334
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1335
1306
|
)
|
|
1336
1307
|
client.pipelines.get_pipeline_document_status(
|
|
1337
1308
|
pipeline_id="string",
|
|
@@ -1384,7 +1355,6 @@ class AsyncPipelinesClient:
|
|
|
1384
1355
|
|
|
1385
1356
|
client = AsyncLlamaCloud(
|
|
1386
1357
|
token="YOUR_TOKEN",
|
|
1387
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1388
1358
|
)
|
|
1389
1359
|
await client.pipelines.search_pipelines(
|
|
1390
1360
|
project_name="string",
|
|
@@ -1434,7 +1404,6 @@ class AsyncPipelinesClient:
|
|
|
1434
1404
|
|
|
1435
1405
|
client = AsyncLlamaCloud(
|
|
1436
1406
|
token="YOUR_TOKEN",
|
|
1437
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1438
1407
|
)
|
|
1439
1408
|
await client.pipelines.create_pipeline(
|
|
1440
1409
|
request=PipelineCreate(
|
|
@@ -1499,7 +1468,6 @@ class AsyncPipelinesClient:
|
|
|
1499
1468
|
|
|
1500
1469
|
client = AsyncLlamaCloud(
|
|
1501
1470
|
token="YOUR_TOKEN",
|
|
1502
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1503
1471
|
)
|
|
1504
1472
|
await client.pipelines.upsert_pipeline(
|
|
1505
1473
|
request=PipelineCreate(
|
|
@@ -1554,7 +1522,6 @@ class AsyncPipelinesClient:
|
|
|
1554
1522
|
|
|
1555
1523
|
client = AsyncLlamaCloud(
|
|
1556
1524
|
token="YOUR_TOKEN",
|
|
1557
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1558
1525
|
)
|
|
1559
1526
|
await client.pipelines.get_pipeline(
|
|
1560
1527
|
pipeline_id="string",
|
|
@@ -1625,7 +1592,6 @@ class AsyncPipelinesClient:
|
|
|
1625
1592
|
|
|
1626
1593
|
client = AsyncLlamaCloud(
|
|
1627
1594
|
token="YOUR_TOKEN",
|
|
1628
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1629
1595
|
)
|
|
1630
1596
|
await client.pipelines.update_existing_pipeline(
|
|
1631
1597
|
pipeline_id="string",
|
|
@@ -1689,7 +1655,6 @@ class AsyncPipelinesClient:
|
|
|
1689
1655
|
|
|
1690
1656
|
client = AsyncLlamaCloud(
|
|
1691
1657
|
token="YOUR_TOKEN",
|
|
1692
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1693
1658
|
)
|
|
1694
1659
|
await client.pipelines.delete_pipeline(
|
|
1695
1660
|
pipeline_id="string",
|
|
@@ -1722,7 +1687,6 @@ class AsyncPipelinesClient:
|
|
|
1722
1687
|
|
|
1723
1688
|
client = AsyncLlamaCloud(
|
|
1724
1689
|
token="YOUR_TOKEN",
|
|
1725
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1726
1690
|
)
|
|
1727
1691
|
await client.pipelines.sync_pipeline(
|
|
1728
1692
|
pipeline_id="string",
|
|
@@ -1759,7 +1723,6 @@ class AsyncPipelinesClient:
|
|
|
1759
1723
|
|
|
1760
1724
|
client = AsyncLlamaCloud(
|
|
1761
1725
|
token="YOUR_TOKEN",
|
|
1762
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1763
1726
|
)
|
|
1764
1727
|
await client.pipelines.get_eval_dataset_executions(
|
|
1765
1728
|
pipeline_id="string",
|
|
@@ -1810,7 +1773,6 @@ class AsyncPipelinesClient:
|
|
|
1810
1773
|
|
|
1811
1774
|
client = AsyncLlamaCloud(
|
|
1812
1775
|
token="YOUR_TOKEN",
|
|
1813
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1814
1776
|
)
|
|
1815
1777
|
await client.pipelines.execute_eval_dataset(
|
|
1816
1778
|
pipeline_id="string",
|
|
@@ -1862,7 +1824,6 @@ class AsyncPipelinesClient:
|
|
|
1862
1824
|
|
|
1863
1825
|
client = AsyncLlamaCloud(
|
|
1864
1826
|
token="YOUR_TOKEN",
|
|
1865
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1866
1827
|
)
|
|
1867
1828
|
await client.pipelines.get_eval_dataset_execution_result(
|
|
1868
1829
|
pipeline_id="string",
|
|
@@ -1905,7 +1866,6 @@ class AsyncPipelinesClient:
|
|
|
1905
1866
|
|
|
1906
1867
|
client = AsyncLlamaCloud(
|
|
1907
1868
|
token="YOUR_TOKEN",
|
|
1908
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1909
1869
|
)
|
|
1910
1870
|
await client.pipelines.get_eval_dataset_execution(
|
|
1911
1871
|
pipeline_id="string",
|
|
@@ -1943,7 +1903,6 @@ class AsyncPipelinesClient:
|
|
|
1943
1903
|
|
|
1944
1904
|
client = AsyncLlamaCloud(
|
|
1945
1905
|
token="YOUR_TOKEN",
|
|
1946
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1947
1906
|
)
|
|
1948
1907
|
await client.pipelines.get_files_for_pipeline(
|
|
1949
1908
|
pipeline_id="string",
|
|
@@ -1980,7 +1939,6 @@ class AsyncPipelinesClient:
|
|
|
1980
1939
|
|
|
1981
1940
|
client = AsyncLlamaCloud(
|
|
1982
1941
|
token="YOUR_TOKEN",
|
|
1983
|
-
base_url="https://yourhost.com/path/to/api",
|
|
1984
1942
|
)
|
|
1985
1943
|
await client.pipelines.add_files_to_pipeline(
|
|
1986
1944
|
pipeline_id="string",
|
|
@@ -2017,7 +1975,6 @@ class AsyncPipelinesClient:
|
|
|
2017
1975
|
|
|
2018
1976
|
client = AsyncLlamaCloud(
|
|
2019
1977
|
token="YOUR_TOKEN",
|
|
2020
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2021
1978
|
)
|
|
2022
1979
|
await client.pipelines.get_pipeline_file_status(
|
|
2023
1980
|
pipeline_id="string",
|
|
@@ -2063,7 +2020,6 @@ class AsyncPipelinesClient:
|
|
|
2063
2020
|
|
|
2064
2021
|
client = AsyncLlamaCloud(
|
|
2065
2022
|
token="YOUR_TOKEN",
|
|
2066
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2067
2023
|
)
|
|
2068
2024
|
await client.pipelines.update_pipeline_file(
|
|
2069
2025
|
pipeline_id="string",
|
|
@@ -2105,7 +2061,6 @@ class AsyncPipelinesClient:
|
|
|
2105
2061
|
|
|
2106
2062
|
client = AsyncLlamaCloud(
|
|
2107
2063
|
token="YOUR_TOKEN",
|
|
2108
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2109
2064
|
)
|
|
2110
2065
|
await client.pipelines.delete_pipeline_file(
|
|
2111
2066
|
pipeline_id="string",
|
|
@@ -2141,7 +2096,6 @@ class AsyncPipelinesClient:
|
|
|
2141
2096
|
|
|
2142
2097
|
client = AsyncLlamaCloud(
|
|
2143
2098
|
token="YOUR_TOKEN",
|
|
2144
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2145
2099
|
)
|
|
2146
2100
|
await client.pipelines.get_pipeline_data_sources(
|
|
2147
2101
|
pipeline_id="string",
|
|
@@ -2180,7 +2134,6 @@ class AsyncPipelinesClient:
|
|
|
2180
2134
|
|
|
2181
2135
|
client = AsyncLlamaCloud(
|
|
2182
2136
|
token="YOUR_TOKEN",
|
|
2183
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2184
2137
|
)
|
|
2185
2138
|
await client.pipelines.add_data_sources_to_pipeline(
|
|
2186
2139
|
pipeline_id="string",
|
|
@@ -2219,7 +2172,6 @@ class AsyncPipelinesClient:
|
|
|
2219
2172
|
|
|
2220
2173
|
client = AsyncLlamaCloud(
|
|
2221
2174
|
token="YOUR_TOKEN",
|
|
2222
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2223
2175
|
)
|
|
2224
2176
|
await client.pipelines.delete_pipeline_data_source(
|
|
2225
2177
|
pipeline_id="string",
|
|
@@ -2258,7 +2210,6 @@ class AsyncPipelinesClient:
|
|
|
2258
2210
|
|
|
2259
2211
|
client = AsyncLlamaCloud(
|
|
2260
2212
|
token="YOUR_TOKEN",
|
|
2261
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2262
2213
|
)
|
|
2263
2214
|
await client.pipelines.sync_pipeline_data_source(
|
|
2264
2215
|
pipeline_id="string",
|
|
@@ -2321,7 +2272,6 @@ class AsyncPipelinesClient:
|
|
|
2321
2272
|
|
|
2322
2273
|
client = AsyncLlamaCloud(
|
|
2323
2274
|
token="YOUR_TOKEN",
|
|
2324
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2325
2275
|
)
|
|
2326
2276
|
await client.pipelines.run_search(
|
|
2327
2277
|
pipeline_id="string",
|
|
@@ -2373,7 +2323,6 @@ class AsyncPipelinesClient:
|
|
|
2373
2323
|
|
|
2374
2324
|
client = AsyncLlamaCloud(
|
|
2375
2325
|
token="YOUR_TOKEN",
|
|
2376
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2377
2326
|
)
|
|
2378
2327
|
await client.pipelines.get_pipeline_jobs(
|
|
2379
2328
|
pipeline_id="string",
|
|
@@ -2408,7 +2357,6 @@ class AsyncPipelinesClient:
|
|
|
2408
2357
|
|
|
2409
2358
|
client = AsyncLlamaCloud(
|
|
2410
2359
|
token="YOUR_TOKEN",
|
|
2411
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2412
2360
|
)
|
|
2413
2361
|
await client.pipelines.get_pipeline_job(
|
|
2414
2362
|
pipeline_id="string",
|
|
@@ -2450,7 +2398,6 @@ class AsyncPipelinesClient:
|
|
|
2450
2398
|
|
|
2451
2399
|
client = AsyncLlamaCloud(
|
|
2452
2400
|
token="YOUR_TOKEN",
|
|
2453
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2454
2401
|
)
|
|
2455
2402
|
await client.pipelines.list_pipeline_documents(
|
|
2456
2403
|
pipeline_id="string",
|
|
@@ -2490,7 +2437,6 @@ class AsyncPipelinesClient:
|
|
|
2490
2437
|
|
|
2491
2438
|
client = AsyncLlamaCloud(
|
|
2492
2439
|
token="YOUR_TOKEN",
|
|
2493
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2494
2440
|
)
|
|
2495
2441
|
await client.pipelines.create_batch_pipeline_documents(
|
|
2496
2442
|
pipeline_id="string",
|
|
@@ -2531,7 +2477,6 @@ class AsyncPipelinesClient:
|
|
|
2531
2477
|
|
|
2532
2478
|
client = AsyncLlamaCloud(
|
|
2533
2479
|
token="YOUR_TOKEN",
|
|
2534
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2535
2480
|
)
|
|
2536
2481
|
await client.pipelines.upsert_batch_pipeline_documents(
|
|
2537
2482
|
pipeline_id="string",
|
|
@@ -2570,7 +2515,6 @@ class AsyncPipelinesClient:
|
|
|
2570
2515
|
|
|
2571
2516
|
client = AsyncLlamaCloud(
|
|
2572
2517
|
token="YOUR_TOKEN",
|
|
2573
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2574
2518
|
)
|
|
2575
2519
|
await client.pipelines.get_pipeline_document(
|
|
2576
2520
|
pipeline_id="string",
|
|
@@ -2608,7 +2552,6 @@ class AsyncPipelinesClient:
|
|
|
2608
2552
|
|
|
2609
2553
|
client = AsyncLlamaCloud(
|
|
2610
2554
|
token="YOUR_TOKEN",
|
|
2611
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2612
2555
|
)
|
|
2613
2556
|
await client.pipelines.delete_pipeline_document(
|
|
2614
2557
|
pipeline_id="string",
|
|
@@ -2646,7 +2589,6 @@ class AsyncPipelinesClient:
|
|
|
2646
2589
|
|
|
2647
2590
|
client = AsyncLlamaCloud(
|
|
2648
2591
|
token="YOUR_TOKEN",
|
|
2649
|
-
base_url="https://yourhost.com/path/to/api",
|
|
2650
2592
|
)
|
|
2651
2593
|
await client.pipelines.get_pipeline_document_status(
|
|
2652
2594
|
pipeline_id="string",
|