llama-cloud 0.0.3__py3-none-any.whl → 0.0.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

@@ -48,7 +48,6 @@ class ParsingClient:
48
48
 
49
49
  client = LlamaCloud(
50
50
  token="YOUR_TOKEN",
51
- base_url="https://yourhost.com/path/to/api",
52
51
  )
53
52
  client.parsing.get_job_image_result(
54
53
  job_id="string",
@@ -82,7 +81,6 @@ class ParsingClient:
82
81
 
83
82
  client = LlamaCloud(
84
83
  token="YOUR_TOKEN",
85
- base_url="https://yourhost.com/path/to/api",
86
84
  )
87
85
  client.parsing.get_supported_file_extensions()
88
86
  """
@@ -181,7 +179,6 @@ class ParsingClient:
181
179
 
182
180
  client = LlamaCloud(
183
181
  token="YOUR_TOKEN",
184
- base_url="https://yourhost.com/path/to/api",
185
182
  )
186
183
  client.parsing.usage()
187
184
  """
@@ -212,7 +209,6 @@ class ParsingClient:
212
209
 
213
210
  client = LlamaCloud(
214
211
  token="YOUR_TOKEN",
215
- base_url="https://yourhost.com/path/to/api",
216
212
  )
217
213
  client.parsing.get_job(
218
214
  job_id="string",
@@ -245,7 +241,6 @@ class ParsingClient:
245
241
 
246
242
  client = LlamaCloud(
247
243
  token="YOUR_TOKEN",
248
- base_url="https://yourhost.com/path/to/api",
249
244
  )
250
245
  client.parsing.get_job_text_result(
251
246
  job_id="string",
@@ -278,7 +273,6 @@ class ParsingClient:
278
273
 
279
274
  client = LlamaCloud(
280
275
  token="YOUR_TOKEN",
281
- base_url="https://yourhost.com/path/to/api",
282
276
  )
283
277
  client.parsing.get_job_raw_text_result(
284
278
  job_id="string",
@@ -313,7 +307,6 @@ class ParsingClient:
313
307
 
314
308
  client = LlamaCloud(
315
309
  token="YOUR_TOKEN",
316
- base_url="https://yourhost.com/path/to/api",
317
310
  )
318
311
  client.parsing.get_job_result(
319
312
  job_id="string",
@@ -348,7 +341,6 @@ class ParsingClient:
348
341
 
349
342
  client = LlamaCloud(
350
343
  token="YOUR_TOKEN",
351
- base_url="https://yourhost.com/path/to/api",
352
344
  )
353
345
  client.parsing.get_job_raw_md_result(
354
346
  job_id="string",
@@ -383,7 +375,6 @@ class ParsingClient:
383
375
 
384
376
  client = LlamaCloud(
385
377
  token="YOUR_TOKEN",
386
- base_url="https://yourhost.com/path/to/api",
387
378
  )
388
379
  client.parsing.get_job_json_result(
389
380
  job_id="string",
@@ -416,7 +407,6 @@ class ParsingClient:
416
407
 
417
408
  client = LlamaCloud(
418
409
  token="YOUR_TOKEN",
419
- base_url="https://yourhost.com/path/to/api",
420
410
  )
421
411
  client.parsing.get_job_json_raw_result(
422
412
  job_id="string",
@@ -449,7 +439,6 @@ class ParsingClient:
449
439
 
450
440
  client = LlamaCloud(
451
441
  token="YOUR_TOKEN",
452
- base_url="https://yourhost.com/path/to/api",
453
442
  )
454
443
  client.parsing.get_parsing_history_result()
455
444
  """
@@ -482,7 +471,6 @@ class ParsingClient:
482
471
 
483
472
  client = LlamaCloud(
484
473
  token="YOUR_TOKEN",
485
- base_url="https://yourhost.com/path/to/api",
486
474
  )
487
475
  client.parsing.generate_presigned_url(
488
476
  job_id="string",
@@ -525,7 +513,6 @@ class AsyncParsingClient:
525
513
 
526
514
  client = AsyncLlamaCloud(
527
515
  token="YOUR_TOKEN",
528
- base_url="https://yourhost.com/path/to/api",
529
516
  )
530
517
  await client.parsing.get_job_image_result(
531
518
  job_id="string",
@@ -559,7 +546,6 @@ class AsyncParsingClient:
559
546
 
560
547
  client = AsyncLlamaCloud(
561
548
  token="YOUR_TOKEN",
562
- base_url="https://yourhost.com/path/to/api",
563
549
  )
564
550
  await client.parsing.get_supported_file_extensions()
565
551
  """
@@ -658,7 +644,6 @@ class AsyncParsingClient:
658
644
 
659
645
  client = AsyncLlamaCloud(
660
646
  token="YOUR_TOKEN",
661
- base_url="https://yourhost.com/path/to/api",
662
647
  )
663
648
  await client.parsing.usage()
664
649
  """
@@ -689,7 +674,6 @@ class AsyncParsingClient:
689
674
 
690
675
  client = AsyncLlamaCloud(
691
676
  token="YOUR_TOKEN",
692
- base_url="https://yourhost.com/path/to/api",
693
677
  )
694
678
  await client.parsing.get_job(
695
679
  job_id="string",
@@ -722,7 +706,6 @@ class AsyncParsingClient:
722
706
 
723
707
  client = AsyncLlamaCloud(
724
708
  token="YOUR_TOKEN",
725
- base_url="https://yourhost.com/path/to/api",
726
709
  )
727
710
  await client.parsing.get_job_text_result(
728
711
  job_id="string",
@@ -755,7 +738,6 @@ class AsyncParsingClient:
755
738
 
756
739
  client = AsyncLlamaCloud(
757
740
  token="YOUR_TOKEN",
758
- base_url="https://yourhost.com/path/to/api",
759
741
  )
760
742
  await client.parsing.get_job_raw_text_result(
761
743
  job_id="string",
@@ -790,7 +772,6 @@ class AsyncParsingClient:
790
772
 
791
773
  client = AsyncLlamaCloud(
792
774
  token="YOUR_TOKEN",
793
- base_url="https://yourhost.com/path/to/api",
794
775
  )
795
776
  await client.parsing.get_job_result(
796
777
  job_id="string",
@@ -825,7 +806,6 @@ class AsyncParsingClient:
825
806
 
826
807
  client = AsyncLlamaCloud(
827
808
  token="YOUR_TOKEN",
828
- base_url="https://yourhost.com/path/to/api",
829
809
  )
830
810
  await client.parsing.get_job_raw_md_result(
831
811
  job_id="string",
@@ -860,7 +840,6 @@ class AsyncParsingClient:
860
840
 
861
841
  client = AsyncLlamaCloud(
862
842
  token="YOUR_TOKEN",
863
- base_url="https://yourhost.com/path/to/api",
864
843
  )
865
844
  await client.parsing.get_job_json_result(
866
845
  job_id="string",
@@ -893,7 +872,6 @@ class AsyncParsingClient:
893
872
 
894
873
  client = AsyncLlamaCloud(
895
874
  token="YOUR_TOKEN",
896
- base_url="https://yourhost.com/path/to/api",
897
875
  )
898
876
  await client.parsing.get_job_json_raw_result(
899
877
  job_id="string",
@@ -926,7 +904,6 @@ class AsyncParsingClient:
926
904
 
927
905
  client = AsyncLlamaCloud(
928
906
  token="YOUR_TOKEN",
929
- base_url="https://yourhost.com/path/to/api",
930
907
  )
931
908
  await client.parsing.get_parsing_history_result()
932
909
  """
@@ -959,7 +936,6 @@ class AsyncParsingClient:
959
936
 
960
937
  client = AsyncLlamaCloud(
961
938
  token="YOUR_TOKEN",
962
- base_url="https://yourhost.com/path/to/api",
963
939
  )
964
940
  await client.parsing.generate_presigned_url(
965
941
  job_id="string",
@@ -71,7 +71,6 @@ class PipelinesClient:
71
71
 
72
72
  client = LlamaCloud(
73
73
  token="YOUR_TOKEN",
74
- base_url="https://yourhost.com/path/to/api",
75
74
  )
76
75
  client.pipelines.search_pipelines(
77
76
  project_name="string",
@@ -121,7 +120,6 @@ class PipelinesClient:
121
120
 
122
121
  client = LlamaCloud(
123
122
  token="YOUR_TOKEN",
124
- base_url="https://yourhost.com/path/to/api",
125
123
  )
126
124
  client.pipelines.create_pipeline(
127
125
  request=PipelineCreate(
@@ -186,7 +184,6 @@ class PipelinesClient:
186
184
 
187
185
  client = LlamaCloud(
188
186
  token="YOUR_TOKEN",
189
- base_url="https://yourhost.com/path/to/api",
190
187
  )
191
188
  client.pipelines.upsert_pipeline(
192
189
  request=PipelineCreate(
@@ -241,7 +238,6 @@ class PipelinesClient:
241
238
 
242
239
  client = LlamaCloud(
243
240
  token="YOUR_TOKEN",
244
- base_url="https://yourhost.com/path/to/api",
245
241
  )
246
242
  client.pipelines.get_pipeline(
247
243
  pipeline_id="string",
@@ -312,7 +308,6 @@ class PipelinesClient:
312
308
 
313
309
  client = LlamaCloud(
314
310
  token="YOUR_TOKEN",
315
- base_url="https://yourhost.com/path/to/api",
316
311
  )
317
312
  client.pipelines.update_existing_pipeline(
318
313
  pipeline_id="string",
@@ -376,7 +371,6 @@ class PipelinesClient:
376
371
 
377
372
  client = LlamaCloud(
378
373
  token="YOUR_TOKEN",
379
- base_url="https://yourhost.com/path/to/api",
380
374
  )
381
375
  client.pipelines.delete_pipeline(
382
376
  pipeline_id="string",
@@ -409,7 +403,6 @@ class PipelinesClient:
409
403
 
410
404
  client = LlamaCloud(
411
405
  token="YOUR_TOKEN",
412
- base_url="https://yourhost.com/path/to/api",
413
406
  )
414
407
  client.pipelines.sync_pipeline(
415
408
  pipeline_id="string",
@@ -444,7 +437,6 @@ class PipelinesClient:
444
437
 
445
438
  client = LlamaCloud(
446
439
  token="YOUR_TOKEN",
447
- base_url="https://yourhost.com/path/to/api",
448
440
  )
449
441
  client.pipelines.get_eval_dataset_executions(
450
442
  pipeline_id="string",
@@ -495,7 +487,6 @@ class PipelinesClient:
495
487
 
496
488
  client = LlamaCloud(
497
489
  token="YOUR_TOKEN",
498
- base_url="https://yourhost.com/path/to/api",
499
490
  )
500
491
  client.pipelines.execute_eval_dataset(
501
492
  pipeline_id="string",
@@ -547,7 +538,6 @@ class PipelinesClient:
547
538
 
548
539
  client = LlamaCloud(
549
540
  token="YOUR_TOKEN",
550
- base_url="https://yourhost.com/path/to/api",
551
541
  )
552
542
  client.pipelines.get_eval_dataset_execution_result(
553
543
  pipeline_id="string",
@@ -590,7 +580,6 @@ class PipelinesClient:
590
580
 
591
581
  client = LlamaCloud(
592
582
  token="YOUR_TOKEN",
593
- base_url="https://yourhost.com/path/to/api",
594
583
  )
595
584
  client.pipelines.get_eval_dataset_execution(
596
585
  pipeline_id="string",
@@ -628,7 +617,6 @@ class PipelinesClient:
628
617
 
629
618
  client = LlamaCloud(
630
619
  token="YOUR_TOKEN",
631
- base_url="https://yourhost.com/path/to/api",
632
620
  )
633
621
  client.pipelines.get_files_for_pipeline(
634
622
  pipeline_id="string",
@@ -665,7 +653,6 @@ class PipelinesClient:
665
653
 
666
654
  client = LlamaCloud(
667
655
  token="YOUR_TOKEN",
668
- base_url="https://yourhost.com/path/to/api",
669
656
  )
670
657
  client.pipelines.add_files_to_pipeline(
671
658
  pipeline_id="string",
@@ -702,7 +689,6 @@ class PipelinesClient:
702
689
 
703
690
  client = LlamaCloud(
704
691
  token="YOUR_TOKEN",
705
- base_url="https://yourhost.com/path/to/api",
706
692
  )
707
693
  client.pipelines.get_pipeline_file_status(
708
694
  pipeline_id="string",
@@ -748,7 +734,6 @@ class PipelinesClient:
748
734
 
749
735
  client = LlamaCloud(
750
736
  token="YOUR_TOKEN",
751
- base_url="https://yourhost.com/path/to/api",
752
737
  )
753
738
  client.pipelines.update_pipeline_file(
754
739
  pipeline_id="string",
@@ -790,7 +775,6 @@ class PipelinesClient:
790
775
 
791
776
  client = LlamaCloud(
792
777
  token="YOUR_TOKEN",
793
- base_url="https://yourhost.com/path/to/api",
794
778
  )
795
779
  client.pipelines.delete_pipeline_file(
796
780
  pipeline_id="string",
@@ -826,7 +810,6 @@ class PipelinesClient:
826
810
 
827
811
  client = LlamaCloud(
828
812
  token="YOUR_TOKEN",
829
- base_url="https://yourhost.com/path/to/api",
830
813
  )
831
814
  client.pipelines.get_pipeline_data_sources(
832
815
  pipeline_id="string",
@@ -865,7 +848,6 @@ class PipelinesClient:
865
848
 
866
849
  client = LlamaCloud(
867
850
  token="YOUR_TOKEN",
868
- base_url="https://yourhost.com/path/to/api",
869
851
  )
870
852
  client.pipelines.add_data_sources_to_pipeline(
871
853
  pipeline_id="string",
@@ -904,7 +886,6 @@ class PipelinesClient:
904
886
 
905
887
  client = LlamaCloud(
906
888
  token="YOUR_TOKEN",
907
- base_url="https://yourhost.com/path/to/api",
908
889
  )
909
890
  client.pipelines.delete_pipeline_data_source(
910
891
  pipeline_id="string",
@@ -943,7 +924,6 @@ class PipelinesClient:
943
924
 
944
925
  client = LlamaCloud(
945
926
  token="YOUR_TOKEN",
946
- base_url="https://yourhost.com/path/to/api",
947
927
  )
948
928
  client.pipelines.sync_pipeline_data_source(
949
929
  pipeline_id="string",
@@ -1006,7 +986,6 @@ class PipelinesClient:
1006
986
 
1007
987
  client = LlamaCloud(
1008
988
  token="YOUR_TOKEN",
1009
- base_url="https://yourhost.com/path/to/api",
1010
989
  )
1011
990
  client.pipelines.run_search(
1012
991
  pipeline_id="string",
@@ -1058,7 +1037,6 @@ class PipelinesClient:
1058
1037
 
1059
1038
  client = LlamaCloud(
1060
1039
  token="YOUR_TOKEN",
1061
- base_url="https://yourhost.com/path/to/api",
1062
1040
  )
1063
1041
  client.pipelines.get_pipeline_jobs(
1064
1042
  pipeline_id="string",
@@ -1093,7 +1071,6 @@ class PipelinesClient:
1093
1071
 
1094
1072
  client = LlamaCloud(
1095
1073
  token="YOUR_TOKEN",
1096
- base_url="https://yourhost.com/path/to/api",
1097
1074
  )
1098
1075
  client.pipelines.get_pipeline_job(
1099
1076
  pipeline_id="string",
@@ -1135,7 +1112,6 @@ class PipelinesClient:
1135
1112
 
1136
1113
  client = LlamaCloud(
1137
1114
  token="YOUR_TOKEN",
1138
- base_url="https://yourhost.com/path/to/api",
1139
1115
  )
1140
1116
  client.pipelines.list_pipeline_documents(
1141
1117
  pipeline_id="string",
@@ -1175,7 +1151,6 @@ class PipelinesClient:
1175
1151
 
1176
1152
  client = LlamaCloud(
1177
1153
  token="YOUR_TOKEN",
1178
- base_url="https://yourhost.com/path/to/api",
1179
1154
  )
1180
1155
  client.pipelines.create_batch_pipeline_documents(
1181
1156
  pipeline_id="string",
@@ -1216,7 +1191,6 @@ class PipelinesClient:
1216
1191
 
1217
1192
  client = LlamaCloud(
1218
1193
  token="YOUR_TOKEN",
1219
- base_url="https://yourhost.com/path/to/api",
1220
1194
  )
1221
1195
  client.pipelines.upsert_batch_pipeline_documents(
1222
1196
  pipeline_id="string",
@@ -1255,7 +1229,6 @@ class PipelinesClient:
1255
1229
 
1256
1230
  client = LlamaCloud(
1257
1231
  token="YOUR_TOKEN",
1258
- base_url="https://yourhost.com/path/to/api",
1259
1232
  )
1260
1233
  client.pipelines.get_pipeline_document(
1261
1234
  pipeline_id="string",
@@ -1293,7 +1266,6 @@ class PipelinesClient:
1293
1266
 
1294
1267
  client = LlamaCloud(
1295
1268
  token="YOUR_TOKEN",
1296
- base_url="https://yourhost.com/path/to/api",
1297
1269
  )
1298
1270
  client.pipelines.delete_pipeline_document(
1299
1271
  pipeline_id="string",
@@ -1331,7 +1303,6 @@ class PipelinesClient:
1331
1303
 
1332
1304
  client = LlamaCloud(
1333
1305
  token="YOUR_TOKEN",
1334
- base_url="https://yourhost.com/path/to/api",
1335
1306
  )
1336
1307
  client.pipelines.get_pipeline_document_status(
1337
1308
  pipeline_id="string",
@@ -1384,7 +1355,6 @@ class AsyncPipelinesClient:
1384
1355
 
1385
1356
  client = AsyncLlamaCloud(
1386
1357
  token="YOUR_TOKEN",
1387
- base_url="https://yourhost.com/path/to/api",
1388
1358
  )
1389
1359
  await client.pipelines.search_pipelines(
1390
1360
  project_name="string",
@@ -1434,7 +1404,6 @@ class AsyncPipelinesClient:
1434
1404
 
1435
1405
  client = AsyncLlamaCloud(
1436
1406
  token="YOUR_TOKEN",
1437
- base_url="https://yourhost.com/path/to/api",
1438
1407
  )
1439
1408
  await client.pipelines.create_pipeline(
1440
1409
  request=PipelineCreate(
@@ -1499,7 +1468,6 @@ class AsyncPipelinesClient:
1499
1468
 
1500
1469
  client = AsyncLlamaCloud(
1501
1470
  token="YOUR_TOKEN",
1502
- base_url="https://yourhost.com/path/to/api",
1503
1471
  )
1504
1472
  await client.pipelines.upsert_pipeline(
1505
1473
  request=PipelineCreate(
@@ -1554,7 +1522,6 @@ class AsyncPipelinesClient:
1554
1522
 
1555
1523
  client = AsyncLlamaCloud(
1556
1524
  token="YOUR_TOKEN",
1557
- base_url="https://yourhost.com/path/to/api",
1558
1525
  )
1559
1526
  await client.pipelines.get_pipeline(
1560
1527
  pipeline_id="string",
@@ -1625,7 +1592,6 @@ class AsyncPipelinesClient:
1625
1592
 
1626
1593
  client = AsyncLlamaCloud(
1627
1594
  token="YOUR_TOKEN",
1628
- base_url="https://yourhost.com/path/to/api",
1629
1595
  )
1630
1596
  await client.pipelines.update_existing_pipeline(
1631
1597
  pipeline_id="string",
@@ -1689,7 +1655,6 @@ class AsyncPipelinesClient:
1689
1655
 
1690
1656
  client = AsyncLlamaCloud(
1691
1657
  token="YOUR_TOKEN",
1692
- base_url="https://yourhost.com/path/to/api",
1693
1658
  )
1694
1659
  await client.pipelines.delete_pipeline(
1695
1660
  pipeline_id="string",
@@ -1722,7 +1687,6 @@ class AsyncPipelinesClient:
1722
1687
 
1723
1688
  client = AsyncLlamaCloud(
1724
1689
  token="YOUR_TOKEN",
1725
- base_url="https://yourhost.com/path/to/api",
1726
1690
  )
1727
1691
  await client.pipelines.sync_pipeline(
1728
1692
  pipeline_id="string",
@@ -1759,7 +1723,6 @@ class AsyncPipelinesClient:
1759
1723
 
1760
1724
  client = AsyncLlamaCloud(
1761
1725
  token="YOUR_TOKEN",
1762
- base_url="https://yourhost.com/path/to/api",
1763
1726
  )
1764
1727
  await client.pipelines.get_eval_dataset_executions(
1765
1728
  pipeline_id="string",
@@ -1810,7 +1773,6 @@ class AsyncPipelinesClient:
1810
1773
 
1811
1774
  client = AsyncLlamaCloud(
1812
1775
  token="YOUR_TOKEN",
1813
- base_url="https://yourhost.com/path/to/api",
1814
1776
  )
1815
1777
  await client.pipelines.execute_eval_dataset(
1816
1778
  pipeline_id="string",
@@ -1862,7 +1824,6 @@ class AsyncPipelinesClient:
1862
1824
 
1863
1825
  client = AsyncLlamaCloud(
1864
1826
  token="YOUR_TOKEN",
1865
- base_url="https://yourhost.com/path/to/api",
1866
1827
  )
1867
1828
  await client.pipelines.get_eval_dataset_execution_result(
1868
1829
  pipeline_id="string",
@@ -1905,7 +1866,6 @@ class AsyncPipelinesClient:
1905
1866
 
1906
1867
  client = AsyncLlamaCloud(
1907
1868
  token="YOUR_TOKEN",
1908
- base_url="https://yourhost.com/path/to/api",
1909
1869
  )
1910
1870
  await client.pipelines.get_eval_dataset_execution(
1911
1871
  pipeline_id="string",
@@ -1943,7 +1903,6 @@ class AsyncPipelinesClient:
1943
1903
 
1944
1904
  client = AsyncLlamaCloud(
1945
1905
  token="YOUR_TOKEN",
1946
- base_url="https://yourhost.com/path/to/api",
1947
1906
  )
1948
1907
  await client.pipelines.get_files_for_pipeline(
1949
1908
  pipeline_id="string",
@@ -1980,7 +1939,6 @@ class AsyncPipelinesClient:
1980
1939
 
1981
1940
  client = AsyncLlamaCloud(
1982
1941
  token="YOUR_TOKEN",
1983
- base_url="https://yourhost.com/path/to/api",
1984
1942
  )
1985
1943
  await client.pipelines.add_files_to_pipeline(
1986
1944
  pipeline_id="string",
@@ -2017,7 +1975,6 @@ class AsyncPipelinesClient:
2017
1975
 
2018
1976
  client = AsyncLlamaCloud(
2019
1977
  token="YOUR_TOKEN",
2020
- base_url="https://yourhost.com/path/to/api",
2021
1978
  )
2022
1979
  await client.pipelines.get_pipeline_file_status(
2023
1980
  pipeline_id="string",
@@ -2063,7 +2020,6 @@ class AsyncPipelinesClient:
2063
2020
 
2064
2021
  client = AsyncLlamaCloud(
2065
2022
  token="YOUR_TOKEN",
2066
- base_url="https://yourhost.com/path/to/api",
2067
2023
  )
2068
2024
  await client.pipelines.update_pipeline_file(
2069
2025
  pipeline_id="string",
@@ -2105,7 +2061,6 @@ class AsyncPipelinesClient:
2105
2061
 
2106
2062
  client = AsyncLlamaCloud(
2107
2063
  token="YOUR_TOKEN",
2108
- base_url="https://yourhost.com/path/to/api",
2109
2064
  )
2110
2065
  await client.pipelines.delete_pipeline_file(
2111
2066
  pipeline_id="string",
@@ -2141,7 +2096,6 @@ class AsyncPipelinesClient:
2141
2096
 
2142
2097
  client = AsyncLlamaCloud(
2143
2098
  token="YOUR_TOKEN",
2144
- base_url="https://yourhost.com/path/to/api",
2145
2099
  )
2146
2100
  await client.pipelines.get_pipeline_data_sources(
2147
2101
  pipeline_id="string",
@@ -2180,7 +2134,6 @@ class AsyncPipelinesClient:
2180
2134
 
2181
2135
  client = AsyncLlamaCloud(
2182
2136
  token="YOUR_TOKEN",
2183
- base_url="https://yourhost.com/path/to/api",
2184
2137
  )
2185
2138
  await client.pipelines.add_data_sources_to_pipeline(
2186
2139
  pipeline_id="string",
@@ -2219,7 +2172,6 @@ class AsyncPipelinesClient:
2219
2172
 
2220
2173
  client = AsyncLlamaCloud(
2221
2174
  token="YOUR_TOKEN",
2222
- base_url="https://yourhost.com/path/to/api",
2223
2175
  )
2224
2176
  await client.pipelines.delete_pipeline_data_source(
2225
2177
  pipeline_id="string",
@@ -2258,7 +2210,6 @@ class AsyncPipelinesClient:
2258
2210
 
2259
2211
  client = AsyncLlamaCloud(
2260
2212
  token="YOUR_TOKEN",
2261
- base_url="https://yourhost.com/path/to/api",
2262
2213
  )
2263
2214
  await client.pipelines.sync_pipeline_data_source(
2264
2215
  pipeline_id="string",
@@ -2321,7 +2272,6 @@ class AsyncPipelinesClient:
2321
2272
 
2322
2273
  client = AsyncLlamaCloud(
2323
2274
  token="YOUR_TOKEN",
2324
- base_url="https://yourhost.com/path/to/api",
2325
2275
  )
2326
2276
  await client.pipelines.run_search(
2327
2277
  pipeline_id="string",
@@ -2373,7 +2323,6 @@ class AsyncPipelinesClient:
2373
2323
 
2374
2324
  client = AsyncLlamaCloud(
2375
2325
  token="YOUR_TOKEN",
2376
- base_url="https://yourhost.com/path/to/api",
2377
2326
  )
2378
2327
  await client.pipelines.get_pipeline_jobs(
2379
2328
  pipeline_id="string",
@@ -2408,7 +2357,6 @@ class AsyncPipelinesClient:
2408
2357
 
2409
2358
  client = AsyncLlamaCloud(
2410
2359
  token="YOUR_TOKEN",
2411
- base_url="https://yourhost.com/path/to/api",
2412
2360
  )
2413
2361
  await client.pipelines.get_pipeline_job(
2414
2362
  pipeline_id="string",
@@ -2450,7 +2398,6 @@ class AsyncPipelinesClient:
2450
2398
 
2451
2399
  client = AsyncLlamaCloud(
2452
2400
  token="YOUR_TOKEN",
2453
- base_url="https://yourhost.com/path/to/api",
2454
2401
  )
2455
2402
  await client.pipelines.list_pipeline_documents(
2456
2403
  pipeline_id="string",
@@ -2490,7 +2437,6 @@ class AsyncPipelinesClient:
2490
2437
 
2491
2438
  client = AsyncLlamaCloud(
2492
2439
  token="YOUR_TOKEN",
2493
- base_url="https://yourhost.com/path/to/api",
2494
2440
  )
2495
2441
  await client.pipelines.create_batch_pipeline_documents(
2496
2442
  pipeline_id="string",
@@ -2531,7 +2477,6 @@ class AsyncPipelinesClient:
2531
2477
 
2532
2478
  client = AsyncLlamaCloud(
2533
2479
  token="YOUR_TOKEN",
2534
- base_url="https://yourhost.com/path/to/api",
2535
2480
  )
2536
2481
  await client.pipelines.upsert_batch_pipeline_documents(
2537
2482
  pipeline_id="string",
@@ -2570,7 +2515,6 @@ class AsyncPipelinesClient:
2570
2515
 
2571
2516
  client = AsyncLlamaCloud(
2572
2517
  token="YOUR_TOKEN",
2573
- base_url="https://yourhost.com/path/to/api",
2574
2518
  )
2575
2519
  await client.pipelines.get_pipeline_document(
2576
2520
  pipeline_id="string",
@@ -2608,7 +2552,6 @@ class AsyncPipelinesClient:
2608
2552
 
2609
2553
  client = AsyncLlamaCloud(
2610
2554
  token="YOUR_TOKEN",
2611
- base_url="https://yourhost.com/path/to/api",
2612
2555
  )
2613
2556
  await client.pipelines.delete_pipeline_document(
2614
2557
  pipeline_id="string",
@@ -2646,7 +2589,6 @@ class AsyncPipelinesClient:
2646
2589
 
2647
2590
  client = AsyncLlamaCloud(
2648
2591
  token="YOUR_TOKEN",
2649
- base_url="https://yourhost.com/path/to/api",
2650
2592
  )
2651
2593
  await client.pipelines.get_pipeline_document_status(
2652
2594
  pipeline_id="string",