llama-cloud 0.1.39__py3-none-any.whl → 0.1.41__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_cloud/__init__.py +24 -72
- llama_cloud/client.py +2 -5
- llama_cloud/resources/__init__.py +0 -4
- llama_cloud/resources/admin/client.py +5 -5
- llama_cloud/resources/beta/client.py +1351 -335
- llama_cloud/resources/llama_extract/client.py +56 -0
- llama_cloud/resources/parsing/client.py +8 -0
- llama_cloud/resources/pipelines/client.py +37 -0
- llama_cloud/types/__init__.py +24 -72
- llama_cloud/types/{src_app_schema_chat_chat_message.py → chat_message.py} +1 -1
- llama_cloud/types/extract_config.py +8 -2
- llama_cloud/types/extract_job_create.py +2 -0
- llama_cloud/types/extract_job_create_priority.py +29 -0
- llama_cloud/types/extract_models.py +28 -28
- llama_cloud/types/job_names.py +0 -4
- llama_cloud/types/{document_block.py → llama_extract_feature_availability.py} +5 -6
- llama_cloud/types/llama_extract_mode_availability.py +4 -3
- llama_cloud/types/llama_extract_settings.py +1 -1
- llama_cloud/types/llama_parse_parameters.py +1 -0
- llama_cloud/types/{progress_event.py → parse_configuration.py} +12 -12
- llama_cloud/types/{llama_index_core_base_llms_types_chat_message.py → parse_configuration_create.py} +9 -7
- llama_cloud/types/{edit_suggestion.py → parse_configuration_filter.py} +8 -6
- llama_cloud/types/{report_update_event.py → parse_configuration_query_response.py} +6 -6
- llama_cloud/types/parse_job_config.py +1 -0
- llama_cloud/types/pipeline.py +4 -0
- llama_cloud/types/pipeline_create.py +2 -0
- llama_cloud/types/playground_session.py +2 -2
- llama_cloud/types/public_model_name.py +97 -0
- llama_cloud/types/{report_create_response.py → schema_generation_availability.py} +4 -2
- llama_cloud/types/schema_generation_availability_status.py +17 -0
- llama_cloud/types/{report_event_item.py → sparse_model_config.py} +10 -8
- llama_cloud/types/sparse_model_type.py +33 -0
- llama_cloud/types/webhook_configuration.py +1 -0
- llama_cloud-0.1.41.dist-info/METADATA +106 -0
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/RECORD +37 -56
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/WHEEL +1 -1
- llama_cloud/resources/reports/__init__.py +0 -5
- llama_cloud/resources/reports/client.py +0 -1230
- llama_cloud/resources/reports/types/__init__.py +0 -7
- llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +0 -25
- llama_cloud/types/audio_block.py +0 -34
- llama_cloud/types/edit_suggestion_blocks_item.py +0 -8
- llama_cloud/types/image_block.py +0 -35
- llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +0 -56
- llama_cloud/types/paginated_report_response.py +0 -35
- llama_cloud/types/progress_event_status.py +0 -33
- llama_cloud/types/report.py +0 -33
- llama_cloud/types/report_block.py +0 -35
- llama_cloud/types/report_block_dependency.py +0 -29
- llama_cloud/types/report_event_item_event_data.py +0 -45
- llama_cloud/types/report_event_type.py +0 -37
- llama_cloud/types/report_metadata.py +0 -43
- llama_cloud/types/report_plan.py +0 -36
- llama_cloud/types/report_plan_block.py +0 -36
- llama_cloud/types/report_query.py +0 -33
- llama_cloud/types/report_response.py +0 -41
- llama_cloud/types/report_state.py +0 -37
- llama_cloud/types/report_state_event.py +0 -38
- llama_cloud/types/text_block.py +0 -31
- llama_cloud-0.1.39.dist-info/METADATA +0 -32
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/LICENSE +0 -0
|
@@ -111,7 +111,9 @@ class LlamaExtractClient:
|
|
|
111
111
|
ExtractConfig,
|
|
112
112
|
ExtractConfigPriority,
|
|
113
113
|
ExtractMode,
|
|
114
|
+
ExtractModels,
|
|
114
115
|
ExtractTarget,
|
|
116
|
+
PublicModelName,
|
|
115
117
|
)
|
|
116
118
|
from llama_cloud.client import LlamaCloud
|
|
117
119
|
|
|
@@ -124,6 +126,8 @@ class LlamaExtractClient:
|
|
|
124
126
|
priority=ExtractConfigPriority.LOW,
|
|
125
127
|
extraction_target=ExtractTarget.PER_DOC,
|
|
126
128
|
extraction_mode=ExtractMode.FAST,
|
|
129
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
130
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
127
131
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
128
132
|
),
|
|
129
133
|
)
|
|
@@ -359,7 +363,9 @@ class LlamaExtractClient:
|
|
|
359
363
|
ExtractConfig,
|
|
360
364
|
ExtractConfigPriority,
|
|
361
365
|
ExtractMode,
|
|
366
|
+
ExtractModels,
|
|
362
367
|
ExtractTarget,
|
|
368
|
+
PublicModelName,
|
|
363
369
|
)
|
|
364
370
|
from llama_cloud.client import LlamaCloud
|
|
365
371
|
|
|
@@ -372,6 +378,8 @@ class LlamaExtractClient:
|
|
|
372
378
|
priority=ExtractConfigPriority.LOW,
|
|
373
379
|
extraction_target=ExtractTarget.PER_DOC,
|
|
374
380
|
extraction_mode=ExtractMode.FAST,
|
|
381
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
382
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
375
383
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
376
384
|
),
|
|
377
385
|
)
|
|
@@ -470,8 +478,11 @@ class LlamaExtractClient:
|
|
|
470
478
|
ExtractConfig,
|
|
471
479
|
ExtractConfigPriority,
|
|
472
480
|
ExtractJobCreate,
|
|
481
|
+
ExtractJobCreatePriority,
|
|
473
482
|
ExtractMode,
|
|
483
|
+
ExtractModels,
|
|
474
484
|
ExtractTarget,
|
|
485
|
+
PublicModelName,
|
|
475
486
|
)
|
|
476
487
|
from llama_cloud.client import LlamaCloud
|
|
477
488
|
|
|
@@ -480,12 +491,15 @@ class LlamaExtractClient:
|
|
|
480
491
|
)
|
|
481
492
|
client.llama_extract.run_job(
|
|
482
493
|
request=ExtractJobCreate(
|
|
494
|
+
priority=ExtractJobCreatePriority.LOW,
|
|
483
495
|
extraction_agent_id="string",
|
|
484
496
|
file_id="string",
|
|
485
497
|
config_override=ExtractConfig(
|
|
486
498
|
priority=ExtractConfigPriority.LOW,
|
|
487
499
|
extraction_target=ExtractTarget.PER_DOC,
|
|
488
500
|
extraction_mode=ExtractMode.FAST,
|
|
501
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
502
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
489
503
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
490
504
|
),
|
|
491
505
|
),
|
|
@@ -560,7 +574,9 @@ class LlamaExtractClient:
|
|
|
560
574
|
ExtractConfig,
|
|
561
575
|
ExtractConfigPriority,
|
|
562
576
|
ExtractJobCreate,
|
|
577
|
+
ExtractJobCreatePriority,
|
|
563
578
|
ExtractMode,
|
|
579
|
+
ExtractModels,
|
|
564
580
|
ExtractTarget,
|
|
565
581
|
FailPageMode,
|
|
566
582
|
LlamaExtractSettings,
|
|
@@ -568,6 +584,7 @@ class LlamaExtractClient:
|
|
|
568
584
|
LlamaParseParametersPriority,
|
|
569
585
|
MultimodalParseResolution,
|
|
570
586
|
ParsingMode,
|
|
587
|
+
PublicModelName,
|
|
571
588
|
)
|
|
572
589
|
from llama_cloud.client import LlamaCloud
|
|
573
590
|
|
|
@@ -576,12 +593,15 @@ class LlamaExtractClient:
|
|
|
576
593
|
)
|
|
577
594
|
client.llama_extract.run_job_test_user(
|
|
578
595
|
job_create=ExtractJobCreate(
|
|
596
|
+
priority=ExtractJobCreatePriority.LOW,
|
|
579
597
|
extraction_agent_id="string",
|
|
580
598
|
file_id="string",
|
|
581
599
|
config_override=ExtractConfig(
|
|
582
600
|
priority=ExtractConfigPriority.LOW,
|
|
583
601
|
extraction_target=ExtractTarget.PER_DOC,
|
|
584
602
|
extraction_mode=ExtractMode.FAST,
|
|
603
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
604
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
585
605
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
586
606
|
),
|
|
587
607
|
),
|
|
@@ -689,7 +709,9 @@ class LlamaExtractClient:
|
|
|
689
709
|
ExtractConfig,
|
|
690
710
|
ExtractConfigPriority,
|
|
691
711
|
ExtractMode,
|
|
712
|
+
ExtractModels,
|
|
692
713
|
ExtractTarget,
|
|
714
|
+
PublicModelName,
|
|
693
715
|
)
|
|
694
716
|
from llama_cloud.client import LlamaCloud
|
|
695
717
|
|
|
@@ -703,6 +725,8 @@ class LlamaExtractClient:
|
|
|
703
725
|
priority=ExtractConfigPriority.LOW,
|
|
704
726
|
extraction_target=ExtractTarget.PER_DOC,
|
|
705
727
|
extraction_mode=ExtractMode.FAST,
|
|
728
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
729
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
706
730
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
707
731
|
),
|
|
708
732
|
)
|
|
@@ -984,8 +1008,10 @@ class LlamaExtractClient:
|
|
|
984
1008
|
ExtractConfig,
|
|
985
1009
|
ExtractConfigPriority,
|
|
986
1010
|
ExtractMode,
|
|
1011
|
+
ExtractModels,
|
|
987
1012
|
ExtractTarget,
|
|
988
1013
|
FileData,
|
|
1014
|
+
PublicModelName,
|
|
989
1015
|
)
|
|
990
1016
|
from llama_cloud.client import LlamaCloud
|
|
991
1017
|
|
|
@@ -997,6 +1023,8 @@ class LlamaExtractClient:
|
|
|
997
1023
|
priority=ExtractConfigPriority.LOW,
|
|
998
1024
|
extraction_target=ExtractTarget.PER_DOC,
|
|
999
1025
|
extraction_mode=ExtractMode.FAST,
|
|
1026
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
1027
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1000
1028
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1001
1029
|
),
|
|
1002
1030
|
file=FileData(
|
|
@@ -1104,7 +1132,9 @@ class AsyncLlamaExtractClient:
|
|
|
1104
1132
|
ExtractConfig,
|
|
1105
1133
|
ExtractConfigPriority,
|
|
1106
1134
|
ExtractMode,
|
|
1135
|
+
ExtractModels,
|
|
1107
1136
|
ExtractTarget,
|
|
1137
|
+
PublicModelName,
|
|
1108
1138
|
)
|
|
1109
1139
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1110
1140
|
|
|
@@ -1117,6 +1147,8 @@ class AsyncLlamaExtractClient:
|
|
|
1117
1147
|
priority=ExtractConfigPriority.LOW,
|
|
1118
1148
|
extraction_target=ExtractTarget.PER_DOC,
|
|
1119
1149
|
extraction_mode=ExtractMode.FAST,
|
|
1150
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
1151
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1120
1152
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1121
1153
|
),
|
|
1122
1154
|
)
|
|
@@ -1352,7 +1384,9 @@ class AsyncLlamaExtractClient:
|
|
|
1352
1384
|
ExtractConfig,
|
|
1353
1385
|
ExtractConfigPriority,
|
|
1354
1386
|
ExtractMode,
|
|
1387
|
+
ExtractModels,
|
|
1355
1388
|
ExtractTarget,
|
|
1389
|
+
PublicModelName,
|
|
1356
1390
|
)
|
|
1357
1391
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1358
1392
|
|
|
@@ -1365,6 +1399,8 @@ class AsyncLlamaExtractClient:
|
|
|
1365
1399
|
priority=ExtractConfigPriority.LOW,
|
|
1366
1400
|
extraction_target=ExtractTarget.PER_DOC,
|
|
1367
1401
|
extraction_mode=ExtractMode.FAST,
|
|
1402
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
1403
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1368
1404
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1369
1405
|
),
|
|
1370
1406
|
)
|
|
@@ -1463,8 +1499,11 @@ class AsyncLlamaExtractClient:
|
|
|
1463
1499
|
ExtractConfig,
|
|
1464
1500
|
ExtractConfigPriority,
|
|
1465
1501
|
ExtractJobCreate,
|
|
1502
|
+
ExtractJobCreatePriority,
|
|
1466
1503
|
ExtractMode,
|
|
1504
|
+
ExtractModels,
|
|
1467
1505
|
ExtractTarget,
|
|
1506
|
+
PublicModelName,
|
|
1468
1507
|
)
|
|
1469
1508
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1470
1509
|
|
|
@@ -1473,12 +1512,15 @@ class AsyncLlamaExtractClient:
|
|
|
1473
1512
|
)
|
|
1474
1513
|
await client.llama_extract.run_job(
|
|
1475
1514
|
request=ExtractJobCreate(
|
|
1515
|
+
priority=ExtractJobCreatePriority.LOW,
|
|
1476
1516
|
extraction_agent_id="string",
|
|
1477
1517
|
file_id="string",
|
|
1478
1518
|
config_override=ExtractConfig(
|
|
1479
1519
|
priority=ExtractConfigPriority.LOW,
|
|
1480
1520
|
extraction_target=ExtractTarget.PER_DOC,
|
|
1481
1521
|
extraction_mode=ExtractMode.FAST,
|
|
1522
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
1523
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1482
1524
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1483
1525
|
),
|
|
1484
1526
|
),
|
|
@@ -1553,7 +1595,9 @@ class AsyncLlamaExtractClient:
|
|
|
1553
1595
|
ExtractConfig,
|
|
1554
1596
|
ExtractConfigPriority,
|
|
1555
1597
|
ExtractJobCreate,
|
|
1598
|
+
ExtractJobCreatePriority,
|
|
1556
1599
|
ExtractMode,
|
|
1600
|
+
ExtractModels,
|
|
1557
1601
|
ExtractTarget,
|
|
1558
1602
|
FailPageMode,
|
|
1559
1603
|
LlamaExtractSettings,
|
|
@@ -1561,6 +1605,7 @@ class AsyncLlamaExtractClient:
|
|
|
1561
1605
|
LlamaParseParametersPriority,
|
|
1562
1606
|
MultimodalParseResolution,
|
|
1563
1607
|
ParsingMode,
|
|
1608
|
+
PublicModelName,
|
|
1564
1609
|
)
|
|
1565
1610
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1566
1611
|
|
|
@@ -1569,12 +1614,15 @@ class AsyncLlamaExtractClient:
|
|
|
1569
1614
|
)
|
|
1570
1615
|
await client.llama_extract.run_job_test_user(
|
|
1571
1616
|
job_create=ExtractJobCreate(
|
|
1617
|
+
priority=ExtractJobCreatePriority.LOW,
|
|
1572
1618
|
extraction_agent_id="string",
|
|
1573
1619
|
file_id="string",
|
|
1574
1620
|
config_override=ExtractConfig(
|
|
1575
1621
|
priority=ExtractConfigPriority.LOW,
|
|
1576
1622
|
extraction_target=ExtractTarget.PER_DOC,
|
|
1577
1623
|
extraction_mode=ExtractMode.FAST,
|
|
1624
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
1625
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1578
1626
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1579
1627
|
),
|
|
1580
1628
|
),
|
|
@@ -1682,7 +1730,9 @@ class AsyncLlamaExtractClient:
|
|
|
1682
1730
|
ExtractConfig,
|
|
1683
1731
|
ExtractConfigPriority,
|
|
1684
1732
|
ExtractMode,
|
|
1733
|
+
ExtractModels,
|
|
1685
1734
|
ExtractTarget,
|
|
1735
|
+
PublicModelName,
|
|
1686
1736
|
)
|
|
1687
1737
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1688
1738
|
|
|
@@ -1696,6 +1746,8 @@ class AsyncLlamaExtractClient:
|
|
|
1696
1746
|
priority=ExtractConfigPriority.LOW,
|
|
1697
1747
|
extraction_target=ExtractTarget.PER_DOC,
|
|
1698
1748
|
extraction_mode=ExtractMode.FAST,
|
|
1749
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
1750
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1699
1751
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1700
1752
|
),
|
|
1701
1753
|
)
|
|
@@ -1977,8 +2029,10 @@ class AsyncLlamaExtractClient:
|
|
|
1977
2029
|
ExtractConfig,
|
|
1978
2030
|
ExtractConfigPriority,
|
|
1979
2031
|
ExtractMode,
|
|
2032
|
+
ExtractModels,
|
|
1980
2033
|
ExtractTarget,
|
|
1981
2034
|
FileData,
|
|
2035
|
+
PublicModelName,
|
|
1982
2036
|
)
|
|
1983
2037
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1984
2038
|
|
|
@@ -1990,6 +2044,8 @@ class AsyncLlamaExtractClient:
|
|
|
1990
2044
|
priority=ExtractConfigPriority.LOW,
|
|
1991
2045
|
extraction_target=ExtractTarget.PER_DOC,
|
|
1992
2046
|
extraction_mode=ExtractMode.FAST,
|
|
2047
|
+
parse_model=PublicModelName.OPENAI_GPT_4_O,
|
|
2048
|
+
extract_model=ExtractModels.OPENAI_GPT_4_1,
|
|
1993
2049
|
chunk_mode=DocumentChunkMode.PAGE,
|
|
1994
2050
|
),
|
|
1995
2051
|
file=FileData(
|
|
@@ -227,6 +227,7 @@ class ParsingClient:
|
|
|
227
227
|
guess_xlsx_sheet_name: bool,
|
|
228
228
|
high_res_ocr: bool,
|
|
229
229
|
html_make_all_elements_visible: bool,
|
|
230
|
+
layout_aware: bool,
|
|
230
231
|
html_remove_fixed_elements: bool,
|
|
231
232
|
html_remove_navigation_elements: bool,
|
|
232
233
|
http_proxy: str,
|
|
@@ -357,6 +358,8 @@ class ParsingClient:
|
|
|
357
358
|
|
|
358
359
|
- html_make_all_elements_visible: bool.
|
|
359
360
|
|
|
361
|
+
- layout_aware: bool.
|
|
362
|
+
|
|
360
363
|
- html_remove_fixed_elements: bool.
|
|
361
364
|
|
|
362
365
|
- html_remove_navigation_elements: bool.
|
|
@@ -522,6 +525,7 @@ class ParsingClient:
|
|
|
522
525
|
"guess_xlsx_sheet_name": guess_xlsx_sheet_name,
|
|
523
526
|
"high_res_ocr": high_res_ocr,
|
|
524
527
|
"html_make_all_elements_visible": html_make_all_elements_visible,
|
|
528
|
+
"layout_aware": layout_aware,
|
|
525
529
|
"html_remove_fixed_elements": html_remove_fixed_elements,
|
|
526
530
|
"html_remove_navigation_elements": html_remove_navigation_elements,
|
|
527
531
|
"http_proxy": http_proxy,
|
|
@@ -1399,6 +1403,7 @@ class AsyncParsingClient:
|
|
|
1399
1403
|
guess_xlsx_sheet_name: bool,
|
|
1400
1404
|
high_res_ocr: bool,
|
|
1401
1405
|
html_make_all_elements_visible: bool,
|
|
1406
|
+
layout_aware: bool,
|
|
1402
1407
|
html_remove_fixed_elements: bool,
|
|
1403
1408
|
html_remove_navigation_elements: bool,
|
|
1404
1409
|
http_proxy: str,
|
|
@@ -1529,6 +1534,8 @@ class AsyncParsingClient:
|
|
|
1529
1534
|
|
|
1530
1535
|
- html_make_all_elements_visible: bool.
|
|
1531
1536
|
|
|
1537
|
+
- layout_aware: bool.
|
|
1538
|
+
|
|
1532
1539
|
- html_remove_fixed_elements: bool.
|
|
1533
1540
|
|
|
1534
1541
|
- html_remove_navigation_elements: bool.
|
|
@@ -1694,6 +1701,7 @@ class AsyncParsingClient:
|
|
|
1694
1701
|
"guess_xlsx_sheet_name": guess_xlsx_sheet_name,
|
|
1695
1702
|
"high_res_ocr": high_res_ocr,
|
|
1696
1703
|
"html_make_all_elements_visible": html_make_all_elements_visible,
|
|
1704
|
+
"layout_aware": layout_aware,
|
|
1697
1705
|
"html_remove_fixed_elements": html_remove_fixed_elements,
|
|
1698
1706
|
"html_remove_navigation_elements": html_remove_navigation_elements,
|
|
1699
1707
|
"http_proxy": http_proxy,
|
|
@@ -35,6 +35,7 @@ from ...types.playground_session import PlaygroundSession
|
|
|
35
35
|
from ...types.preset_retrieval_params import PresetRetrievalParams
|
|
36
36
|
from ...types.retrieval_mode import RetrievalMode
|
|
37
37
|
from ...types.retrieve_results import RetrieveResults
|
|
38
|
+
from ...types.sparse_model_config import SparseModelConfig
|
|
38
39
|
from ...types.text_node import TextNode
|
|
39
40
|
from .types.pipeline_file_update_custom_metadata_value import PipelineFileUpdateCustomMetadataValue
|
|
40
41
|
from .types.pipeline_update_embedding_config import PipelineUpdateEmbeddingConfig
|
|
@@ -217,6 +218,7 @@ class PipelinesClient:
|
|
|
217
218
|
*,
|
|
218
219
|
embedding_config: typing.Optional[PipelineUpdateEmbeddingConfig] = OMIT,
|
|
219
220
|
transform_config: typing.Optional[PipelineUpdateTransformConfig] = OMIT,
|
|
221
|
+
sparse_model_config: typing.Optional[SparseModelConfig] = OMIT,
|
|
220
222
|
data_sink_id: typing.Optional[str] = OMIT,
|
|
221
223
|
embedding_model_config_id: typing.Optional[str] = OMIT,
|
|
222
224
|
data_sink: typing.Optional[DataSinkCreate] = OMIT,
|
|
@@ -238,6 +240,8 @@ class PipelinesClient:
|
|
|
238
240
|
|
|
239
241
|
- transform_config: typing.Optional[PipelineUpdateTransformConfig]. Configuration for the transformation.
|
|
240
242
|
|
|
243
|
+
- sparse_model_config: typing.Optional[SparseModelConfig].
|
|
244
|
+
|
|
241
245
|
- data_sink_id: typing.Optional[str].
|
|
242
246
|
|
|
243
247
|
- embedding_model_config_id: typing.Optional[str].
|
|
@@ -263,6 +267,8 @@ class PipelinesClient:
|
|
|
263
267
|
_request["embedding_config"] = embedding_config
|
|
264
268
|
if transform_config is not OMIT:
|
|
265
269
|
_request["transform_config"] = transform_config
|
|
270
|
+
if sparse_model_config is not OMIT:
|
|
271
|
+
_request["sparse_model_config"] = sparse_model_config
|
|
266
272
|
if data_sink_id is not OMIT:
|
|
267
273
|
_request["data_sink_id"] = data_sink_id
|
|
268
274
|
if embedding_model_config_id is not OMIT:
|
|
@@ -554,6 +560,7 @@ class PipelinesClient:
|
|
|
554
560
|
*,
|
|
555
561
|
data_source_id: typing.Optional[str] = None,
|
|
556
562
|
only_manually_uploaded: typing.Optional[bool] = None,
|
|
563
|
+
file_name_contains: typing.Optional[str] = None,
|
|
557
564
|
limit: typing.Optional[int] = None,
|
|
558
565
|
offset: typing.Optional[int] = None,
|
|
559
566
|
order_by: typing.Optional[str] = None,
|
|
@@ -561,6 +568,15 @@ class PipelinesClient:
|
|
|
561
568
|
"""
|
|
562
569
|
Get files for a pipeline.
|
|
563
570
|
|
|
571
|
+
Args:
|
|
572
|
+
pipeline_id: ID of the pipeline
|
|
573
|
+
data_source_id: Optional filter by data source ID
|
|
574
|
+
only_manually_uploaded: Filter for only manually uploaded files
|
|
575
|
+
file_name_contains: Optional filter by file name (substring match)
|
|
576
|
+
limit: Limit number of results
|
|
577
|
+
offset: Offset for pagination
|
|
578
|
+
order_by: Field to order by
|
|
579
|
+
|
|
564
580
|
Parameters:
|
|
565
581
|
- pipeline_id: str.
|
|
566
582
|
|
|
@@ -568,6 +584,8 @@ class PipelinesClient:
|
|
|
568
584
|
|
|
569
585
|
- only_manually_uploaded: typing.Optional[bool].
|
|
570
586
|
|
|
587
|
+
- file_name_contains: typing.Optional[str].
|
|
588
|
+
|
|
571
589
|
- limit: typing.Optional[int].
|
|
572
590
|
|
|
573
591
|
- offset: typing.Optional[int].
|
|
@@ -590,6 +608,7 @@ class PipelinesClient:
|
|
|
590
608
|
{
|
|
591
609
|
"data_source_id": data_source_id,
|
|
592
610
|
"only_manually_uploaded": only_manually_uploaded,
|
|
611
|
+
"file_name_contains": file_name_contains,
|
|
593
612
|
"limit": limit,
|
|
594
613
|
"offset": offset,
|
|
595
614
|
"order_by": order_by,
|
|
@@ -1888,6 +1907,7 @@ class AsyncPipelinesClient:
|
|
|
1888
1907
|
*,
|
|
1889
1908
|
embedding_config: typing.Optional[PipelineUpdateEmbeddingConfig] = OMIT,
|
|
1890
1909
|
transform_config: typing.Optional[PipelineUpdateTransformConfig] = OMIT,
|
|
1910
|
+
sparse_model_config: typing.Optional[SparseModelConfig] = OMIT,
|
|
1891
1911
|
data_sink_id: typing.Optional[str] = OMIT,
|
|
1892
1912
|
embedding_model_config_id: typing.Optional[str] = OMIT,
|
|
1893
1913
|
data_sink: typing.Optional[DataSinkCreate] = OMIT,
|
|
@@ -1909,6 +1929,8 @@ class AsyncPipelinesClient:
|
|
|
1909
1929
|
|
|
1910
1930
|
- transform_config: typing.Optional[PipelineUpdateTransformConfig]. Configuration for the transformation.
|
|
1911
1931
|
|
|
1932
|
+
- sparse_model_config: typing.Optional[SparseModelConfig].
|
|
1933
|
+
|
|
1912
1934
|
- data_sink_id: typing.Optional[str].
|
|
1913
1935
|
|
|
1914
1936
|
- embedding_model_config_id: typing.Optional[str].
|
|
@@ -1934,6 +1956,8 @@ class AsyncPipelinesClient:
|
|
|
1934
1956
|
_request["embedding_config"] = embedding_config
|
|
1935
1957
|
if transform_config is not OMIT:
|
|
1936
1958
|
_request["transform_config"] = transform_config
|
|
1959
|
+
if sparse_model_config is not OMIT:
|
|
1960
|
+
_request["sparse_model_config"] = sparse_model_config
|
|
1937
1961
|
if data_sink_id is not OMIT:
|
|
1938
1962
|
_request["data_sink_id"] = data_sink_id
|
|
1939
1963
|
if embedding_model_config_id is not OMIT:
|
|
@@ -2225,6 +2249,7 @@ class AsyncPipelinesClient:
|
|
|
2225
2249
|
*,
|
|
2226
2250
|
data_source_id: typing.Optional[str] = None,
|
|
2227
2251
|
only_manually_uploaded: typing.Optional[bool] = None,
|
|
2252
|
+
file_name_contains: typing.Optional[str] = None,
|
|
2228
2253
|
limit: typing.Optional[int] = None,
|
|
2229
2254
|
offset: typing.Optional[int] = None,
|
|
2230
2255
|
order_by: typing.Optional[str] = None,
|
|
@@ -2232,6 +2257,15 @@ class AsyncPipelinesClient:
|
|
|
2232
2257
|
"""
|
|
2233
2258
|
Get files for a pipeline.
|
|
2234
2259
|
|
|
2260
|
+
Args:
|
|
2261
|
+
pipeline_id: ID of the pipeline
|
|
2262
|
+
data_source_id: Optional filter by data source ID
|
|
2263
|
+
only_manually_uploaded: Filter for only manually uploaded files
|
|
2264
|
+
file_name_contains: Optional filter by file name (substring match)
|
|
2265
|
+
limit: Limit number of results
|
|
2266
|
+
offset: Offset for pagination
|
|
2267
|
+
order_by: Field to order by
|
|
2268
|
+
|
|
2235
2269
|
Parameters:
|
|
2236
2270
|
- pipeline_id: str.
|
|
2237
2271
|
|
|
@@ -2239,6 +2273,8 @@ class AsyncPipelinesClient:
|
|
|
2239
2273
|
|
|
2240
2274
|
- only_manually_uploaded: typing.Optional[bool].
|
|
2241
2275
|
|
|
2276
|
+
- file_name_contains: typing.Optional[str].
|
|
2277
|
+
|
|
2242
2278
|
- limit: typing.Optional[int].
|
|
2243
2279
|
|
|
2244
2280
|
- offset: typing.Optional[int].
|
|
@@ -2261,6 +2297,7 @@ class AsyncPipelinesClient:
|
|
|
2261
2297
|
{
|
|
2262
2298
|
"data_source_id": data_source_id,
|
|
2263
2299
|
"only_manually_uploaded": only_manually_uploaded,
|
|
2300
|
+
"file_name_contains": file_name_contains,
|
|
2264
2301
|
"limit": limit,
|
|
2265
2302
|
"offset": offset,
|
|
2266
2303
|
"order_by": order_by,
|