llama-cloud 0.1.39__py3-none-any.whl → 0.1.41__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (61) hide show
  1. llama_cloud/__init__.py +24 -72
  2. llama_cloud/client.py +2 -5
  3. llama_cloud/resources/__init__.py +0 -4
  4. llama_cloud/resources/admin/client.py +5 -5
  5. llama_cloud/resources/beta/client.py +1351 -335
  6. llama_cloud/resources/llama_extract/client.py +56 -0
  7. llama_cloud/resources/parsing/client.py +8 -0
  8. llama_cloud/resources/pipelines/client.py +37 -0
  9. llama_cloud/types/__init__.py +24 -72
  10. llama_cloud/types/{src_app_schema_chat_chat_message.py → chat_message.py} +1 -1
  11. llama_cloud/types/extract_config.py +8 -2
  12. llama_cloud/types/extract_job_create.py +2 -0
  13. llama_cloud/types/extract_job_create_priority.py +29 -0
  14. llama_cloud/types/extract_models.py +28 -28
  15. llama_cloud/types/job_names.py +0 -4
  16. llama_cloud/types/{document_block.py → llama_extract_feature_availability.py} +5 -6
  17. llama_cloud/types/llama_extract_mode_availability.py +4 -3
  18. llama_cloud/types/llama_extract_settings.py +1 -1
  19. llama_cloud/types/llama_parse_parameters.py +1 -0
  20. llama_cloud/types/{progress_event.py → parse_configuration.py} +12 -12
  21. llama_cloud/types/{llama_index_core_base_llms_types_chat_message.py → parse_configuration_create.py} +9 -7
  22. llama_cloud/types/{edit_suggestion.py → parse_configuration_filter.py} +8 -6
  23. llama_cloud/types/{report_update_event.py → parse_configuration_query_response.py} +6 -6
  24. llama_cloud/types/parse_job_config.py +1 -0
  25. llama_cloud/types/pipeline.py +4 -0
  26. llama_cloud/types/pipeline_create.py +2 -0
  27. llama_cloud/types/playground_session.py +2 -2
  28. llama_cloud/types/public_model_name.py +97 -0
  29. llama_cloud/types/{report_create_response.py → schema_generation_availability.py} +4 -2
  30. llama_cloud/types/schema_generation_availability_status.py +17 -0
  31. llama_cloud/types/{report_event_item.py → sparse_model_config.py} +10 -8
  32. llama_cloud/types/sparse_model_type.py +33 -0
  33. llama_cloud/types/webhook_configuration.py +1 -0
  34. llama_cloud-0.1.41.dist-info/METADATA +106 -0
  35. {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/RECORD +37 -56
  36. {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/WHEEL +1 -1
  37. llama_cloud/resources/reports/__init__.py +0 -5
  38. llama_cloud/resources/reports/client.py +0 -1230
  39. llama_cloud/resources/reports/types/__init__.py +0 -7
  40. llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +0 -25
  41. llama_cloud/types/audio_block.py +0 -34
  42. llama_cloud/types/edit_suggestion_blocks_item.py +0 -8
  43. llama_cloud/types/image_block.py +0 -35
  44. llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +0 -56
  45. llama_cloud/types/paginated_report_response.py +0 -35
  46. llama_cloud/types/progress_event_status.py +0 -33
  47. llama_cloud/types/report.py +0 -33
  48. llama_cloud/types/report_block.py +0 -35
  49. llama_cloud/types/report_block_dependency.py +0 -29
  50. llama_cloud/types/report_event_item_event_data.py +0 -45
  51. llama_cloud/types/report_event_type.py +0 -37
  52. llama_cloud/types/report_metadata.py +0 -43
  53. llama_cloud/types/report_plan.py +0 -36
  54. llama_cloud/types/report_plan_block.py +0 -36
  55. llama_cloud/types/report_query.py +0 -33
  56. llama_cloud/types/report_response.py +0 -41
  57. llama_cloud/types/report_state.py +0 -37
  58. llama_cloud/types/report_state_event.py +0 -38
  59. llama_cloud/types/text_block.py +0 -31
  60. llama_cloud-0.1.39.dist-info/METADATA +0 -32
  61. {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/LICENSE +0 -0
@@ -111,7 +111,9 @@ class LlamaExtractClient:
111
111
  ExtractConfig,
112
112
  ExtractConfigPriority,
113
113
  ExtractMode,
114
+ ExtractModels,
114
115
  ExtractTarget,
116
+ PublicModelName,
115
117
  )
116
118
  from llama_cloud.client import LlamaCloud
117
119
 
@@ -124,6 +126,8 @@ class LlamaExtractClient:
124
126
  priority=ExtractConfigPriority.LOW,
125
127
  extraction_target=ExtractTarget.PER_DOC,
126
128
  extraction_mode=ExtractMode.FAST,
129
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
130
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
127
131
  chunk_mode=DocumentChunkMode.PAGE,
128
132
  ),
129
133
  )
@@ -359,7 +363,9 @@ class LlamaExtractClient:
359
363
  ExtractConfig,
360
364
  ExtractConfigPriority,
361
365
  ExtractMode,
366
+ ExtractModels,
362
367
  ExtractTarget,
368
+ PublicModelName,
363
369
  )
364
370
  from llama_cloud.client import LlamaCloud
365
371
 
@@ -372,6 +378,8 @@ class LlamaExtractClient:
372
378
  priority=ExtractConfigPriority.LOW,
373
379
  extraction_target=ExtractTarget.PER_DOC,
374
380
  extraction_mode=ExtractMode.FAST,
381
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
382
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
375
383
  chunk_mode=DocumentChunkMode.PAGE,
376
384
  ),
377
385
  )
@@ -470,8 +478,11 @@ class LlamaExtractClient:
470
478
  ExtractConfig,
471
479
  ExtractConfigPriority,
472
480
  ExtractJobCreate,
481
+ ExtractJobCreatePriority,
473
482
  ExtractMode,
483
+ ExtractModels,
474
484
  ExtractTarget,
485
+ PublicModelName,
475
486
  )
476
487
  from llama_cloud.client import LlamaCloud
477
488
 
@@ -480,12 +491,15 @@ class LlamaExtractClient:
480
491
  )
481
492
  client.llama_extract.run_job(
482
493
  request=ExtractJobCreate(
494
+ priority=ExtractJobCreatePriority.LOW,
483
495
  extraction_agent_id="string",
484
496
  file_id="string",
485
497
  config_override=ExtractConfig(
486
498
  priority=ExtractConfigPriority.LOW,
487
499
  extraction_target=ExtractTarget.PER_DOC,
488
500
  extraction_mode=ExtractMode.FAST,
501
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
502
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
489
503
  chunk_mode=DocumentChunkMode.PAGE,
490
504
  ),
491
505
  ),
@@ -560,7 +574,9 @@ class LlamaExtractClient:
560
574
  ExtractConfig,
561
575
  ExtractConfigPriority,
562
576
  ExtractJobCreate,
577
+ ExtractJobCreatePriority,
563
578
  ExtractMode,
579
+ ExtractModels,
564
580
  ExtractTarget,
565
581
  FailPageMode,
566
582
  LlamaExtractSettings,
@@ -568,6 +584,7 @@ class LlamaExtractClient:
568
584
  LlamaParseParametersPriority,
569
585
  MultimodalParseResolution,
570
586
  ParsingMode,
587
+ PublicModelName,
571
588
  )
572
589
  from llama_cloud.client import LlamaCloud
573
590
 
@@ -576,12 +593,15 @@ class LlamaExtractClient:
576
593
  )
577
594
  client.llama_extract.run_job_test_user(
578
595
  job_create=ExtractJobCreate(
596
+ priority=ExtractJobCreatePriority.LOW,
579
597
  extraction_agent_id="string",
580
598
  file_id="string",
581
599
  config_override=ExtractConfig(
582
600
  priority=ExtractConfigPriority.LOW,
583
601
  extraction_target=ExtractTarget.PER_DOC,
584
602
  extraction_mode=ExtractMode.FAST,
603
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
604
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
585
605
  chunk_mode=DocumentChunkMode.PAGE,
586
606
  ),
587
607
  ),
@@ -689,7 +709,9 @@ class LlamaExtractClient:
689
709
  ExtractConfig,
690
710
  ExtractConfigPriority,
691
711
  ExtractMode,
712
+ ExtractModels,
692
713
  ExtractTarget,
714
+ PublicModelName,
693
715
  )
694
716
  from llama_cloud.client import LlamaCloud
695
717
 
@@ -703,6 +725,8 @@ class LlamaExtractClient:
703
725
  priority=ExtractConfigPriority.LOW,
704
726
  extraction_target=ExtractTarget.PER_DOC,
705
727
  extraction_mode=ExtractMode.FAST,
728
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
729
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
706
730
  chunk_mode=DocumentChunkMode.PAGE,
707
731
  ),
708
732
  )
@@ -984,8 +1008,10 @@ class LlamaExtractClient:
984
1008
  ExtractConfig,
985
1009
  ExtractConfigPriority,
986
1010
  ExtractMode,
1011
+ ExtractModels,
987
1012
  ExtractTarget,
988
1013
  FileData,
1014
+ PublicModelName,
989
1015
  )
990
1016
  from llama_cloud.client import LlamaCloud
991
1017
 
@@ -997,6 +1023,8 @@ class LlamaExtractClient:
997
1023
  priority=ExtractConfigPriority.LOW,
998
1024
  extraction_target=ExtractTarget.PER_DOC,
999
1025
  extraction_mode=ExtractMode.FAST,
1026
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
1027
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1000
1028
  chunk_mode=DocumentChunkMode.PAGE,
1001
1029
  ),
1002
1030
  file=FileData(
@@ -1104,7 +1132,9 @@ class AsyncLlamaExtractClient:
1104
1132
  ExtractConfig,
1105
1133
  ExtractConfigPriority,
1106
1134
  ExtractMode,
1135
+ ExtractModels,
1107
1136
  ExtractTarget,
1137
+ PublicModelName,
1108
1138
  )
1109
1139
  from llama_cloud.client import AsyncLlamaCloud
1110
1140
 
@@ -1117,6 +1147,8 @@ class AsyncLlamaExtractClient:
1117
1147
  priority=ExtractConfigPriority.LOW,
1118
1148
  extraction_target=ExtractTarget.PER_DOC,
1119
1149
  extraction_mode=ExtractMode.FAST,
1150
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
1151
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1120
1152
  chunk_mode=DocumentChunkMode.PAGE,
1121
1153
  ),
1122
1154
  )
@@ -1352,7 +1384,9 @@ class AsyncLlamaExtractClient:
1352
1384
  ExtractConfig,
1353
1385
  ExtractConfigPriority,
1354
1386
  ExtractMode,
1387
+ ExtractModels,
1355
1388
  ExtractTarget,
1389
+ PublicModelName,
1356
1390
  )
1357
1391
  from llama_cloud.client import AsyncLlamaCloud
1358
1392
 
@@ -1365,6 +1399,8 @@ class AsyncLlamaExtractClient:
1365
1399
  priority=ExtractConfigPriority.LOW,
1366
1400
  extraction_target=ExtractTarget.PER_DOC,
1367
1401
  extraction_mode=ExtractMode.FAST,
1402
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
1403
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1368
1404
  chunk_mode=DocumentChunkMode.PAGE,
1369
1405
  ),
1370
1406
  )
@@ -1463,8 +1499,11 @@ class AsyncLlamaExtractClient:
1463
1499
  ExtractConfig,
1464
1500
  ExtractConfigPriority,
1465
1501
  ExtractJobCreate,
1502
+ ExtractJobCreatePriority,
1466
1503
  ExtractMode,
1504
+ ExtractModels,
1467
1505
  ExtractTarget,
1506
+ PublicModelName,
1468
1507
  )
1469
1508
  from llama_cloud.client import AsyncLlamaCloud
1470
1509
 
@@ -1473,12 +1512,15 @@ class AsyncLlamaExtractClient:
1473
1512
  )
1474
1513
  await client.llama_extract.run_job(
1475
1514
  request=ExtractJobCreate(
1515
+ priority=ExtractJobCreatePriority.LOW,
1476
1516
  extraction_agent_id="string",
1477
1517
  file_id="string",
1478
1518
  config_override=ExtractConfig(
1479
1519
  priority=ExtractConfigPriority.LOW,
1480
1520
  extraction_target=ExtractTarget.PER_DOC,
1481
1521
  extraction_mode=ExtractMode.FAST,
1522
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
1523
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1482
1524
  chunk_mode=DocumentChunkMode.PAGE,
1483
1525
  ),
1484
1526
  ),
@@ -1553,7 +1595,9 @@ class AsyncLlamaExtractClient:
1553
1595
  ExtractConfig,
1554
1596
  ExtractConfigPriority,
1555
1597
  ExtractJobCreate,
1598
+ ExtractJobCreatePriority,
1556
1599
  ExtractMode,
1600
+ ExtractModels,
1557
1601
  ExtractTarget,
1558
1602
  FailPageMode,
1559
1603
  LlamaExtractSettings,
@@ -1561,6 +1605,7 @@ class AsyncLlamaExtractClient:
1561
1605
  LlamaParseParametersPriority,
1562
1606
  MultimodalParseResolution,
1563
1607
  ParsingMode,
1608
+ PublicModelName,
1564
1609
  )
1565
1610
  from llama_cloud.client import AsyncLlamaCloud
1566
1611
 
@@ -1569,12 +1614,15 @@ class AsyncLlamaExtractClient:
1569
1614
  )
1570
1615
  await client.llama_extract.run_job_test_user(
1571
1616
  job_create=ExtractJobCreate(
1617
+ priority=ExtractJobCreatePriority.LOW,
1572
1618
  extraction_agent_id="string",
1573
1619
  file_id="string",
1574
1620
  config_override=ExtractConfig(
1575
1621
  priority=ExtractConfigPriority.LOW,
1576
1622
  extraction_target=ExtractTarget.PER_DOC,
1577
1623
  extraction_mode=ExtractMode.FAST,
1624
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
1625
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1578
1626
  chunk_mode=DocumentChunkMode.PAGE,
1579
1627
  ),
1580
1628
  ),
@@ -1682,7 +1730,9 @@ class AsyncLlamaExtractClient:
1682
1730
  ExtractConfig,
1683
1731
  ExtractConfigPriority,
1684
1732
  ExtractMode,
1733
+ ExtractModels,
1685
1734
  ExtractTarget,
1735
+ PublicModelName,
1686
1736
  )
1687
1737
  from llama_cloud.client import AsyncLlamaCloud
1688
1738
 
@@ -1696,6 +1746,8 @@ class AsyncLlamaExtractClient:
1696
1746
  priority=ExtractConfigPriority.LOW,
1697
1747
  extraction_target=ExtractTarget.PER_DOC,
1698
1748
  extraction_mode=ExtractMode.FAST,
1749
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
1750
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1699
1751
  chunk_mode=DocumentChunkMode.PAGE,
1700
1752
  ),
1701
1753
  )
@@ -1977,8 +2029,10 @@ class AsyncLlamaExtractClient:
1977
2029
  ExtractConfig,
1978
2030
  ExtractConfigPriority,
1979
2031
  ExtractMode,
2032
+ ExtractModels,
1980
2033
  ExtractTarget,
1981
2034
  FileData,
2035
+ PublicModelName,
1982
2036
  )
1983
2037
  from llama_cloud.client import AsyncLlamaCloud
1984
2038
 
@@ -1990,6 +2044,8 @@ class AsyncLlamaExtractClient:
1990
2044
  priority=ExtractConfigPriority.LOW,
1991
2045
  extraction_target=ExtractTarget.PER_DOC,
1992
2046
  extraction_mode=ExtractMode.FAST,
2047
+ parse_model=PublicModelName.OPENAI_GPT_4_O,
2048
+ extract_model=ExtractModels.OPENAI_GPT_4_1,
1993
2049
  chunk_mode=DocumentChunkMode.PAGE,
1994
2050
  ),
1995
2051
  file=FileData(
@@ -227,6 +227,7 @@ class ParsingClient:
227
227
  guess_xlsx_sheet_name: bool,
228
228
  high_res_ocr: bool,
229
229
  html_make_all_elements_visible: bool,
230
+ layout_aware: bool,
230
231
  html_remove_fixed_elements: bool,
231
232
  html_remove_navigation_elements: bool,
232
233
  http_proxy: str,
@@ -357,6 +358,8 @@ class ParsingClient:
357
358
 
358
359
  - html_make_all_elements_visible: bool.
359
360
 
361
+ - layout_aware: bool.
362
+
360
363
  - html_remove_fixed_elements: bool.
361
364
 
362
365
  - html_remove_navigation_elements: bool.
@@ -522,6 +525,7 @@ class ParsingClient:
522
525
  "guess_xlsx_sheet_name": guess_xlsx_sheet_name,
523
526
  "high_res_ocr": high_res_ocr,
524
527
  "html_make_all_elements_visible": html_make_all_elements_visible,
528
+ "layout_aware": layout_aware,
525
529
  "html_remove_fixed_elements": html_remove_fixed_elements,
526
530
  "html_remove_navigation_elements": html_remove_navigation_elements,
527
531
  "http_proxy": http_proxy,
@@ -1399,6 +1403,7 @@ class AsyncParsingClient:
1399
1403
  guess_xlsx_sheet_name: bool,
1400
1404
  high_res_ocr: bool,
1401
1405
  html_make_all_elements_visible: bool,
1406
+ layout_aware: bool,
1402
1407
  html_remove_fixed_elements: bool,
1403
1408
  html_remove_navigation_elements: bool,
1404
1409
  http_proxy: str,
@@ -1529,6 +1534,8 @@ class AsyncParsingClient:
1529
1534
 
1530
1535
  - html_make_all_elements_visible: bool.
1531
1536
 
1537
+ - layout_aware: bool.
1538
+
1532
1539
  - html_remove_fixed_elements: bool.
1533
1540
 
1534
1541
  - html_remove_navigation_elements: bool.
@@ -1694,6 +1701,7 @@ class AsyncParsingClient:
1694
1701
  "guess_xlsx_sheet_name": guess_xlsx_sheet_name,
1695
1702
  "high_res_ocr": high_res_ocr,
1696
1703
  "html_make_all_elements_visible": html_make_all_elements_visible,
1704
+ "layout_aware": layout_aware,
1697
1705
  "html_remove_fixed_elements": html_remove_fixed_elements,
1698
1706
  "html_remove_navigation_elements": html_remove_navigation_elements,
1699
1707
  "http_proxy": http_proxy,
@@ -35,6 +35,7 @@ from ...types.playground_session import PlaygroundSession
35
35
  from ...types.preset_retrieval_params import PresetRetrievalParams
36
36
  from ...types.retrieval_mode import RetrievalMode
37
37
  from ...types.retrieve_results import RetrieveResults
38
+ from ...types.sparse_model_config import SparseModelConfig
38
39
  from ...types.text_node import TextNode
39
40
  from .types.pipeline_file_update_custom_metadata_value import PipelineFileUpdateCustomMetadataValue
40
41
  from .types.pipeline_update_embedding_config import PipelineUpdateEmbeddingConfig
@@ -217,6 +218,7 @@ class PipelinesClient:
217
218
  *,
218
219
  embedding_config: typing.Optional[PipelineUpdateEmbeddingConfig] = OMIT,
219
220
  transform_config: typing.Optional[PipelineUpdateTransformConfig] = OMIT,
221
+ sparse_model_config: typing.Optional[SparseModelConfig] = OMIT,
220
222
  data_sink_id: typing.Optional[str] = OMIT,
221
223
  embedding_model_config_id: typing.Optional[str] = OMIT,
222
224
  data_sink: typing.Optional[DataSinkCreate] = OMIT,
@@ -238,6 +240,8 @@ class PipelinesClient:
238
240
 
239
241
  - transform_config: typing.Optional[PipelineUpdateTransformConfig]. Configuration for the transformation.
240
242
 
243
+ - sparse_model_config: typing.Optional[SparseModelConfig].
244
+
241
245
  - data_sink_id: typing.Optional[str].
242
246
 
243
247
  - embedding_model_config_id: typing.Optional[str].
@@ -263,6 +267,8 @@ class PipelinesClient:
263
267
  _request["embedding_config"] = embedding_config
264
268
  if transform_config is not OMIT:
265
269
  _request["transform_config"] = transform_config
270
+ if sparse_model_config is not OMIT:
271
+ _request["sparse_model_config"] = sparse_model_config
266
272
  if data_sink_id is not OMIT:
267
273
  _request["data_sink_id"] = data_sink_id
268
274
  if embedding_model_config_id is not OMIT:
@@ -554,6 +560,7 @@ class PipelinesClient:
554
560
  *,
555
561
  data_source_id: typing.Optional[str] = None,
556
562
  only_manually_uploaded: typing.Optional[bool] = None,
563
+ file_name_contains: typing.Optional[str] = None,
557
564
  limit: typing.Optional[int] = None,
558
565
  offset: typing.Optional[int] = None,
559
566
  order_by: typing.Optional[str] = None,
@@ -561,6 +568,15 @@ class PipelinesClient:
561
568
  """
562
569
  Get files for a pipeline.
563
570
 
571
+ Args:
572
+ pipeline_id: ID of the pipeline
573
+ data_source_id: Optional filter by data source ID
574
+ only_manually_uploaded: Filter for only manually uploaded files
575
+ file_name_contains: Optional filter by file name (substring match)
576
+ limit: Limit number of results
577
+ offset: Offset for pagination
578
+ order_by: Field to order by
579
+
564
580
  Parameters:
565
581
  - pipeline_id: str.
566
582
 
@@ -568,6 +584,8 @@ class PipelinesClient:
568
584
 
569
585
  - only_manually_uploaded: typing.Optional[bool].
570
586
 
587
+ - file_name_contains: typing.Optional[str].
588
+
571
589
  - limit: typing.Optional[int].
572
590
 
573
591
  - offset: typing.Optional[int].
@@ -590,6 +608,7 @@ class PipelinesClient:
590
608
  {
591
609
  "data_source_id": data_source_id,
592
610
  "only_manually_uploaded": only_manually_uploaded,
611
+ "file_name_contains": file_name_contains,
593
612
  "limit": limit,
594
613
  "offset": offset,
595
614
  "order_by": order_by,
@@ -1888,6 +1907,7 @@ class AsyncPipelinesClient:
1888
1907
  *,
1889
1908
  embedding_config: typing.Optional[PipelineUpdateEmbeddingConfig] = OMIT,
1890
1909
  transform_config: typing.Optional[PipelineUpdateTransformConfig] = OMIT,
1910
+ sparse_model_config: typing.Optional[SparseModelConfig] = OMIT,
1891
1911
  data_sink_id: typing.Optional[str] = OMIT,
1892
1912
  embedding_model_config_id: typing.Optional[str] = OMIT,
1893
1913
  data_sink: typing.Optional[DataSinkCreate] = OMIT,
@@ -1909,6 +1929,8 @@ class AsyncPipelinesClient:
1909
1929
 
1910
1930
  - transform_config: typing.Optional[PipelineUpdateTransformConfig]. Configuration for the transformation.
1911
1931
 
1932
+ - sparse_model_config: typing.Optional[SparseModelConfig].
1933
+
1912
1934
  - data_sink_id: typing.Optional[str].
1913
1935
 
1914
1936
  - embedding_model_config_id: typing.Optional[str].
@@ -1934,6 +1956,8 @@ class AsyncPipelinesClient:
1934
1956
  _request["embedding_config"] = embedding_config
1935
1957
  if transform_config is not OMIT:
1936
1958
  _request["transform_config"] = transform_config
1959
+ if sparse_model_config is not OMIT:
1960
+ _request["sparse_model_config"] = sparse_model_config
1937
1961
  if data_sink_id is not OMIT:
1938
1962
  _request["data_sink_id"] = data_sink_id
1939
1963
  if embedding_model_config_id is not OMIT:
@@ -2225,6 +2249,7 @@ class AsyncPipelinesClient:
2225
2249
  *,
2226
2250
  data_source_id: typing.Optional[str] = None,
2227
2251
  only_manually_uploaded: typing.Optional[bool] = None,
2252
+ file_name_contains: typing.Optional[str] = None,
2228
2253
  limit: typing.Optional[int] = None,
2229
2254
  offset: typing.Optional[int] = None,
2230
2255
  order_by: typing.Optional[str] = None,
@@ -2232,6 +2257,15 @@ class AsyncPipelinesClient:
2232
2257
  """
2233
2258
  Get files for a pipeline.
2234
2259
 
2260
+ Args:
2261
+ pipeline_id: ID of the pipeline
2262
+ data_source_id: Optional filter by data source ID
2263
+ only_manually_uploaded: Filter for only manually uploaded files
2264
+ file_name_contains: Optional filter by file name (substring match)
2265
+ limit: Limit number of results
2266
+ offset: Offset for pagination
2267
+ order_by: Field to order by
2268
+
2235
2269
  Parameters:
2236
2270
  - pipeline_id: str.
2237
2271
 
@@ -2239,6 +2273,8 @@ class AsyncPipelinesClient:
2239
2273
 
2240
2274
  - only_manually_uploaded: typing.Optional[bool].
2241
2275
 
2276
+ - file_name_contains: typing.Optional[str].
2277
+
2242
2278
  - limit: typing.Optional[int].
2243
2279
 
2244
2280
  - offset: typing.Optional[int].
@@ -2261,6 +2297,7 @@ class AsyncPipelinesClient:
2261
2297
  {
2262
2298
  "data_source_id": data_source_id,
2263
2299
  "only_manually_uploaded": only_manually_uploaded,
2300
+ "file_name_contains": file_name_contains,
2264
2301
  "limit": limit,
2265
2302
  "offset": offset,
2266
2303
  "order_by": order_by,