llama-cloud 0.1.20__py3-none-any.whl → 0.1.22__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (55) hide show
  1. llama_cloud/__init__.py +18 -136
  2. llama_cloud/client.py +3 -3
  3. llama_cloud/resources/__init__.py +4 -43
  4. llama_cloud/resources/admin/client.py +78 -0
  5. llama_cloud/resources/data_sinks/__init__.py +2 -18
  6. llama_cloud/resources/data_sinks/client.py +94 -2
  7. llama_cloud/resources/data_sinks/types/__init__.py +2 -18
  8. llama_cloud/resources/data_sinks/types/data_sink_update_component.py +7 -65
  9. llama_cloud/resources/data_sources/__init__.py +2 -30
  10. llama_cloud/resources/data_sources/types/__init__.py +1 -28
  11. llama_cloud/resources/data_sources/types/data_source_update_component.py +23 -2
  12. llama_cloud/resources/jobs/client.py +10 -2
  13. llama_cloud/resources/llama_extract/client.py +50 -6
  14. llama_cloud/resources/organizations/client.py +12 -2
  15. llama_cloud/resources/parsing/client.py +30 -0
  16. llama_cloud/resources/pipelines/client.py +22 -0
  17. llama_cloud/resources/retrievers/client.py +14 -0
  18. llama_cloud/types/__init__.py +18 -112
  19. llama_cloud/types/cloud_jira_data_source.py +4 -0
  20. llama_cloud/types/cloud_s_3_data_source.py +1 -0
  21. llama_cloud/types/data_sink_component.py +7 -65
  22. llama_cloud/types/data_sink_create_component.py +7 -65
  23. llama_cloud/types/data_source_component.py +23 -2
  24. llama_cloud/types/data_source_create_component.py +23 -2
  25. llama_cloud/types/{data_sink_definition.py → document_block.py} +6 -15
  26. llama_cloud/types/document_chunk_mode.py +17 -0
  27. llama_cloud/types/extract_config.py +7 -0
  28. llama_cloud/types/extract_mode.py +4 -0
  29. llama_cloud/types/extract_models.py +33 -0
  30. llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +11 -0
  31. llama_cloud/types/{data_source_definition.py → llm_config_result.py} +6 -15
  32. llama_cloud/types/llm_config_result_llm_type.py +33 -0
  33. llama_cloud/types/llm_configs_response.py +33 -0
  34. llama_cloud/types/pipeline.py +2 -0
  35. llama_cloud/types/pipeline_create.py +1 -3
  36. llama_cloud/types/pipeline_data_source_component.py +23 -2
  37. llama_cloud/types/pipeline_status.py +17 -0
  38. llama_cloud/types/prompt_conf.py +1 -0
  39. llama_cloud/types/struct_parse_conf.py +2 -1
  40. llama_cloud/types/supported_llm_model.py +1 -2
  41. llama_cloud/types/supported_llm_model_names.py +4 -4
  42. llama_cloud/types/user_organization_role.py +1 -0
  43. {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/METADATA +5 -3
  44. {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/RECORD +47 -49
  45. {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/WHEEL +1 -1
  46. llama_cloud/resources/component_definitions/client.py +0 -189
  47. llama_cloud/resources/data_sources/types/data_source_update_component_one.py +0 -122
  48. llama_cloud/types/configurable_transformation_definition.py +0 -48
  49. llama_cloud/types/configurable_transformation_names.py +0 -41
  50. llama_cloud/types/data_source_component_one.py +0 -122
  51. llama_cloud/types/data_source_create_component_one.py +0 -122
  52. llama_cloud/types/pipeline_data_source_component_one.py +0 -122
  53. llama_cloud/types/transformation_category_names.py +0 -17
  54. /llama_cloud/resources/{component_definitions → admin}/__init__.py +0 -0
  55. {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/LICENSE +0 -0
@@ -295,6 +295,8 @@ class RetrieversClient:
295
295
  self,
296
296
  retriever_id: str,
297
297
  *,
298
+ project_id: typing.Optional[str] = None,
299
+ organization_id: typing.Optional[str] = None,
298
300
  mode: typing.Optional[CompositeRetrievalMode] = OMIT,
299
301
  rerank_top_n: typing.Optional[int] = OMIT,
300
302
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
@@ -306,6 +308,10 @@ class RetrieversClient:
306
308
  Parameters:
307
309
  - retriever_id: str.
308
310
 
311
+ - project_id: typing.Optional[str].
312
+
313
+ - organization_id: typing.Optional[str].
314
+
309
315
  - mode: typing.Optional[CompositeRetrievalMode]. The mode of composite retrieval.
310
316
 
311
317
  - rerank_top_n: typing.Optional[int].
@@ -341,6 +347,7 @@ class RetrieversClient:
341
347
  urllib.parse.urljoin(
342
348
  f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}/retrieve"
343
349
  ),
350
+ params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
344
351
  json=jsonable_encoder(_request),
345
352
  headers=self._client_wrapper.get_headers(),
346
353
  timeout=60,
@@ -692,6 +699,8 @@ class AsyncRetrieversClient:
692
699
  self,
693
700
  retriever_id: str,
694
701
  *,
702
+ project_id: typing.Optional[str] = None,
703
+ organization_id: typing.Optional[str] = None,
695
704
  mode: typing.Optional[CompositeRetrievalMode] = OMIT,
696
705
  rerank_top_n: typing.Optional[int] = OMIT,
697
706
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
@@ -703,6 +712,10 @@ class AsyncRetrieversClient:
703
712
  Parameters:
704
713
  - retriever_id: str.
705
714
 
715
+ - project_id: typing.Optional[str].
716
+
717
+ - organization_id: typing.Optional[str].
718
+
706
719
  - mode: typing.Optional[CompositeRetrievalMode]. The mode of composite retrieval.
707
720
 
708
721
  - rerank_top_n: typing.Optional[int].
@@ -738,6 +751,7 @@ class AsyncRetrieversClient:
738
751
  urllib.parse.urljoin(
739
752
  f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}/retrieve"
740
753
  ),
754
+ params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
741
755
  json=jsonable_encoder(_request),
742
756
  headers=self._client_wrapper.get_headers(),
743
757
  timeout=60,
@@ -63,65 +63,21 @@ from .composite_retrieved_text_node import CompositeRetrievedTextNode
63
63
  from .composite_retrieved_text_node_with_score import CompositeRetrievedTextNodeWithScore
64
64
  from .configurable_data_sink_names import ConfigurableDataSinkNames
65
65
  from .configurable_data_source_names import ConfigurableDataSourceNames
66
- from .configurable_transformation_definition import ConfigurableTransformationDefinition
67
- from .configurable_transformation_names import ConfigurableTransformationNames
68
66
  from .credit_type import CreditType
69
67
  from .data_sink import DataSink
70
- from .data_sink_component import (
71
- DataSinkComponent,
72
- DataSinkComponent_AzureAiSearch,
73
- DataSinkComponent_Milvus,
74
- DataSinkComponent_MongodbAtlas,
75
- DataSinkComponent_Pinecone,
76
- DataSinkComponent_Postgres,
77
- DataSinkComponent_Qdrant,
78
- )
68
+ from .data_sink_component import DataSinkComponent
79
69
  from .data_sink_create import DataSinkCreate
80
- from .data_sink_create_component import (
81
- DataSinkCreateComponent,
82
- DataSinkCreateComponent_AzureAiSearch,
83
- DataSinkCreateComponent_Milvus,
84
- DataSinkCreateComponent_MongodbAtlas,
85
- DataSinkCreateComponent_Pinecone,
86
- DataSinkCreateComponent_Postgres,
87
- DataSinkCreateComponent_Qdrant,
88
- )
89
- from .data_sink_definition import DataSinkDefinition
70
+ from .data_sink_create_component import DataSinkCreateComponent
90
71
  from .data_source import DataSource
91
72
  from .data_source_component import DataSourceComponent
92
- from .data_source_component_one import (
93
- DataSourceComponentOne,
94
- DataSourceComponentOne_AzureStorageBlob,
95
- DataSourceComponentOne_Box,
96
- DataSourceComponentOne_Confluence,
97
- DataSourceComponentOne_GoogleDrive,
98
- DataSourceComponentOne_Jira,
99
- DataSourceComponentOne_MicrosoftOnedrive,
100
- DataSourceComponentOne_MicrosoftSharepoint,
101
- DataSourceComponentOne_NotionPage,
102
- DataSourceComponentOne_S3,
103
- DataSourceComponentOne_Slack,
104
- )
105
73
  from .data_source_create import DataSourceCreate
106
74
  from .data_source_create_component import DataSourceCreateComponent
107
- from .data_source_create_component_one import (
108
- DataSourceCreateComponentOne,
109
- DataSourceCreateComponentOne_AzureStorageBlob,
110
- DataSourceCreateComponentOne_Box,
111
- DataSourceCreateComponentOne_Confluence,
112
- DataSourceCreateComponentOne_GoogleDrive,
113
- DataSourceCreateComponentOne_Jira,
114
- DataSourceCreateComponentOne_MicrosoftOnedrive,
115
- DataSourceCreateComponentOne_MicrosoftSharepoint,
116
- DataSourceCreateComponentOne_NotionPage,
117
- DataSourceCreateComponentOne_S3,
118
- DataSourceCreateComponentOne_Slack,
119
- )
120
75
  from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
121
76
  from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
122
- from .data_source_definition import DataSourceDefinition
123
77
  from .data_source_update_dispatcher_config import DataSourceUpdateDispatcherConfig
124
78
  from .delete_params import DeleteParams
79
+ from .document_block import DocumentBlock
80
+ from .document_chunk_mode import DocumentChunkMode
125
81
  from .document_ingestion_job_params import DocumentIngestionJobParams
126
82
  from .edit_suggestion import EditSuggestion
127
83
  from .edit_suggestion_blocks_item import EditSuggestionBlocksItem
@@ -157,6 +113,7 @@ from .extract_job_create import ExtractJobCreate
157
113
  from .extract_job_create_data_schema_override import ExtractJobCreateDataSchemaOverride
158
114
  from .extract_job_create_data_schema_override_zero_value import ExtractJobCreateDataSchemaOverrideZeroValue
159
115
  from .extract_mode import ExtractMode
116
+ from .extract_models import ExtractModels
160
117
  from .extract_resultset import ExtractResultset
161
118
  from .extract_resultset_data import ExtractResultsetData
162
119
  from .extract_resultset_data_item_value import ExtractResultsetDataItemValue
@@ -213,11 +170,15 @@ from .llama_index_core_base_llms_types_chat_message import LlamaIndexCoreBaseLlm
213
170
  from .llama_index_core_base_llms_types_chat_message_blocks_item import (
214
171
  LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem,
215
172
  LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Audio,
173
+ LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Document,
216
174
  LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image,
217
175
  LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text,
218
176
  )
219
177
  from .llama_parse_parameters import LlamaParseParameters
220
178
  from .llama_parse_supported_file_extensions import LlamaParseSupportedFileExtensions
179
+ from .llm_config_result import LlmConfigResult
180
+ from .llm_config_result_llm_type import LlmConfigResultLlmType
181
+ from .llm_configs_response import LlmConfigsResponse
221
182
  from .llm_model_data import LlmModelData
222
183
  from .llm_parameters import LlmParameters
223
184
  from .load_files_job_config import LoadFilesJobConfig
@@ -277,19 +238,6 @@ from .pipeline_create_embedding_config import (
277
238
  from .pipeline_create_transform_config import PipelineCreateTransformConfig
278
239
  from .pipeline_data_source import PipelineDataSource
279
240
  from .pipeline_data_source_component import PipelineDataSourceComponent
280
- from .pipeline_data_source_component_one import (
281
- PipelineDataSourceComponentOne,
282
- PipelineDataSourceComponentOne_AzureStorageBlob,
283
- PipelineDataSourceComponentOne_Box,
284
- PipelineDataSourceComponentOne_Confluence,
285
- PipelineDataSourceComponentOne_GoogleDrive,
286
- PipelineDataSourceComponentOne_Jira,
287
- PipelineDataSourceComponentOne_MicrosoftOnedrive,
288
- PipelineDataSourceComponentOne_MicrosoftSharepoint,
289
- PipelineDataSourceComponentOne_NotionPage,
290
- PipelineDataSourceComponentOne_S3,
291
- PipelineDataSourceComponentOne_Slack,
292
- )
293
241
  from .pipeline_data_source_create import PipelineDataSourceCreate
294
242
  from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
295
243
  from .pipeline_data_source_status import PipelineDataSourceStatus
@@ -316,6 +264,7 @@ from .pipeline_file_update_dispatcher_config import PipelineFileUpdateDispatcher
316
264
  from .pipeline_file_updater_config import PipelineFileUpdaterConfig
317
265
  from .pipeline_managed_ingestion_job_params import PipelineManagedIngestionJobParams
318
266
  from .pipeline_metadata_config import PipelineMetadataConfig
267
+ from .pipeline_status import PipelineStatus
319
268
  from .pipeline_transform_config import (
320
269
  PipelineTransformConfig,
321
270
  PipelineTransformConfig_Advanced,
@@ -377,7 +326,6 @@ from .text_node import TextNode
377
326
  from .text_node_relationships_value import TextNodeRelationshipsValue
378
327
  from .text_node_with_score import TextNodeWithScore
379
328
  from .token_chunking_config import TokenChunkingConfig
380
- from .transformation_category_names import TransformationCategoryNames
381
329
  from .usage_and_plan import UsageAndPlan
382
330
  from .usage_metric_response import UsageMetricResponse
383
331
  from .usage_response import UsageResponse
@@ -453,57 +401,21 @@ __all__ = [
453
401
  "CompositeRetrievedTextNodeWithScore",
454
402
  "ConfigurableDataSinkNames",
455
403
  "ConfigurableDataSourceNames",
456
- "ConfigurableTransformationDefinition",
457
- "ConfigurableTransformationNames",
458
404
  "CreditType",
459
405
  "DataSink",
460
406
  "DataSinkComponent",
461
- "DataSinkComponent_AzureAiSearch",
462
- "DataSinkComponent_Milvus",
463
- "DataSinkComponent_MongodbAtlas",
464
- "DataSinkComponent_Pinecone",
465
- "DataSinkComponent_Postgres",
466
- "DataSinkComponent_Qdrant",
467
407
  "DataSinkCreate",
468
408
  "DataSinkCreateComponent",
469
- "DataSinkCreateComponent_AzureAiSearch",
470
- "DataSinkCreateComponent_Milvus",
471
- "DataSinkCreateComponent_MongodbAtlas",
472
- "DataSinkCreateComponent_Pinecone",
473
- "DataSinkCreateComponent_Postgres",
474
- "DataSinkCreateComponent_Qdrant",
475
- "DataSinkDefinition",
476
409
  "DataSource",
477
410
  "DataSourceComponent",
478
- "DataSourceComponentOne",
479
- "DataSourceComponentOne_AzureStorageBlob",
480
- "DataSourceComponentOne_Box",
481
- "DataSourceComponentOne_Confluence",
482
- "DataSourceComponentOne_GoogleDrive",
483
- "DataSourceComponentOne_Jira",
484
- "DataSourceComponentOne_MicrosoftOnedrive",
485
- "DataSourceComponentOne_MicrosoftSharepoint",
486
- "DataSourceComponentOne_NotionPage",
487
- "DataSourceComponentOne_S3",
488
- "DataSourceComponentOne_Slack",
489
411
  "DataSourceCreate",
490
412
  "DataSourceCreateComponent",
491
- "DataSourceCreateComponentOne",
492
- "DataSourceCreateComponentOne_AzureStorageBlob",
493
- "DataSourceCreateComponentOne_Box",
494
- "DataSourceCreateComponentOne_Confluence",
495
- "DataSourceCreateComponentOne_GoogleDrive",
496
- "DataSourceCreateComponentOne_Jira",
497
- "DataSourceCreateComponentOne_MicrosoftOnedrive",
498
- "DataSourceCreateComponentOne_MicrosoftSharepoint",
499
- "DataSourceCreateComponentOne_NotionPage",
500
- "DataSourceCreateComponentOne_S3",
501
- "DataSourceCreateComponentOne_Slack",
502
413
  "DataSourceCreateCustomMetadataValue",
503
414
  "DataSourceCustomMetadataValue",
504
- "DataSourceDefinition",
505
415
  "DataSourceUpdateDispatcherConfig",
506
416
  "DeleteParams",
417
+ "DocumentBlock",
418
+ "DocumentChunkMode",
507
419
  "DocumentIngestionJobParams",
508
420
  "EditSuggestion",
509
421
  "EditSuggestionBlocksItem",
@@ -535,6 +447,7 @@ __all__ = [
535
447
  "ExtractJobCreateDataSchemaOverride",
536
448
  "ExtractJobCreateDataSchemaOverrideZeroValue",
537
449
  "ExtractMode",
450
+ "ExtractModels",
538
451
  "ExtractResultset",
539
452
  "ExtractResultsetData",
540
453
  "ExtractResultsetDataItemValue",
@@ -588,10 +501,14 @@ __all__ = [
588
501
  "LlamaIndexCoreBaseLlmsTypesChatMessage",
589
502
  "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem",
590
503
  "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Audio",
504
+ "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Document",
591
505
  "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image",
592
506
  "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text",
593
507
  "LlamaParseParameters",
594
508
  "LlamaParseSupportedFileExtensions",
509
+ "LlmConfigResult",
510
+ "LlmConfigResultLlmType",
511
+ "LlmConfigsResponse",
595
512
  "LlmModelData",
596
513
  "LlmParameters",
597
514
  "LoadFilesJobConfig",
@@ -649,17 +566,6 @@ __all__ = [
649
566
  "PipelineCreateTransformConfig",
650
567
  "PipelineDataSource",
651
568
  "PipelineDataSourceComponent",
652
- "PipelineDataSourceComponentOne",
653
- "PipelineDataSourceComponentOne_AzureStorageBlob",
654
- "PipelineDataSourceComponentOne_Box",
655
- "PipelineDataSourceComponentOne_Confluence",
656
- "PipelineDataSourceComponentOne_GoogleDrive",
657
- "PipelineDataSourceComponentOne_Jira",
658
- "PipelineDataSourceComponentOne_MicrosoftOnedrive",
659
- "PipelineDataSourceComponentOne_MicrosoftSharepoint",
660
- "PipelineDataSourceComponentOne_NotionPage",
661
- "PipelineDataSourceComponentOne_S3",
662
- "PipelineDataSourceComponentOne_Slack",
663
569
  "PipelineDataSourceCreate",
664
570
  "PipelineDataSourceCustomMetadataValue",
665
571
  "PipelineDataSourceStatus",
@@ -684,6 +590,7 @@ __all__ = [
684
590
  "PipelineFileUpdaterConfig",
685
591
  "PipelineManagedIngestionJobParams",
686
592
  "PipelineMetadataConfig",
593
+ "PipelineStatus",
687
594
  "PipelineTransformConfig",
688
595
  "PipelineTransformConfig_Advanced",
689
596
  "PipelineTransformConfig_Auto",
@@ -741,7 +648,6 @@ __all__ = [
741
648
  "TextNodeRelationshipsValue",
742
649
  "TextNodeWithScore",
743
650
  "TokenChunkingConfig",
744
- "TransformationCategoryNames",
745
651
  "UsageAndPlan",
746
652
  "UsageMetricResponse",
747
653
  "UsageResponse",
@@ -15,6 +15,10 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudJiraDataSource(pydantic.BaseModel):
18
+ """
19
+ Cloud Jira Data Source integrating JiraReader.
20
+ """
21
+
18
22
  supports_access_control: typing.Optional[bool]
19
23
  email: typing.Optional[str]
20
24
  api_token: typing.Optional[str]
@@ -18,6 +18,7 @@ class CloudS3DataSource(pydantic.BaseModel):
18
18
  supports_access_control: typing.Optional[bool]
19
19
  bucket: str = pydantic.Field(description="The name of the S3 bucket to read from.")
20
20
  prefix: typing.Optional[str]
21
+ regex_pattern: typing.Optional[str]
21
22
  aws_access_id: typing.Optional[str]
22
23
  aws_access_secret: typing.Optional[str]
23
24
  s_3_endpoint_url: typing.Optional[str] = pydantic.Field(alias="s3_endpoint_url")
@@ -1,11 +1,7 @@
1
1
  # This file was auto-generated by Fern from our API Definition.
2
2
 
3
- from __future__ import annotations
4
-
5
3
  import typing
6
4
 
7
- import typing_extensions
8
-
9
5
  from .cloud_azure_ai_search_vector_store import CloudAzureAiSearchVectorStore
10
6
  from .cloud_milvus_vector_store import CloudMilvusVectorStore
11
7
  from .cloud_mongo_db_atlas_vector_search import CloudMongoDbAtlasVectorSearch
@@ -13,66 +9,12 @@ from .cloud_pinecone_vector_store import CloudPineconeVectorStore
13
9
  from .cloud_postgres_vector_store import CloudPostgresVectorStore
14
10
  from .cloud_qdrant_vector_store import CloudQdrantVectorStore
15
11
 
16
-
17
- class DataSinkComponent_AzureAiSearch(CloudAzureAiSearchVectorStore):
18
- type: typing_extensions.Literal["azure_ai_search"]
19
-
20
- class Config:
21
- frozen = True
22
- smart_union = True
23
- allow_population_by_field_name = True
24
-
25
-
26
- class DataSinkComponent_Milvus(CloudMilvusVectorStore):
27
- type: typing_extensions.Literal["milvus"]
28
-
29
- class Config:
30
- frozen = True
31
- smart_union = True
32
- allow_population_by_field_name = True
33
-
34
-
35
- class DataSinkComponent_MongodbAtlas(CloudMongoDbAtlasVectorSearch):
36
- type: typing_extensions.Literal["mongodb_atlas"]
37
-
38
- class Config:
39
- frozen = True
40
- smart_union = True
41
- allow_population_by_field_name = True
42
-
43
-
44
- class DataSinkComponent_Pinecone(CloudPineconeVectorStore):
45
- type: typing_extensions.Literal["pinecone"]
46
-
47
- class Config:
48
- frozen = True
49
- smart_union = True
50
- allow_population_by_field_name = True
51
-
52
-
53
- class DataSinkComponent_Postgres(CloudPostgresVectorStore):
54
- type: typing_extensions.Literal["postgres"]
55
-
56
- class Config:
57
- frozen = True
58
- smart_union = True
59
- allow_population_by_field_name = True
60
-
61
-
62
- class DataSinkComponent_Qdrant(CloudQdrantVectorStore):
63
- type: typing_extensions.Literal["qdrant"]
64
-
65
- class Config:
66
- frozen = True
67
- smart_union = True
68
- allow_population_by_field_name = True
69
-
70
-
71
12
  DataSinkComponent = typing.Union[
72
- DataSinkComponent_AzureAiSearch,
73
- DataSinkComponent_Milvus,
74
- DataSinkComponent_MongodbAtlas,
75
- DataSinkComponent_Pinecone,
76
- DataSinkComponent_Postgres,
77
- DataSinkComponent_Qdrant,
13
+ typing.Dict[str, typing.Any],
14
+ CloudPineconeVectorStore,
15
+ CloudPostgresVectorStore,
16
+ CloudQdrantVectorStore,
17
+ CloudAzureAiSearchVectorStore,
18
+ CloudMongoDbAtlasVectorSearch,
19
+ CloudMilvusVectorStore,
78
20
  ]
@@ -1,11 +1,7 @@
1
1
  # This file was auto-generated by Fern from our API Definition.
2
2
 
3
- from __future__ import annotations
4
-
5
3
  import typing
6
4
 
7
- import typing_extensions
8
-
9
5
  from .cloud_azure_ai_search_vector_store import CloudAzureAiSearchVectorStore
10
6
  from .cloud_milvus_vector_store import CloudMilvusVectorStore
11
7
  from .cloud_mongo_db_atlas_vector_search import CloudMongoDbAtlasVectorSearch
@@ -13,66 +9,12 @@ from .cloud_pinecone_vector_store import CloudPineconeVectorStore
13
9
  from .cloud_postgres_vector_store import CloudPostgresVectorStore
14
10
  from .cloud_qdrant_vector_store import CloudQdrantVectorStore
15
11
 
16
-
17
- class DataSinkCreateComponent_AzureAiSearch(CloudAzureAiSearchVectorStore):
18
- type: typing_extensions.Literal["azure_ai_search"]
19
-
20
- class Config:
21
- frozen = True
22
- smart_union = True
23
- allow_population_by_field_name = True
24
-
25
-
26
- class DataSinkCreateComponent_Milvus(CloudMilvusVectorStore):
27
- type: typing_extensions.Literal["milvus"]
28
-
29
- class Config:
30
- frozen = True
31
- smart_union = True
32
- allow_population_by_field_name = True
33
-
34
-
35
- class DataSinkCreateComponent_MongodbAtlas(CloudMongoDbAtlasVectorSearch):
36
- type: typing_extensions.Literal["mongodb_atlas"]
37
-
38
- class Config:
39
- frozen = True
40
- smart_union = True
41
- allow_population_by_field_name = True
42
-
43
-
44
- class DataSinkCreateComponent_Pinecone(CloudPineconeVectorStore):
45
- type: typing_extensions.Literal["pinecone"]
46
-
47
- class Config:
48
- frozen = True
49
- smart_union = True
50
- allow_population_by_field_name = True
51
-
52
-
53
- class DataSinkCreateComponent_Postgres(CloudPostgresVectorStore):
54
- type: typing_extensions.Literal["postgres"]
55
-
56
- class Config:
57
- frozen = True
58
- smart_union = True
59
- allow_population_by_field_name = True
60
-
61
-
62
- class DataSinkCreateComponent_Qdrant(CloudQdrantVectorStore):
63
- type: typing_extensions.Literal["qdrant"]
64
-
65
- class Config:
66
- frozen = True
67
- smart_union = True
68
- allow_population_by_field_name = True
69
-
70
-
71
12
  DataSinkCreateComponent = typing.Union[
72
- DataSinkCreateComponent_AzureAiSearch,
73
- DataSinkCreateComponent_Milvus,
74
- DataSinkCreateComponent_MongodbAtlas,
75
- DataSinkCreateComponent_Pinecone,
76
- DataSinkCreateComponent_Postgres,
77
- DataSinkCreateComponent_Qdrant,
13
+ typing.Dict[str, typing.Any],
14
+ CloudPineconeVectorStore,
15
+ CloudPostgresVectorStore,
16
+ CloudQdrantVectorStore,
17
+ CloudAzureAiSearchVectorStore,
18
+ CloudMongoDbAtlasVectorSearch,
19
+ CloudMilvusVectorStore,
78
20
  ]
@@ -2,6 +2,27 @@
2
2
 
3
3
  import typing
4
4
 
5
- from .data_source_component_one import DataSourceComponentOne
5
+ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
6
+ from .cloud_box_data_source import CloudBoxDataSource
7
+ from .cloud_confluence_data_source import CloudConfluenceDataSource
8
+ from .cloud_google_drive_data_source import CloudGoogleDriveDataSource
9
+ from .cloud_jira_data_source import CloudJiraDataSource
10
+ from .cloud_notion_page_data_source import CloudNotionPageDataSource
11
+ from .cloud_one_drive_data_source import CloudOneDriveDataSource
12
+ from .cloud_s_3_data_source import CloudS3DataSource
13
+ from .cloud_sharepoint_data_source import CloudSharepointDataSource
14
+ from .cloud_slack_data_source import CloudSlackDataSource
6
15
 
7
- DataSourceComponent = typing.Union[typing.Dict[str, typing.Any], DataSourceComponentOne]
16
+ DataSourceComponent = typing.Union[
17
+ typing.Dict[str, typing.Any],
18
+ CloudS3DataSource,
19
+ CloudAzStorageBlobDataSource,
20
+ CloudGoogleDriveDataSource,
21
+ CloudOneDriveDataSource,
22
+ CloudSharepointDataSource,
23
+ CloudSlackDataSource,
24
+ CloudNotionPageDataSource,
25
+ CloudConfluenceDataSource,
26
+ CloudJiraDataSource,
27
+ CloudBoxDataSource,
28
+ ]
@@ -2,6 +2,27 @@
2
2
 
3
3
  import typing
4
4
 
5
- from .data_source_create_component_one import DataSourceCreateComponentOne
5
+ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
6
+ from .cloud_box_data_source import CloudBoxDataSource
7
+ from .cloud_confluence_data_source import CloudConfluenceDataSource
8
+ from .cloud_google_drive_data_source import CloudGoogleDriveDataSource
9
+ from .cloud_jira_data_source import CloudJiraDataSource
10
+ from .cloud_notion_page_data_source import CloudNotionPageDataSource
11
+ from .cloud_one_drive_data_source import CloudOneDriveDataSource
12
+ from .cloud_s_3_data_source import CloudS3DataSource
13
+ from .cloud_sharepoint_data_source import CloudSharepointDataSource
14
+ from .cloud_slack_data_source import CloudSlackDataSource
6
15
 
7
- DataSourceCreateComponent = typing.Union[typing.Dict[str, typing.Any], DataSourceCreateComponentOne]
16
+ DataSourceCreateComponent = typing.Union[
17
+ typing.Dict[str, typing.Any],
18
+ CloudS3DataSource,
19
+ CloudAzStorageBlobDataSource,
20
+ CloudGoogleDriveDataSource,
21
+ CloudOneDriveDataSource,
22
+ CloudSharepointDataSource,
23
+ CloudSlackDataSource,
24
+ CloudNotionPageDataSource,
25
+ CloudConfluenceDataSource,
26
+ CloudJiraDataSource,
27
+ CloudBoxDataSource,
28
+ ]
@@ -4,7 +4,6 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
- from .configurable_data_sink_names import ConfigurableDataSinkNames
8
7
 
9
8
  try:
10
9
  import pydantic
@@ -15,20 +14,12 @@ except ImportError:
15
14
  import pydantic # type: ignore
16
15
 
17
16
 
18
- class DataSinkDefinition(pydantic.BaseModel):
19
- """
20
- Schema for a data sink definition.
21
- """
22
-
23
- label: str = pydantic.Field(
24
- description="The label field will be used to display the name of the component in the UI"
25
- )
26
- json_schema: typing.Dict[str, typing.Any] = pydantic.Field(
27
- description="The json_schema field can be used by clients to determine how to construct the component"
28
- )
29
- sink_type: ConfigurableDataSinkNames = pydantic.Field(
30
- description="The name field will act as the unique identifier of DataSinkDefinition objects"
31
- )
17
+ class DocumentBlock(pydantic.BaseModel):
18
+ data: typing.Optional[str]
19
+ path: typing.Optional[str]
20
+ url: typing.Optional[str]
21
+ title: typing.Optional[str]
22
+ document_mimetype: typing.Optional[str]
32
23
 
33
24
  def json(self, **kwargs: typing.Any) -> str:
34
25
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -0,0 +1,17 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import enum
4
+ import typing
5
+
6
+ T_Result = typing.TypeVar("T_Result")
7
+
8
+
9
+ class DocumentChunkMode(str, enum.Enum):
10
+ PAGE = "PAGE"
11
+ SECTION = "SECTION"
12
+
13
+ def visit(self, page: typing.Callable[[], T_Result], section: typing.Callable[[], T_Result]) -> T_Result:
14
+ if self is DocumentChunkMode.PAGE:
15
+ return page()
16
+ if self is DocumentChunkMode.SECTION:
17
+ return section()
@@ -4,6 +4,7 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
+ from .document_chunk_mode import DocumentChunkMode
7
8
  from .extract_mode import ExtractMode
8
9
  from .extract_target import ExtractTarget
9
10
 
@@ -26,6 +27,12 @@ class ExtractConfig(pydantic.BaseModel):
26
27
  system_prompt: typing.Optional[str]
27
28
  use_reasoning: typing.Optional[bool] = pydantic.Field(description="Whether to use reasoning for the extraction.")
28
29
  cite_sources: typing.Optional[bool] = pydantic.Field(description="Whether to cite sources for the extraction.")
30
+ chunk_mode: typing.Optional[DocumentChunkMode] = pydantic.Field(
31
+ description="The mode to use for chunking the document."
32
+ )
33
+ invalidate_cache: typing.Optional[bool] = pydantic.Field(
34
+ description="Whether to invalidate the cache for the extraction."
35
+ )
29
36
 
30
37
  def json(self, **kwargs: typing.Any) -> str:
31
38
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -9,6 +9,7 @@ T_Result = typing.TypeVar("T_Result")
9
9
  class ExtractMode(str, enum.Enum):
10
10
  FAST = "FAST"
11
11
  BALANCED = "BALANCED"
12
+ PREMIUM = "PREMIUM"
12
13
  MULTIMODAL = "MULTIMODAL"
13
14
  ACCURATE = "ACCURATE"
14
15
 
@@ -16,6 +17,7 @@ class ExtractMode(str, enum.Enum):
16
17
  self,
17
18
  fast: typing.Callable[[], T_Result],
18
19
  balanced: typing.Callable[[], T_Result],
20
+ premium: typing.Callable[[], T_Result],
19
21
  multimodal: typing.Callable[[], T_Result],
20
22
  accurate: typing.Callable[[], T_Result],
21
23
  ) -> T_Result:
@@ -23,6 +25,8 @@ class ExtractMode(str, enum.Enum):
23
25
  return fast()
24
26
  if self is ExtractMode.BALANCED:
25
27
  return balanced()
28
+ if self is ExtractMode.PREMIUM:
29
+ return premium()
26
30
  if self is ExtractMode.MULTIMODAL:
27
31
  return multimodal()
28
32
  if self is ExtractMode.ACCURATE: