llama-cloud 0.1.4__py3-none-any.whl → 0.1.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (117) hide show
  1. llama_cloud/__init__.py +76 -10
  2. llama_cloud/client.py +3 -0
  3. llama_cloud/environment.py +1 -1
  4. llama_cloud/resources/__init__.py +23 -1
  5. llama_cloud/resources/data_sinks/client.py +26 -20
  6. llama_cloud/resources/data_sources/client.py +16 -16
  7. llama_cloud/resources/embedding_model_configs/__init__.py +23 -0
  8. llama_cloud/resources/embedding_model_configs/client.py +416 -0
  9. llama_cloud/resources/embedding_model_configs/types/__init__.py +23 -0
  10. llama_cloud/resources/embedding_model_configs/types/embedding_model_config_create_embedding_config.py +89 -0
  11. llama_cloud/resources/evals/client.py +36 -26
  12. llama_cloud/resources/extraction/client.py +32 -32
  13. llama_cloud/resources/files/__init__.py +2 -2
  14. llama_cloud/resources/files/client.py +310 -54
  15. llama_cloud/resources/files/types/__init__.py +3 -1
  16. llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py +7 -0
  17. llama_cloud/resources/files/types/file_create_permission_info_value.py +7 -0
  18. llama_cloud/resources/organizations/client.py +125 -56
  19. llama_cloud/resources/parsing/client.py +652 -264
  20. llama_cloud/resources/pipelines/client.py +617 -310
  21. llama_cloud/resources/projects/client.py +341 -136
  22. llama_cloud/types/__init__.py +58 -10
  23. llama_cloud/types/azure_open_ai_embedding.py +12 -6
  24. llama_cloud/types/base_prompt_template.py +6 -2
  25. llama_cloud/types/bedrock_embedding.py +12 -6
  26. llama_cloud/types/character_splitter.py +4 -2
  27. llama_cloud/types/chat_message.py +1 -1
  28. llama_cloud/types/cloud_az_storage_blob_data_source.py +16 -7
  29. llama_cloud/types/cloud_box_data_source.py +13 -6
  30. llama_cloud/types/cloud_confluence_data_source.py +7 -6
  31. llama_cloud/types/cloud_document.py +3 -1
  32. llama_cloud/types/cloud_document_create.py +3 -1
  33. llama_cloud/types/cloud_google_drive_data_source.py +1 -0
  34. llama_cloud/types/cloud_jira_data_source.py +7 -4
  35. llama_cloud/types/cloud_notion_page_data_source.py +3 -2
  36. llama_cloud/types/cloud_one_drive_data_source.py +6 -2
  37. llama_cloud/types/cloud_postgres_vector_store.py +1 -1
  38. llama_cloud/types/cloud_s_3_data_source.py +9 -4
  39. llama_cloud/types/cloud_sharepoint_data_source.py +9 -5
  40. llama_cloud/types/cloud_slack_data_source.py +7 -6
  41. llama_cloud/types/code_splitter.py +1 -1
  42. llama_cloud/types/cohere_embedding.py +7 -3
  43. llama_cloud/types/data_sink.py +4 -4
  44. llama_cloud/types/data_sink_create.py +1 -1
  45. llama_cloud/types/data_source.py +7 -5
  46. llama_cloud/types/data_source_create.py +4 -2
  47. llama_cloud/types/embedding_model_config.py +43 -0
  48. llama_cloud/types/embedding_model_config_embedding_config.py +89 -0
  49. llama_cloud/types/embedding_model_config_update.py +35 -0
  50. llama_cloud/types/embedding_model_config_update_embedding_config.py +89 -0
  51. llama_cloud/types/eval_dataset.py +2 -2
  52. llama_cloud/types/eval_dataset_job_record.py +13 -7
  53. llama_cloud/types/eval_execution_params_override.py +6 -2
  54. llama_cloud/types/eval_question.py +2 -2
  55. llama_cloud/types/extraction_result.py +2 -2
  56. llama_cloud/types/extraction_schema.py +5 -3
  57. llama_cloud/types/file.py +15 -7
  58. llama_cloud/types/file_permission_info_value.py +5 -0
  59. llama_cloud/types/filter_operator.py +2 -2
  60. llama_cloud/types/gemini_embedding.py +10 -6
  61. llama_cloud/types/hugging_face_inference_api_embedding.py +27 -11
  62. llama_cloud/types/input_message.py +3 -1
  63. llama_cloud/types/interval_usage_and_plan.py +36 -0
  64. llama_cloud/types/job_name_mapping.py +4 -0
  65. llama_cloud/types/llama_parse_parameters.py +21 -0
  66. llama_cloud/types/llm.py +4 -2
  67. llama_cloud/types/llm_parameters.py +5 -2
  68. llama_cloud/types/local_eval.py +10 -8
  69. llama_cloud/types/local_eval_results.py +1 -1
  70. llama_cloud/types/managed_ingestion_status_response.py +5 -3
  71. llama_cloud/types/markdown_element_node_parser.py +5 -3
  72. llama_cloud/types/markdown_node_parser.py +3 -2
  73. llama_cloud/types/metadata_filter.py +2 -2
  74. llama_cloud/types/metric_result.py +3 -3
  75. llama_cloud/types/node_parser.py +1 -1
  76. llama_cloud/types/open_ai_embedding.py +12 -6
  77. llama_cloud/types/organization.py +2 -2
  78. llama_cloud/types/page_splitter_node_parser.py +2 -2
  79. llama_cloud/types/paginated_list_pipeline_files_response.py +35 -0
  80. llama_cloud/types/parsing_job_structured_result.py +32 -0
  81. llama_cloud/types/permission.py +3 -3
  82. llama_cloud/types/pipeline.py +17 -6
  83. llama_cloud/types/pipeline_configuration_hashes.py +3 -3
  84. llama_cloud/types/pipeline_create.py +15 -4
  85. llama_cloud/types/pipeline_data_source.py +13 -7
  86. llama_cloud/types/pipeline_data_source_create.py +3 -1
  87. llama_cloud/types/pipeline_deployment.py +4 -4
  88. llama_cloud/types/pipeline_file.py +25 -10
  89. llama_cloud/types/pipeline_file_create.py +3 -1
  90. llama_cloud/types/pipeline_file_permission_info_value.py +7 -0
  91. llama_cloud/types/plan.py +40 -0
  92. llama_cloud/types/playground_session.py +2 -2
  93. llama_cloud/types/preset_retrieval_params.py +14 -7
  94. llama_cloud/types/presigned_url.py +3 -1
  95. llama_cloud/types/project.py +2 -2
  96. llama_cloud/types/prompt_mixin_prompts.py +1 -1
  97. llama_cloud/types/prompt_spec.py +4 -2
  98. llama_cloud/types/role.py +3 -3
  99. llama_cloud/types/sentence_splitter.py +4 -2
  100. llama_cloud/types/text_node.py +3 -3
  101. llama_cloud/types/{hugging_face_inference_api_embedding_token.py → token.py} +1 -1
  102. llama_cloud/types/token_text_splitter.py +1 -1
  103. llama_cloud/types/usage.py +41 -0
  104. llama_cloud/types/user_organization.py +9 -5
  105. llama_cloud/types/user_organization_create.py +4 -4
  106. llama_cloud/types/user_organization_delete.py +2 -2
  107. llama_cloud/types/user_organization_role.py +2 -2
  108. llama_cloud/types/value.py +5 -0
  109. llama_cloud/types/vertex_text_embedding.py +9 -5
  110. {llama_cloud-0.1.4.dist-info → llama_cloud-0.1.6.dist-info}/METADATA +1 -1
  111. {llama_cloud-0.1.4.dist-info → llama_cloud-0.1.6.dist-info}/RECORD +113 -99
  112. llama_cloud/types/data_sink_component.py +0 -20
  113. llama_cloud/types/data_source_component.py +0 -28
  114. llama_cloud/types/metadata_filter_value.py +0 -5
  115. llama_cloud/types/pipeline_data_source_component.py +0 -28
  116. {llama_cloud-0.1.4.dist-info → llama_cloud-0.1.6.dist-info}/LICENSE +0 -0
  117. {llama_cloud-0.1.4.dist-info → llama_cloud-0.1.6.dist-info}/WHEEL +0 -0
@@ -55,18 +55,38 @@ from .configurable_transformation_names import ConfigurableTransformationNames
55
55
  from .configured_transformation_item import ConfiguredTransformationItem
56
56
  from .configured_transformation_item_component import ConfiguredTransformationItemComponent
57
57
  from .data_sink import DataSink
58
- from .data_sink_component import DataSinkComponent
59
58
  from .data_sink_create import DataSinkCreate
60
59
  from .data_sink_create_component import DataSinkCreateComponent
61
60
  from .data_sink_definition import DataSinkDefinition
62
61
  from .data_source import DataSource
63
- from .data_source_component import DataSourceComponent
64
62
  from .data_source_create import DataSourceCreate
65
63
  from .data_source_create_component import DataSourceCreateComponent
66
64
  from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
67
65
  from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
68
66
  from .data_source_definition import DataSourceDefinition
69
67
  from .element_segmentation_config import ElementSegmentationConfig
68
+ from .embedding_model_config import EmbeddingModelConfig
69
+ from .embedding_model_config_embedding_config import (
70
+ EmbeddingModelConfigEmbeddingConfig,
71
+ EmbeddingModelConfigEmbeddingConfig_AzureEmbedding,
72
+ EmbeddingModelConfigEmbeddingConfig_BedrockEmbedding,
73
+ EmbeddingModelConfigEmbeddingConfig_CohereEmbedding,
74
+ EmbeddingModelConfigEmbeddingConfig_GeminiEmbedding,
75
+ EmbeddingModelConfigEmbeddingConfig_HuggingfaceApiEmbedding,
76
+ EmbeddingModelConfigEmbeddingConfig_OpenaiEmbedding,
77
+ EmbeddingModelConfigEmbeddingConfig_VertexaiEmbedding,
78
+ )
79
+ from .embedding_model_config_update import EmbeddingModelConfigUpdate
80
+ from .embedding_model_config_update_embedding_config import (
81
+ EmbeddingModelConfigUpdateEmbeddingConfig,
82
+ EmbeddingModelConfigUpdateEmbeddingConfig_AzureEmbedding,
83
+ EmbeddingModelConfigUpdateEmbeddingConfig_BedrockEmbedding,
84
+ EmbeddingModelConfigUpdateEmbeddingConfig_CohereEmbedding,
85
+ EmbeddingModelConfigUpdateEmbeddingConfig_GeminiEmbedding,
86
+ EmbeddingModelConfigUpdateEmbeddingConfig_HuggingfaceApiEmbedding,
87
+ EmbeddingModelConfigUpdateEmbeddingConfig_OpenaiEmbedding,
88
+ EmbeddingModelConfigUpdateEmbeddingConfig_VertexaiEmbedding,
89
+ )
70
90
  from .eval_dataset import EvalDataset
71
91
  from .eval_dataset_job_params import EvalDatasetJobParams
72
92
  from .eval_dataset_job_record import EvalDatasetJobRecord
@@ -82,6 +102,7 @@ from .extraction_result_data_value import ExtractionResultDataValue
82
102
  from .extraction_schema import ExtractionSchema
83
103
  from .extraction_schema_data_schema_value import ExtractionSchemaDataSchemaValue
84
104
  from .file import File
105
+ from .file_permission_info_value import FilePermissionInfoValue
85
106
  from .file_resource_info_value import FileResourceInfoValue
86
107
  from .filter_condition import FilterCondition
87
108
  from .filter_operator import FilterOperator
@@ -90,9 +111,9 @@ from .gemini_embedding_config import GeminiEmbeddingConfig
90
111
  from .http_validation_error import HttpValidationError
91
112
  from .hugging_face_inference_api_embedding import HuggingFaceInferenceApiEmbedding
92
113
  from .hugging_face_inference_api_embedding_config import HuggingFaceInferenceApiEmbeddingConfig
93
- from .hugging_face_inference_api_embedding_token import HuggingFaceInferenceApiEmbeddingToken
94
114
  from .ingestion_error_response import IngestionErrorResponse
95
115
  from .input_message import InputMessage
116
+ from .interval_usage_and_plan import IntervalUsageAndPlan
96
117
  from .job_name_mapping import JobNameMapping
97
118
  from .llama_parse_parameters import LlamaParseParameters
98
119
  from .llama_parse_supported_file_extensions import LlamaParseSupportedFileExtensions
@@ -109,7 +130,6 @@ from .markdown_node_parser import MarkdownNodeParser
109
130
  from .message_annotation import MessageAnnotation
110
131
  from .message_role import MessageRole
111
132
  from .metadata_filter import MetadataFilter
112
- from .metadata_filter_value import MetadataFilterValue
113
133
  from .metadata_filters import MetadataFilters
114
134
  from .metadata_filters_filters_item import MetadataFiltersFiltersItem
115
135
  from .metric_result import MetricResult
@@ -126,11 +146,13 @@ from .page_screenshot_metadata import PageScreenshotMetadata
126
146
  from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
127
147
  from .page_segmentation_config import PageSegmentationConfig
128
148
  from .page_splitter_node_parser import PageSplitterNodeParser
149
+ from .paginated_list_pipeline_files_response import PaginatedListPipelineFilesResponse
129
150
  from .parser_languages import ParserLanguages
130
151
  from .parsing_history_item import ParsingHistoryItem
131
152
  from .parsing_job import ParsingJob
132
153
  from .parsing_job_json_result import ParsingJobJsonResult
133
154
  from .parsing_job_markdown_result import ParsingJobMarkdownResult
155
+ from .parsing_job_structured_result import ParsingJobStructuredResult
134
156
  from .parsing_job_text_result import ParsingJobTextResult
135
157
  from .parsing_usage import ParsingUsage
136
158
  from .partition_names import PartitionNames
@@ -150,7 +172,6 @@ from .pipeline_create_embedding_config import (
150
172
  )
151
173
  from .pipeline_create_transform_config import PipelineCreateTransformConfig
152
174
  from .pipeline_data_source import PipelineDataSource
153
- from .pipeline_data_source_component import PipelineDataSourceComponent
154
175
  from .pipeline_data_source_create import PipelineDataSourceCreate
155
176
  from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
156
177
  from .pipeline_deployment import PipelineDeployment
@@ -169,6 +190,7 @@ from .pipeline_file_config_hash_value import PipelineFileConfigHashValue
169
190
  from .pipeline_file_create import PipelineFileCreate
170
191
  from .pipeline_file_create_custom_metadata_value import PipelineFileCreateCustomMetadataValue
171
192
  from .pipeline_file_custom_metadata_value import PipelineFileCustomMetadataValue
193
+ from .pipeline_file_permission_info_value import PipelineFilePermissionInfoValue
172
194
  from .pipeline_file_resource_info_value import PipelineFileResourceInfoValue
173
195
  from .pipeline_transform_config import (
174
196
  PipelineTransformConfig,
@@ -176,6 +198,7 @@ from .pipeline_transform_config import (
176
198
  PipelineTransformConfig_Auto,
177
199
  )
178
200
  from .pipeline_type import PipelineType
201
+ from .plan import Plan
179
202
  from .playground_session import PlaygroundSession
180
203
  from .pooling import Pooling
181
204
  from .preset_retrieval_params import PresetRetrievalParams
@@ -198,15 +221,18 @@ from .supported_llm_model_names import SupportedLlmModelNames
198
221
  from .text_node import TextNode
199
222
  from .text_node_relationships_value import TextNodeRelationshipsValue
200
223
  from .text_node_with_score import TextNodeWithScore
224
+ from .token import Token
201
225
  from .token_chunking_config import TokenChunkingConfig
202
226
  from .token_text_splitter import TokenTextSplitter
203
227
  from .transformation_category_names import TransformationCategoryNames
228
+ from .usage import Usage
204
229
  from .user_organization import UserOrganization
205
230
  from .user_organization_create import UserOrganizationCreate
206
231
  from .user_organization_delete import UserOrganizationDelete
207
232
  from .user_organization_role import UserOrganizationRole
208
233
  from .validation_error import ValidationError
209
234
  from .validation_error_loc_item import ValidationErrorLocItem
235
+ from .value import Value
210
236
  from .vertex_ai_embedding_config import VertexAiEmbeddingConfig
211
237
  from .vertex_embedding_mode import VertexEmbeddingMode
212
238
  from .vertex_text_embedding import VertexTextEmbedding
@@ -263,18 +289,34 @@ __all__ = [
263
289
  "ConfiguredTransformationItem",
264
290
  "ConfiguredTransformationItemComponent",
265
291
  "DataSink",
266
- "DataSinkComponent",
267
292
  "DataSinkCreate",
268
293
  "DataSinkCreateComponent",
269
294
  "DataSinkDefinition",
270
295
  "DataSource",
271
- "DataSourceComponent",
272
296
  "DataSourceCreate",
273
297
  "DataSourceCreateComponent",
274
298
  "DataSourceCreateCustomMetadataValue",
275
299
  "DataSourceCustomMetadataValue",
276
300
  "DataSourceDefinition",
277
301
  "ElementSegmentationConfig",
302
+ "EmbeddingModelConfig",
303
+ "EmbeddingModelConfigEmbeddingConfig",
304
+ "EmbeddingModelConfigEmbeddingConfig_AzureEmbedding",
305
+ "EmbeddingModelConfigEmbeddingConfig_BedrockEmbedding",
306
+ "EmbeddingModelConfigEmbeddingConfig_CohereEmbedding",
307
+ "EmbeddingModelConfigEmbeddingConfig_GeminiEmbedding",
308
+ "EmbeddingModelConfigEmbeddingConfig_HuggingfaceApiEmbedding",
309
+ "EmbeddingModelConfigEmbeddingConfig_OpenaiEmbedding",
310
+ "EmbeddingModelConfigEmbeddingConfig_VertexaiEmbedding",
311
+ "EmbeddingModelConfigUpdate",
312
+ "EmbeddingModelConfigUpdateEmbeddingConfig",
313
+ "EmbeddingModelConfigUpdateEmbeddingConfig_AzureEmbedding",
314
+ "EmbeddingModelConfigUpdateEmbeddingConfig_BedrockEmbedding",
315
+ "EmbeddingModelConfigUpdateEmbeddingConfig_CohereEmbedding",
316
+ "EmbeddingModelConfigUpdateEmbeddingConfig_GeminiEmbedding",
317
+ "EmbeddingModelConfigUpdateEmbeddingConfig_HuggingfaceApiEmbedding",
318
+ "EmbeddingModelConfigUpdateEmbeddingConfig_OpenaiEmbedding",
319
+ "EmbeddingModelConfigUpdateEmbeddingConfig_VertexaiEmbedding",
278
320
  "EvalDataset",
279
321
  "EvalDatasetJobParams",
280
322
  "EvalDatasetJobRecord",
@@ -290,6 +332,7 @@ __all__ = [
290
332
  "ExtractionSchema",
291
333
  "ExtractionSchemaDataSchemaValue",
292
334
  "File",
335
+ "FilePermissionInfoValue",
293
336
  "FileResourceInfoValue",
294
337
  "FilterCondition",
295
338
  "FilterOperator",
@@ -298,9 +341,9 @@ __all__ = [
298
341
  "HttpValidationError",
299
342
  "HuggingFaceInferenceApiEmbedding",
300
343
  "HuggingFaceInferenceApiEmbeddingConfig",
301
- "HuggingFaceInferenceApiEmbeddingToken",
302
344
  "IngestionErrorResponse",
303
345
  "InputMessage",
346
+ "IntervalUsageAndPlan",
304
347
  "JobNameMapping",
305
348
  "LlamaParseParameters",
306
349
  "LlamaParseSupportedFileExtensions",
@@ -317,7 +360,6 @@ __all__ = [
317
360
  "MessageAnnotation",
318
361
  "MessageRole",
319
362
  "MetadataFilter",
320
- "MetadataFilterValue",
321
363
  "MetadataFilters",
322
364
  "MetadataFiltersFiltersItem",
323
365
  "MetricResult",
@@ -334,11 +376,13 @@ __all__ = [
334
376
  "PageScreenshotNodeWithScore",
335
377
  "PageSegmentationConfig",
336
378
  "PageSplitterNodeParser",
379
+ "PaginatedListPipelineFilesResponse",
337
380
  "ParserLanguages",
338
381
  "ParsingHistoryItem",
339
382
  "ParsingJob",
340
383
  "ParsingJobJsonResult",
341
384
  "ParsingJobMarkdownResult",
385
+ "ParsingJobStructuredResult",
342
386
  "ParsingJobTextResult",
343
387
  "ParsingUsage",
344
388
  "PartitionNames",
@@ -356,7 +400,6 @@ __all__ = [
356
400
  "PipelineCreateEmbeddingConfig_VertexaiEmbedding",
357
401
  "PipelineCreateTransformConfig",
358
402
  "PipelineDataSource",
359
- "PipelineDataSourceComponent",
360
403
  "PipelineDataSourceCreate",
361
404
  "PipelineDataSourceCustomMetadataValue",
362
405
  "PipelineDeployment",
@@ -373,11 +416,13 @@ __all__ = [
373
416
  "PipelineFileCreate",
374
417
  "PipelineFileCreateCustomMetadataValue",
375
418
  "PipelineFileCustomMetadataValue",
419
+ "PipelineFilePermissionInfoValue",
376
420
  "PipelineFileResourceInfoValue",
377
421
  "PipelineTransformConfig",
378
422
  "PipelineTransformConfig_Advanced",
379
423
  "PipelineTransformConfig_Auto",
380
424
  "PipelineType",
425
+ "Plan",
381
426
  "PlaygroundSession",
382
427
  "Pooling",
383
428
  "PresetRetrievalParams",
@@ -400,15 +445,18 @@ __all__ = [
400
445
  "TextNode",
401
446
  "TextNodeRelationshipsValue",
402
447
  "TextNodeWithScore",
448
+ "Token",
403
449
  "TokenChunkingConfig",
404
450
  "TokenTextSplitter",
405
451
  "TransformationCategoryNames",
452
+ "Usage",
406
453
  "UserOrganization",
407
454
  "UserOrganizationCreate",
408
455
  "UserOrganizationDelete",
409
456
  "UserOrganizationRole",
410
457
  "ValidationError",
411
458
  "ValidationErrorLocItem",
459
+ "Value",
412
460
  "VertexAiEmbeddingConfig",
413
461
  "VertexEmbeddingMode",
414
462
  "VertexTextEmbedding",
@@ -17,22 +17,28 @@ except ImportError:
17
17
  class AzureOpenAiEmbedding(pydantic.BaseModel):
18
18
  model_name: typing.Optional[str] = pydantic.Field(description="The name of the OpenAI embedding model.")
19
19
  embed_batch_size: typing.Optional[int] = pydantic.Field(description="The batch size for embedding calls.")
20
- num_workers: typing.Optional[int]
20
+ num_workers: typing.Optional[int] = pydantic.Field(
21
+ description="The number of workers to use for async embedding calls."
22
+ )
21
23
  additional_kwargs: typing.Optional[typing.Dict[str, typing.Any]] = pydantic.Field(
22
24
  description="Additional kwargs for the OpenAI API."
23
25
  )
24
- api_key: typing.Optional[str]
26
+ api_key: typing.Optional[str] = pydantic.Field(description="The OpenAI API key.")
25
27
  api_base: typing.Optional[str] = pydantic.Field(description="The base URL for Azure deployment.")
26
28
  api_version: typing.Optional[str] = pydantic.Field(description="The version for Azure OpenAI API.")
27
29
  max_retries: typing.Optional[int] = pydantic.Field(description="Maximum number of retries.")
28
30
  timeout: typing.Optional[float] = pydantic.Field(description="Timeout for each request.")
29
- default_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]]
31
+ default_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]] = pydantic.Field(
32
+ description="The default headers for API requests."
33
+ )
30
34
  reuse_client: typing.Optional[bool] = pydantic.Field(
31
35
  description="Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability."
32
36
  )
33
- dimensions: typing.Optional[int]
34
- azure_endpoint: typing.Optional[str]
35
- azure_deployment: typing.Optional[str]
37
+ dimensions: typing.Optional[int] = pydantic.Field(
38
+ description="The number of dimensions on the output embedding vectors. Works only with v3 embedding models."
39
+ )
40
+ azure_endpoint: typing.Optional[str] = pydantic.Field(description="The Azure endpoint to use.")
41
+ azure_deployment: typing.Optional[str] = pydantic.Field(description="The Azure deployment to use.")
36
42
  class_name: typing.Optional[str]
37
43
 
38
44
  def json(self, **kwargs: typing.Any) -> str:
@@ -19,8 +19,12 @@ class BasePromptTemplate(pydantic.BaseModel):
19
19
  template_vars: typing.List[str]
20
20
  kwargs: typing.Dict[str, str]
21
21
  output_parser: typing.Any
22
- template_var_mappings: typing.Optional[typing.Dict[str, typing.Any]]
23
- function_mappings: typing.Optional[typing.Dict[str, typing.Optional[str]]]
22
+ template_var_mappings: typing.Optional[typing.Dict[str, typing.Any]] = pydantic.Field(
23
+ description="Template variable mappings (Optional)."
24
+ )
25
+ function_mappings: typing.Optional[typing.Dict[str, typing.Optional[str]]] = pydantic.Field(
26
+ description="Function mappings (Optional). This is a mapping from template variable names to functions that take in the current kwargs and return a string."
27
+ )
24
28
 
25
29
  def json(self, **kwargs: typing.Any) -> str:
26
30
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -17,12 +17,18 @@ except ImportError:
17
17
  class BedrockEmbedding(pydantic.BaseModel):
18
18
  model_name: typing.Optional[str] = pydantic.Field(description="The modelId of the Bedrock model to use.")
19
19
  embed_batch_size: typing.Optional[int] = pydantic.Field(description="The batch size for embedding calls.")
20
- num_workers: typing.Optional[int]
21
- profile_name: typing.Optional[str]
22
- aws_access_key_id: typing.Optional[str]
23
- aws_secret_access_key: typing.Optional[str]
24
- aws_session_token: typing.Optional[str]
25
- region_name: typing.Optional[str]
20
+ num_workers: typing.Optional[int] = pydantic.Field(
21
+ description="The number of workers to use for async embedding calls."
22
+ )
23
+ profile_name: typing.Optional[str] = pydantic.Field(
24
+ description="The name of aws profile to use. If not given, then the default profile is used."
25
+ )
26
+ aws_access_key_id: typing.Optional[str] = pydantic.Field(description="AWS Access Key ID to use")
27
+ aws_secret_access_key: typing.Optional[str] = pydantic.Field(description="AWS Secret Access Key to use")
28
+ aws_session_token: typing.Optional[str] = pydantic.Field(description="AWS Session Token to use")
29
+ region_name: typing.Optional[str] = pydantic.Field(
30
+ description="AWS region name to use. Uses region configured in AWS CLI if not passed"
31
+ )
26
32
  max_retries: typing.Optional[int] = pydantic.Field(description="The maximum number of API retries.")
27
33
  timeout: typing.Optional[float] = pydantic.Field(
28
34
  description="The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts."
@@ -24,12 +24,14 @@ class CharacterSplitter(pydantic.BaseModel):
24
24
  )
25
25
  include_prev_next_rel: typing.Optional[bool] = pydantic.Field(description="Include prev/next node relationships.")
26
26
  callback_manager: typing.Optional[typing.Any]
27
- id_func: typing.Optional[str]
27
+ id_func: typing.Optional[str] = pydantic.Field(description="Function to generate node IDs.")
28
28
  chunk_size: typing.Optional[int] = pydantic.Field(description="The token chunk size for each chunk.")
29
29
  chunk_overlap: typing.Optional[int] = pydantic.Field(description="The token overlap of each chunk when splitting.")
30
30
  separator: typing.Optional[str] = pydantic.Field(description="Default separator for splitting into words")
31
31
  paragraph_separator: typing.Optional[str] = pydantic.Field(description="Separator between paragraphs.")
32
- secondary_chunking_regex: typing.Optional[str]
32
+ secondary_chunking_regex: typing.Optional[str] = pydantic.Field(
33
+ description="Backup regex for splitting into sentences."
34
+ )
33
35
  class_name: typing.Optional[str]
34
36
 
35
37
  def json(self, **kwargs: typing.Any) -> str:
@@ -23,7 +23,7 @@ class ChatMessage(pydantic.BaseModel):
23
23
  description="Retrieval annotations for the message."
24
24
  )
25
25
  role: MessageRole
26
- content: typing.Optional[str]
26
+ content: typing.Optional[str] = pydantic.Field(description="Text content of the generation")
27
27
  additional_kwargs: typing.Optional[typing.Dict[str, str]] = pydantic.Field(
28
28
  description="Additional arguments passed to the model"
29
29
  )
@@ -15,15 +15,24 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudAzStorageBlobDataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  container_name: str = pydantic.Field(description="The name of the Azure Storage Blob container to read from.")
19
20
  account_url: str = pydantic.Field(description="The Azure Storage Blob account URL to use for authentication.")
20
- blob: typing.Optional[str]
21
- prefix: typing.Optional[str]
22
- account_name: typing.Optional[str]
23
- account_key: typing.Optional[str]
24
- tenant_id: typing.Optional[str]
25
- client_id: typing.Optional[str]
26
- client_secret: typing.Optional[str]
21
+ blob: typing.Optional[str] = pydantic.Field(description="The blob name to read from.")
22
+ prefix: typing.Optional[str] = pydantic.Field(
23
+ description="The prefix of the Azure Storage Blob objects to read from."
24
+ )
25
+ account_name: typing.Optional[str] = pydantic.Field(
26
+ description="The Azure Storage Blob account name to use for authentication."
27
+ )
28
+ account_key: typing.Optional[str] = pydantic.Field(
29
+ description="The Azure Storage Blob account key to use for authentication."
30
+ )
31
+ tenant_id: typing.Optional[str] = pydantic.Field(description="The Azure AD tenant ID to use for authentication.")
32
+ client_id: typing.Optional[str] = pydantic.Field(description="The Azure AD client ID to use for authentication.")
33
+ client_secret: typing.Optional[str] = pydantic.Field(
34
+ description="The Azure AD client secret to use for authentication."
35
+ )
27
36
  class_name: typing.Optional[str]
28
37
 
29
38
  def json(self, **kwargs: typing.Any) -> str:
@@ -16,15 +16,22 @@ except ImportError:
16
16
 
17
17
 
18
18
  class CloudBoxDataSource(pydantic.BaseModel):
19
- folder_id: typing.Optional[str]
19
+ supports_access_control: typing.Optional[bool]
20
+ folder_id: typing.Optional[str] = pydantic.Field(description="The ID of the Box folder to read from.")
20
21
  authentication_mechanism: BoxAuthMechanism = pydantic.Field(
21
22
  description="The type of authentication to use (Developer Token or CCG)"
22
23
  )
23
- developer_token: typing.Optional[str]
24
- client_id: typing.Optional[str]
25
- client_secret: typing.Optional[str]
26
- user_id: typing.Optional[str]
27
- enterprise_id: typing.Optional[str]
24
+ developer_token: typing.Optional[str] = pydantic.Field(
25
+ description="Developer token for authentication if authentication_mechanism is 'developer_token'."
26
+ )
27
+ client_id: typing.Optional[str] = pydantic.Field(
28
+ description="Box API key used for identifying the application the user is authenticating with"
29
+ )
30
+ client_secret: typing.Optional[str] = pydantic.Field(description="Box API secret used for making auth requests.")
31
+ user_id: typing.Optional[str] = pydantic.Field(description="Box User ID, if provided authenticates as user.")
32
+ enterprise_id: typing.Optional[str] = pydantic.Field(
33
+ description="Box Enterprise ID, if provided authenticates as service."
34
+ )
28
35
  class_name: typing.Optional[str]
29
36
 
30
37
  def json(self, **kwargs: typing.Any) -> str:
@@ -15,16 +15,17 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudConfluenceDataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  server_url: str = pydantic.Field(description="The server URL of the Confluence instance.")
19
20
  authentication_mechanism: str = pydantic.Field(
20
21
  description="Type of Authentication for connecting to Confluence APIs."
21
22
  )
22
- user_name: typing.Optional[str]
23
- api_token: typing.Optional[str]
24
- space_key: typing.Optional[str]
25
- page_ids: typing.Optional[str]
26
- cql: typing.Optional[str]
27
- label: typing.Optional[str]
23
+ user_name: typing.Optional[str] = pydantic.Field(description="The username to use for authentication.")
24
+ api_token: typing.Optional[str] = pydantic.Field(description="The API token to use for authentication.")
25
+ space_key: typing.Optional[str] = pydantic.Field(description="The space key to read from.")
26
+ page_ids: typing.Optional[str] = pydantic.Field(description="The page IDs of the Confluence to read from.")
27
+ cql: typing.Optional[str] = pydantic.Field(description="The CQL query to use for fetching pages.")
28
+ label: typing.Optional[str] = pydantic.Field(description="The label to use for fetching pages.")
28
29
  class_name: typing.Optional[str]
29
30
 
30
31
  def json(self, **kwargs: typing.Any) -> str:
@@ -23,7 +23,9 @@ class CloudDocument(pydantic.BaseModel):
23
23
  metadata: typing.Dict[str, typing.Any]
24
24
  excluded_embed_metadata_keys: typing.Optional[typing.List[str]]
25
25
  excluded_llm_metadata_keys: typing.Optional[typing.List[str]]
26
- page_positions: typing.Optional[typing.List[int]]
26
+ page_positions: typing.Optional[typing.List[int]] = pydantic.Field(
27
+ description="indices in the CloudDocument.text where a new page begins. e.g. Second page starts at index specified by page_positions[1]."
28
+ )
27
29
  id: str
28
30
 
29
31
  def json(self, **kwargs: typing.Any) -> str:
@@ -23,7 +23,9 @@ class CloudDocumentCreate(pydantic.BaseModel):
23
23
  metadata: typing.Dict[str, typing.Any]
24
24
  excluded_embed_metadata_keys: typing.Optional[typing.List[str]]
25
25
  excluded_llm_metadata_keys: typing.Optional[typing.List[str]]
26
- page_positions: typing.Optional[typing.List[int]]
26
+ page_positions: typing.Optional[typing.List[int]] = pydantic.Field(
27
+ description="indices in the CloudDocument.text where a new page begins. e.g. Second page starts at index specified by page_positions[1]."
28
+ )
27
29
  id: typing.Optional[str]
28
30
 
29
31
  def json(self, **kwargs: typing.Any) -> str:
@@ -15,6 +15,7 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudGoogleDriveDataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  folder_id: str = pydantic.Field(description="The ID of the Google Drive folder to read from.")
19
20
  service_account_key: typing.Dict[str, typing.Any] = pydantic.Field(
20
21
  description="The service account key JSON to use for authentication."
@@ -19,10 +19,13 @@ class CloudJiraDataSource(pydantic.BaseModel):
19
19
  Cloud Jira Data Source integrating JiraReader.
20
20
  """
21
21
 
22
- email: typing.Optional[str]
23
- api_token: typing.Optional[str]
24
- server_url: typing.Optional[str]
25
- cloud_id: typing.Optional[str]
22
+ supports_access_control: typing.Optional[bool]
23
+ email: typing.Optional[str] = pydantic.Field(description="The email address to use for authentication.")
24
+ api_token: typing.Optional[str] = pydantic.Field(
25
+ description="The API/ Access Token used for Basic, PAT and OAuth2 authentication."
26
+ )
27
+ server_url: typing.Optional[str] = pydantic.Field(description="The server url for Jira Cloud.")
28
+ cloud_id: typing.Optional[str] = pydantic.Field(description="The cloud ID, used in case of OAuth2.")
26
29
  authentication_mechanism: str = pydantic.Field(description="Type of Authentication for connecting to Jira APIs.")
27
30
  query: str = pydantic.Field(description="JQL (Jira Query Language) query to search.")
28
31
  class_name: typing.Optional[str]
@@ -15,9 +15,10 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudNotionPageDataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  integration_token: str = pydantic.Field(description="The integration token to use for authentication.")
19
- database_ids: typing.Optional[str]
20
- page_ids: typing.Optional[str]
20
+ database_ids: typing.Optional[str] = pydantic.Field(description="The Notion Database Id to read content from.")
21
+ page_ids: typing.Optional[str] = pydantic.Field(description="The Page ID's of the Notion to read from.")
21
22
  class_name: typing.Optional[str]
22
23
 
23
24
  def json(self, **kwargs: typing.Any) -> str:
@@ -15,12 +15,16 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudOneDriveDataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  user_principal_name: str = pydantic.Field(description="The user principal name to use for authentication.")
19
- folder_path: typing.Optional[str]
20
- folder_id: typing.Optional[str]
20
+ folder_path: typing.Optional[str] = pydantic.Field(description="The path of the OneDrive folder to read from.")
21
+ folder_id: typing.Optional[str] = pydantic.Field(description="The ID of the OneDrive folder to read from.")
21
22
  client_id: str = pydantic.Field(description="The client ID to use for authentication.")
22
23
  client_secret: str = pydantic.Field(description="The client secret to use for authentication.")
23
24
  tenant_id: str = pydantic.Field(description="The tenant ID to use for authentication.")
25
+ required_exts: typing.Optional[typing.List[str]] = pydantic.Field(
26
+ description="The list of required file extensions."
27
+ )
24
28
  class_name: typing.Optional[str]
25
29
 
26
30
  def json(self, **kwargs: typing.Any) -> str:
@@ -19,7 +19,7 @@ class CloudPostgresVectorStore(pydantic.BaseModel):
19
19
  database: str
20
20
  host: str
21
21
  password: str
22
- port: str
22
+ port: int
23
23
  user: str
24
24
  table_name: str
25
25
  schema_name: str
@@ -15,11 +15,16 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudS3DataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  bucket: str = pydantic.Field(description="The name of the S3 bucket to read from.")
19
- prefix: typing.Optional[str]
20
- aws_access_id: typing.Optional[str]
21
- aws_access_secret: typing.Optional[str]
22
- s_3_endpoint_url: typing.Optional[str] = pydantic.Field(alias="s3_endpoint_url")
20
+ prefix: typing.Optional[str] = pydantic.Field(description="The prefix of the S3 objects to read from.")
21
+ aws_access_id: typing.Optional[str] = pydantic.Field(description="The AWS access ID to use for authentication.")
22
+ aws_access_secret: typing.Optional[str] = pydantic.Field(
23
+ description="The AWS access secret to use for authentication."
24
+ )
25
+ s_3_endpoint_url: typing.Optional[str] = pydantic.Field(
26
+ alias="s3_endpoint_url", description="The S3 endpoint URL to use for authentication."
27
+ )
23
28
  class_name: typing.Optional[str]
24
29
 
25
30
  def json(self, **kwargs: typing.Any) -> str:
@@ -15,14 +15,18 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudSharepointDataSource(pydantic.BaseModel):
18
- site_name: typing.Optional[str]
19
- site_id: typing.Optional[str]
20
- folder_path: typing.Optional[str]
21
- folder_id: typing.Optional[str]
22
- drive_name: typing.Optional[str]
18
+ supports_access_control: typing.Optional[bool]
19
+ site_name: typing.Optional[str] = pydantic.Field(description="The name of the SharePoint site to download from.")
20
+ site_id: typing.Optional[str] = pydantic.Field(description="The ID of the SharePoint site to download from.")
21
+ folder_path: typing.Optional[str] = pydantic.Field(description="The path of the Sharepoint folder to read from.")
22
+ folder_id: typing.Optional[str] = pydantic.Field(description="The ID of the Sharepoint folder to read from.")
23
+ drive_name: typing.Optional[str] = pydantic.Field(description="The name of the Sharepoint drive to read from.")
23
24
  client_id: str = pydantic.Field(description="The client ID to use for authentication.")
24
25
  client_secret: str = pydantic.Field(description="The client secret to use for authentication.")
25
26
  tenant_id: str = pydantic.Field(description="The tenant ID to use for authentication.")
27
+ required_exts: typing.Optional[typing.List[str]] = pydantic.Field(
28
+ description="The list of required file extensions."
29
+ )
26
30
  class_name: typing.Optional[str]
27
31
 
28
32
  def json(self, **kwargs: typing.Any) -> str:
@@ -15,13 +15,14 @@ except ImportError:
15
15
 
16
16
 
17
17
  class CloudSlackDataSource(pydantic.BaseModel):
18
+ supports_access_control: typing.Optional[bool]
18
19
  slack_token: str = pydantic.Field(description="Slack Bot Token.")
19
- channel_ids: typing.Optional[str]
20
- latest_date: typing.Optional[str]
21
- earliest_date: typing.Optional[str]
22
- earliest_date_timestamp: typing.Optional[float]
23
- latest_date_timestamp: typing.Optional[float]
24
- channel_patterns: typing.Optional[str]
20
+ channel_ids: typing.Optional[str] = pydantic.Field(description="Slack Channel.")
21
+ latest_date: typing.Optional[str] = pydantic.Field(description="Latest date.")
22
+ earliest_date: typing.Optional[str] = pydantic.Field(description="Earliest date.")
23
+ earliest_date_timestamp: typing.Optional[float] = pydantic.Field(description="Earliest date timestamp.")
24
+ latest_date_timestamp: typing.Optional[float] = pydantic.Field(description="Latest date timestamp.")
25
+ channel_patterns: typing.Optional[str] = pydantic.Field(description="Slack Channel name pattern.")
25
26
  class_name: typing.Optional[str]
26
27
 
27
28
  def json(self, **kwargs: typing.Any) -> str:
@@ -27,7 +27,7 @@ class CodeSplitter(pydantic.BaseModel):
27
27
  )
28
28
  include_prev_next_rel: typing.Optional[bool] = pydantic.Field(description="Include prev/next node relationships.")
29
29
  callback_manager: typing.Optional[typing.Any]
30
- id_func: typing.Optional[str]
30
+ id_func: typing.Optional[str] = pydantic.Field(description="Function to generate node IDs.")
31
31
  language: str = pydantic.Field(description="The programming language of the code being split.")
32
32
  chunk_lines: typing.Optional[int] = pydantic.Field(description="The number of lines to include in each chunk.")
33
33
  chunk_lines_overlap: typing.Optional[int] = pydantic.Field(
@@ -17,10 +17,14 @@ except ImportError:
17
17
  class CohereEmbedding(pydantic.BaseModel):
18
18
  model_name: typing.Optional[str] = pydantic.Field(description="The modelId of the Cohere model to use.")
19
19
  embed_batch_size: typing.Optional[int] = pydantic.Field(description="The batch size for embedding calls.")
20
- num_workers: typing.Optional[int]
21
- api_key: typing.Optional[str]
20
+ num_workers: typing.Optional[int] = pydantic.Field(
21
+ description="The number of workers to use for async embedding calls."
22
+ )
23
+ api_key: typing.Optional[str] = pydantic.Field(description="The Cohere API key.")
22
24
  truncate: typing.Optional[str] = pydantic.Field(description="Truncation type - START/ END/ NONE")
23
- input_type: typing.Optional[str]
25
+ input_type: typing.Optional[str] = pydantic.Field(
26
+ description="Model Input type. If not provided, search_document and search_query are used when needed."
27
+ )
24
28
  embedding_type: typing.Optional[str] = pydantic.Field(
25
29
  description="Embedding type. If not provided float embedding_type is used when needed."
26
30
  )