llama-cloud 0.1.6__py3-none-any.whl → 0.1.7a1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (173) hide show
  1. llama_cloud/__init__.py +140 -6
  2. llama_cloud/client.py +15 -0
  3. llama_cloud/environment.py +1 -1
  4. llama_cloud/resources/__init__.py +15 -0
  5. llama_cloud/{types/token.py → resources/chat_apps/__init__.py} +0 -3
  6. llama_cloud/resources/chat_apps/client.py +620 -0
  7. llama_cloud/resources/data_sinks/client.py +12 -12
  8. llama_cloud/resources/data_sources/client.py +14 -14
  9. llama_cloud/resources/embedding_model_configs/client.py +20 -76
  10. llama_cloud/resources/evals/client.py +26 -36
  11. llama_cloud/resources/extraction/client.py +32 -32
  12. llama_cloud/resources/files/client.py +40 -44
  13. llama_cloud/resources/jobs/__init__.py +2 -0
  14. llama_cloud/resources/jobs/client.py +148 -0
  15. llama_cloud/resources/llama_extract/__init__.py +5 -0
  16. llama_cloud/resources/llama_extract/client.py +1038 -0
  17. llama_cloud/resources/llama_extract/types/__init__.py +6 -0
  18. llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_value.py +7 -0
  19. llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema_value.py +7 -0
  20. llama_cloud/resources/organizations/client.py +66 -70
  21. llama_cloud/resources/parsing/client.py +448 -428
  22. llama_cloud/resources/pipelines/client.py +256 -344
  23. llama_cloud/resources/projects/client.py +34 -60
  24. llama_cloud/resources/reports/__init__.py +5 -0
  25. llama_cloud/resources/reports/client.py +1198 -0
  26. llama_cloud/resources/reports/types/__init__.py +7 -0
  27. llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +25 -0
  28. llama_cloud/resources/retrievers/__init__.py +2 -0
  29. llama_cloud/resources/retrievers/client.py +654 -0
  30. llama_cloud/types/__init__.py +128 -6
  31. llama_cloud/types/{chat_message.py → app_schema_chat_chat_message.py} +3 -3
  32. llama_cloud/types/azure_open_ai_embedding.py +6 -12
  33. llama_cloud/types/base_prompt_template.py +2 -6
  34. llama_cloud/types/bedrock_embedding.py +6 -12
  35. llama_cloud/types/character_splitter.py +2 -4
  36. llama_cloud/types/chat_app.py +44 -0
  37. llama_cloud/types/chat_app_response.py +41 -0
  38. llama_cloud/types/cloud_az_storage_blob_data_source.py +7 -15
  39. llama_cloud/types/cloud_box_data_source.py +6 -12
  40. llama_cloud/types/cloud_confluence_data_source.py +6 -6
  41. llama_cloud/types/cloud_document.py +1 -3
  42. llama_cloud/types/cloud_document_create.py +1 -3
  43. llama_cloud/types/cloud_jira_data_source.py +4 -6
  44. llama_cloud/types/cloud_notion_page_data_source.py +2 -2
  45. llama_cloud/types/cloud_one_drive_data_source.py +3 -5
  46. llama_cloud/types/cloud_postgres_vector_store.py +1 -0
  47. llama_cloud/types/cloud_s_3_data_source.py +4 -8
  48. llama_cloud/types/cloud_sharepoint_data_source.py +6 -8
  49. llama_cloud/types/cloud_slack_data_source.py +6 -6
  50. llama_cloud/types/code_splitter.py +1 -1
  51. llama_cloud/types/cohere_embedding.py +3 -7
  52. llama_cloud/types/composite_retrieval_mode.py +21 -0
  53. llama_cloud/types/composite_retrieval_result.py +38 -0
  54. llama_cloud/types/composite_retrieved_text_node.py +42 -0
  55. llama_cloud/types/data_sink.py +4 -4
  56. llama_cloud/types/data_sink_component.py +20 -0
  57. llama_cloud/types/data_source.py +5 -7
  58. llama_cloud/types/data_source_component.py +28 -0
  59. llama_cloud/types/data_source_create.py +1 -3
  60. llama_cloud/types/edit_suggestion.py +39 -0
  61. llama_cloud/types/embedding_model_config.py +2 -2
  62. llama_cloud/types/embedding_model_config_update.py +2 -4
  63. llama_cloud/types/eval_dataset.py +2 -2
  64. llama_cloud/types/eval_dataset_job_record.py +8 -13
  65. llama_cloud/types/eval_execution_params_override.py +2 -6
  66. llama_cloud/types/eval_question.py +2 -2
  67. llama_cloud/types/extract_agent.py +45 -0
  68. llama_cloud/types/extract_agent_data_schema_value.py +5 -0
  69. llama_cloud/types/extract_config.py +40 -0
  70. llama_cloud/types/extract_job.py +35 -0
  71. llama_cloud/types/extract_job_create.py +40 -0
  72. llama_cloud/types/extract_job_create_data_schema_override_value.py +7 -0
  73. llama_cloud/types/extract_mode.py +17 -0
  74. llama_cloud/types/extract_resultset.py +46 -0
  75. llama_cloud/types/extract_resultset_data.py +11 -0
  76. llama_cloud/types/extract_resultset_data_item_value.py +7 -0
  77. llama_cloud/types/extract_resultset_data_zero_value.py +7 -0
  78. llama_cloud/types/extract_resultset_extraction_metadata_value.py +7 -0
  79. llama_cloud/types/extraction_result.py +2 -2
  80. llama_cloud/types/extraction_schema.py +3 -5
  81. llama_cloud/types/file.py +9 -14
  82. llama_cloud/types/filter_condition.py +9 -1
  83. llama_cloud/types/filter_operator.py +6 -2
  84. llama_cloud/types/gemini_embedding.py +6 -10
  85. llama_cloud/types/hugging_face_inference_api_embedding.py +11 -27
  86. llama_cloud/types/hugging_face_inference_api_embedding_token.py +5 -0
  87. llama_cloud/types/image_block.py +35 -0
  88. llama_cloud/types/input_message.py +2 -4
  89. llama_cloud/types/job_names.py +89 -0
  90. llama_cloud/types/job_record.py +57 -0
  91. llama_cloud/types/job_record_with_usage_metrics.py +36 -0
  92. llama_cloud/types/llama_index_core_base_llms_types_chat_message.py +39 -0
  93. llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +33 -0
  94. llama_cloud/types/llama_parse_parameters.py +4 -0
  95. llama_cloud/types/llm.py +3 -4
  96. llama_cloud/types/llm_model_data.py +1 -0
  97. llama_cloud/types/llm_parameters.py +3 -5
  98. llama_cloud/types/local_eval.py +8 -10
  99. llama_cloud/types/local_eval_results.py +1 -1
  100. llama_cloud/types/managed_ingestion_status.py +4 -0
  101. llama_cloud/types/managed_ingestion_status_response.py +4 -5
  102. llama_cloud/types/markdown_element_node_parser.py +3 -5
  103. llama_cloud/types/markdown_node_parser.py +1 -1
  104. llama_cloud/types/metadata_filter.py +2 -2
  105. llama_cloud/types/metadata_filter_value.py +5 -0
  106. llama_cloud/types/metric_result.py +3 -3
  107. llama_cloud/types/node_parser.py +1 -1
  108. llama_cloud/types/object_type.py +4 -0
  109. llama_cloud/types/open_ai_embedding.py +6 -12
  110. llama_cloud/types/organization.py +7 -2
  111. llama_cloud/types/page_splitter_node_parser.py +2 -2
  112. llama_cloud/types/paginated_jobs_history_with_metrics.py +35 -0
  113. llama_cloud/types/paginated_report_response.py +35 -0
  114. llama_cloud/types/parse_plan_level.py +21 -0
  115. llama_cloud/types/permission.py +3 -3
  116. llama_cloud/types/pipeline.py +7 -17
  117. llama_cloud/types/pipeline_configuration_hashes.py +3 -3
  118. llama_cloud/types/pipeline_create.py +8 -16
  119. llama_cloud/types/pipeline_data_source.py +7 -13
  120. llama_cloud/types/pipeline_data_source_component.py +28 -0
  121. llama_cloud/types/pipeline_data_source_create.py +1 -3
  122. llama_cloud/types/pipeline_deployment.py +4 -4
  123. llama_cloud/types/pipeline_file.py +13 -24
  124. llama_cloud/types/pipeline_file_create.py +1 -3
  125. llama_cloud/types/playground_session.py +4 -4
  126. llama_cloud/types/preset_retrieval_params.py +8 -14
  127. llama_cloud/types/presigned_url.py +1 -3
  128. llama_cloud/types/progress_event.py +44 -0
  129. llama_cloud/types/progress_event_status.py +33 -0
  130. llama_cloud/types/project.py +2 -2
  131. llama_cloud/types/prompt_mixin_prompts.py +1 -1
  132. llama_cloud/types/prompt_spec.py +3 -5
  133. llama_cloud/types/related_node_info.py +2 -2
  134. llama_cloud/types/related_node_info_node_type.py +7 -0
  135. llama_cloud/types/report.py +33 -0
  136. llama_cloud/types/report_block.py +34 -0
  137. llama_cloud/types/report_block_dependency.py +29 -0
  138. llama_cloud/types/report_create_response.py +31 -0
  139. llama_cloud/types/report_event_item.py +40 -0
  140. llama_cloud/types/report_event_item_event_data.py +45 -0
  141. llama_cloud/types/report_event_type.py +37 -0
  142. llama_cloud/types/report_metadata.py +39 -0
  143. llama_cloud/types/report_plan.py +36 -0
  144. llama_cloud/types/report_plan_block.py +36 -0
  145. llama_cloud/types/report_query.py +33 -0
  146. llama_cloud/types/report_response.py +41 -0
  147. llama_cloud/types/report_state.py +37 -0
  148. llama_cloud/types/report_state_event.py +38 -0
  149. llama_cloud/types/report_update_event.py +38 -0
  150. llama_cloud/types/retrieve_results.py +1 -1
  151. llama_cloud/types/retriever.py +45 -0
  152. llama_cloud/types/retriever_create.py +37 -0
  153. llama_cloud/types/retriever_pipeline.py +37 -0
  154. llama_cloud/types/role.py +3 -3
  155. llama_cloud/types/sentence_splitter.py +2 -4
  156. llama_cloud/types/status_enum.py +4 -0
  157. llama_cloud/types/supported_llm_model_names.py +4 -0
  158. llama_cloud/types/text_block.py +31 -0
  159. llama_cloud/types/text_node.py +15 -8
  160. llama_cloud/types/token_text_splitter.py +1 -1
  161. llama_cloud/types/usage_metric_response.py +34 -0
  162. llama_cloud/types/user_job_record.py +32 -0
  163. llama_cloud/types/user_organization.py +5 -9
  164. llama_cloud/types/user_organization_create.py +4 -4
  165. llama_cloud/types/user_organization_delete.py +2 -2
  166. llama_cloud/types/user_organization_role.py +2 -2
  167. llama_cloud/types/vertex_text_embedding.py +5 -9
  168. {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7a1.dist-info}/METADATA +2 -1
  169. llama_cloud-0.1.7a1.dist-info/RECORD +310 -0
  170. llama_cloud/types/value.py +0 -5
  171. llama_cloud-0.1.6.dist-info/RECORD +0 -241
  172. {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7a1.dist-info}/LICENSE +0 -0
  173. {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7a1.dist-info}/WHEEL +0 -0
@@ -15,6 +15,7 @@ from .advanced_mode_transform_config_segmentation_config import (
15
15
  AdvancedModeTransformConfigSegmentationConfig_None,
16
16
  AdvancedModeTransformConfigSegmentationConfig_Page,
17
17
  )
18
+ from .app_schema_chat_chat_message import AppSchemaChatChatMessage
18
19
  from .auto_transform_config import AutoTransformConfig
19
20
  from .azure_open_ai_embedding import AzureOpenAiEmbedding
20
21
  from .azure_open_ai_embedding_config import AzureOpenAiEmbeddingConfig
@@ -25,8 +26,9 @@ from .bedrock_embedding_config import BedrockEmbeddingConfig
25
26
  from .box_auth_mechanism import BoxAuthMechanism
26
27
  from .character_chunking_config import CharacterChunkingConfig
27
28
  from .character_splitter import CharacterSplitter
29
+ from .chat_app import ChatApp
30
+ from .chat_app_response import ChatAppResponse
28
31
  from .chat_data import ChatData
29
- from .chat_message import ChatMessage
30
32
  from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
31
33
  from .cloud_azure_ai_search_vector_store import CloudAzureAiSearchVectorStore
32
34
  from .cloud_box_data_source import CloudBoxDataSource
@@ -48,6 +50,9 @@ from .cloud_slack_data_source import CloudSlackDataSource
48
50
  from .code_splitter import CodeSplitter
49
51
  from .cohere_embedding import CohereEmbedding
50
52
  from .cohere_embedding_config import CohereEmbeddingConfig
53
+ from .composite_retrieval_mode import CompositeRetrievalMode
54
+ from .composite_retrieval_result import CompositeRetrievalResult
55
+ from .composite_retrieved_text_node import CompositeRetrievedTextNode
51
56
  from .configurable_data_sink_names import ConfigurableDataSinkNames
52
57
  from .configurable_data_source_names import ConfigurableDataSourceNames
53
58
  from .configurable_transformation_definition import ConfigurableTransformationDefinition
@@ -55,15 +60,18 @@ from .configurable_transformation_names import ConfigurableTransformationNames
55
60
  from .configured_transformation_item import ConfiguredTransformationItem
56
61
  from .configured_transformation_item_component import ConfiguredTransformationItemComponent
57
62
  from .data_sink import DataSink
63
+ from .data_sink_component import DataSinkComponent
58
64
  from .data_sink_create import DataSinkCreate
59
65
  from .data_sink_create_component import DataSinkCreateComponent
60
66
  from .data_sink_definition import DataSinkDefinition
61
67
  from .data_source import DataSource
68
+ from .data_source_component import DataSourceComponent
62
69
  from .data_source_create import DataSourceCreate
63
70
  from .data_source_create_component import DataSourceCreateComponent
64
71
  from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
65
72
  from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
66
73
  from .data_source_definition import DataSourceDefinition
74
+ from .edit_suggestion import EditSuggestion
67
75
  from .element_segmentation_config import ElementSegmentationConfig
68
76
  from .embedding_model_config import EmbeddingModelConfig
69
77
  from .embedding_model_config_embedding_config import (
@@ -96,6 +104,18 @@ from .eval_metric import EvalMetric
96
104
  from .eval_question import EvalQuestion
97
105
  from .eval_question_create import EvalQuestionCreate
98
106
  from .eval_question_result import EvalQuestionResult
107
+ from .extract_agent import ExtractAgent
108
+ from .extract_agent_data_schema_value import ExtractAgentDataSchemaValue
109
+ from .extract_config import ExtractConfig
110
+ from .extract_job import ExtractJob
111
+ from .extract_job_create import ExtractJobCreate
112
+ from .extract_job_create_data_schema_override_value import ExtractJobCreateDataSchemaOverrideValue
113
+ from .extract_mode import ExtractMode
114
+ from .extract_resultset import ExtractResultset
115
+ from .extract_resultset_data import ExtractResultsetData
116
+ from .extract_resultset_data_item_value import ExtractResultsetDataItemValue
117
+ from .extract_resultset_data_zero_value import ExtractResultsetDataZeroValue
118
+ from .extract_resultset_extraction_metadata_value import ExtractResultsetExtractionMetadataValue
99
119
  from .extraction_job import ExtractionJob
100
120
  from .extraction_result import ExtractionResult
101
121
  from .extraction_result_data_value import ExtractionResultDataValue
@@ -111,10 +131,21 @@ from .gemini_embedding_config import GeminiEmbeddingConfig
111
131
  from .http_validation_error import HttpValidationError
112
132
  from .hugging_face_inference_api_embedding import HuggingFaceInferenceApiEmbedding
113
133
  from .hugging_face_inference_api_embedding_config import HuggingFaceInferenceApiEmbeddingConfig
134
+ from .hugging_face_inference_api_embedding_token import HuggingFaceInferenceApiEmbeddingToken
135
+ from .image_block import ImageBlock
114
136
  from .ingestion_error_response import IngestionErrorResponse
115
137
  from .input_message import InputMessage
116
138
  from .interval_usage_and_plan import IntervalUsageAndPlan
117
139
  from .job_name_mapping import JobNameMapping
140
+ from .job_names import JobNames
141
+ from .job_record import JobRecord
142
+ from .job_record_with_usage_metrics import JobRecordWithUsageMetrics
143
+ from .llama_index_core_base_llms_types_chat_message import LlamaIndexCoreBaseLlmsTypesChatMessage
144
+ from .llama_index_core_base_llms_types_chat_message_blocks_item import (
145
+ LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem,
146
+ LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image,
147
+ LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text,
148
+ )
118
149
  from .llama_parse_parameters import LlamaParseParameters
119
150
  from .llama_parse_supported_file_extensions import LlamaParseSupportedFileExtensions
120
151
  from .llm import Llm
@@ -130,6 +161,7 @@ from .markdown_node_parser import MarkdownNodeParser
130
161
  from .message_annotation import MessageAnnotation
131
162
  from .message_role import MessageRole
132
163
  from .metadata_filter import MetadataFilter
164
+ from .metadata_filter_value import MetadataFilterValue
133
165
  from .metadata_filters import MetadataFilters
134
166
  from .metadata_filters_filters_item import MetadataFiltersFiltersItem
135
167
  from .metric_result import MetricResult
@@ -146,7 +178,10 @@ from .page_screenshot_metadata import PageScreenshotMetadata
146
178
  from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
147
179
  from .page_segmentation_config import PageSegmentationConfig
148
180
  from .page_splitter_node_parser import PageSplitterNodeParser
181
+ from .paginated_jobs_history_with_metrics import PaginatedJobsHistoryWithMetrics
149
182
  from .paginated_list_pipeline_files_response import PaginatedListPipelineFilesResponse
183
+ from .paginated_report_response import PaginatedReportResponse
184
+ from .parse_plan_level import ParsePlanLevel
150
185
  from .parser_languages import ParserLanguages
151
186
  from .parsing_history_item import ParsingHistoryItem
152
187
  from .parsing_job import ParsingJob
@@ -172,6 +207,7 @@ from .pipeline_create_embedding_config import (
172
207
  )
173
208
  from .pipeline_create_transform_config import PipelineCreateTransformConfig
174
209
  from .pipeline_data_source import PipelineDataSource
210
+ from .pipeline_data_source_component import PipelineDataSourceComponent
175
211
  from .pipeline_data_source_create import PipelineDataSourceCreate
176
212
  from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
177
213
  from .pipeline_deployment import PipelineDeployment
@@ -203,14 +239,40 @@ from .playground_session import PlaygroundSession
203
239
  from .pooling import Pooling
204
240
  from .preset_retrieval_params import PresetRetrievalParams
205
241
  from .presigned_url import PresignedUrl
242
+ from .progress_event import ProgressEvent
243
+ from .progress_event_status import ProgressEventStatus
206
244
  from .project import Project
207
245
  from .project_create import ProjectCreate
208
246
  from .prompt_mixin_prompts import PromptMixinPrompts
209
247
  from .prompt_spec import PromptSpec
210
248
  from .pydantic_program_mode import PydanticProgramMode
211
249
  from .related_node_info import RelatedNodeInfo
250
+ from .related_node_info_node_type import RelatedNodeInfoNodeType
251
+ from .report import Report
252
+ from .report_block import ReportBlock
253
+ from .report_block_dependency import ReportBlockDependency
254
+ from .report_create_response import ReportCreateResponse
255
+ from .report_event_item import ReportEventItem
256
+ from .report_event_item_event_data import (
257
+ ReportEventItemEventData,
258
+ ReportEventItemEventData_Progress,
259
+ ReportEventItemEventData_ReportBlockUpdate,
260
+ ReportEventItemEventData_ReportStateUpdate,
261
+ )
262
+ from .report_event_type import ReportEventType
263
+ from .report_metadata import ReportMetadata
264
+ from .report_plan import ReportPlan
265
+ from .report_plan_block import ReportPlanBlock
266
+ from .report_query import ReportQuery
267
+ from .report_response import ReportResponse
268
+ from .report_state import ReportState
269
+ from .report_state_event import ReportStateEvent
270
+ from .report_update_event import ReportUpdateEvent
212
271
  from .retrieval_mode import RetrievalMode
213
272
  from .retrieve_results import RetrieveResults
273
+ from .retriever import Retriever
274
+ from .retriever_create import RetrieverCreate
275
+ from .retriever_pipeline import RetrieverPipeline
214
276
  from .role import Role
215
277
  from .semantic_chunking_config import SemanticChunkingConfig
216
278
  from .sentence_chunking_config import SentenceChunkingConfig
@@ -218,21 +280,22 @@ from .sentence_splitter import SentenceSplitter
218
280
  from .status_enum import StatusEnum
219
281
  from .supported_llm_model import SupportedLlmModel
220
282
  from .supported_llm_model_names import SupportedLlmModelNames
283
+ from .text_block import TextBlock
221
284
  from .text_node import TextNode
222
285
  from .text_node_relationships_value import TextNodeRelationshipsValue
223
286
  from .text_node_with_score import TextNodeWithScore
224
- from .token import Token
225
287
  from .token_chunking_config import TokenChunkingConfig
226
288
  from .token_text_splitter import TokenTextSplitter
227
289
  from .transformation_category_names import TransformationCategoryNames
228
290
  from .usage import Usage
291
+ from .usage_metric_response import UsageMetricResponse
292
+ from .user_job_record import UserJobRecord
229
293
  from .user_organization import UserOrganization
230
294
  from .user_organization_create import UserOrganizationCreate
231
295
  from .user_organization_delete import UserOrganizationDelete
232
296
  from .user_organization_role import UserOrganizationRole
233
297
  from .validation_error import ValidationError
234
298
  from .validation_error_loc_item import ValidationErrorLocItem
235
- from .value import Value
236
299
  from .vertex_ai_embedding_config import VertexAiEmbeddingConfig
237
300
  from .vertex_embedding_mode import VertexEmbeddingMode
238
301
  from .vertex_text_embedding import VertexTextEmbedding
@@ -249,6 +312,7 @@ __all__ = [
249
312
  "AdvancedModeTransformConfigSegmentationConfig_Element",
250
313
  "AdvancedModeTransformConfigSegmentationConfig_None",
251
314
  "AdvancedModeTransformConfigSegmentationConfig_Page",
315
+ "AppSchemaChatChatMessage",
252
316
  "AutoTransformConfig",
253
317
  "AzureOpenAiEmbedding",
254
318
  "AzureOpenAiEmbeddingConfig",
@@ -259,8 +323,9 @@ __all__ = [
259
323
  "BoxAuthMechanism",
260
324
  "CharacterChunkingConfig",
261
325
  "CharacterSplitter",
326
+ "ChatApp",
327
+ "ChatAppResponse",
262
328
  "ChatData",
263
- "ChatMessage",
264
329
  "CloudAzStorageBlobDataSource",
265
330
  "CloudAzureAiSearchVectorStore",
266
331
  "CloudBoxDataSource",
@@ -282,6 +347,9 @@ __all__ = [
282
347
  "CodeSplitter",
283
348
  "CohereEmbedding",
284
349
  "CohereEmbeddingConfig",
350
+ "CompositeRetrievalMode",
351
+ "CompositeRetrievalResult",
352
+ "CompositeRetrievedTextNode",
285
353
  "ConfigurableDataSinkNames",
286
354
  "ConfigurableDataSourceNames",
287
355
  "ConfigurableTransformationDefinition",
@@ -289,15 +357,18 @@ __all__ = [
289
357
  "ConfiguredTransformationItem",
290
358
  "ConfiguredTransformationItemComponent",
291
359
  "DataSink",
360
+ "DataSinkComponent",
292
361
  "DataSinkCreate",
293
362
  "DataSinkCreateComponent",
294
363
  "DataSinkDefinition",
295
364
  "DataSource",
365
+ "DataSourceComponent",
296
366
  "DataSourceCreate",
297
367
  "DataSourceCreateComponent",
298
368
  "DataSourceCreateCustomMetadataValue",
299
369
  "DataSourceCustomMetadataValue",
300
370
  "DataSourceDefinition",
371
+ "EditSuggestion",
301
372
  "ElementSegmentationConfig",
302
373
  "EmbeddingModelConfig",
303
374
  "EmbeddingModelConfigEmbeddingConfig",
@@ -326,6 +397,18 @@ __all__ = [
326
397
  "EvalQuestion",
327
398
  "EvalQuestionCreate",
328
399
  "EvalQuestionResult",
400
+ "ExtractAgent",
401
+ "ExtractAgentDataSchemaValue",
402
+ "ExtractConfig",
403
+ "ExtractJob",
404
+ "ExtractJobCreate",
405
+ "ExtractJobCreateDataSchemaOverrideValue",
406
+ "ExtractMode",
407
+ "ExtractResultset",
408
+ "ExtractResultsetData",
409
+ "ExtractResultsetDataItemValue",
410
+ "ExtractResultsetDataZeroValue",
411
+ "ExtractResultsetExtractionMetadataValue",
329
412
  "ExtractionJob",
330
413
  "ExtractionResult",
331
414
  "ExtractionResultDataValue",
@@ -341,10 +424,19 @@ __all__ = [
341
424
  "HttpValidationError",
342
425
  "HuggingFaceInferenceApiEmbedding",
343
426
  "HuggingFaceInferenceApiEmbeddingConfig",
427
+ "HuggingFaceInferenceApiEmbeddingToken",
428
+ "ImageBlock",
344
429
  "IngestionErrorResponse",
345
430
  "InputMessage",
346
431
  "IntervalUsageAndPlan",
347
432
  "JobNameMapping",
433
+ "JobNames",
434
+ "JobRecord",
435
+ "JobRecordWithUsageMetrics",
436
+ "LlamaIndexCoreBaseLlmsTypesChatMessage",
437
+ "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem",
438
+ "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image",
439
+ "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text",
348
440
  "LlamaParseParameters",
349
441
  "LlamaParseSupportedFileExtensions",
350
442
  "Llm",
@@ -360,6 +452,7 @@ __all__ = [
360
452
  "MessageAnnotation",
361
453
  "MessageRole",
362
454
  "MetadataFilter",
455
+ "MetadataFilterValue",
363
456
  "MetadataFilters",
364
457
  "MetadataFiltersFiltersItem",
365
458
  "MetricResult",
@@ -376,7 +469,10 @@ __all__ = [
376
469
  "PageScreenshotNodeWithScore",
377
470
  "PageSegmentationConfig",
378
471
  "PageSplitterNodeParser",
472
+ "PaginatedJobsHistoryWithMetrics",
379
473
  "PaginatedListPipelineFilesResponse",
474
+ "PaginatedReportResponse",
475
+ "ParsePlanLevel",
380
476
  "ParserLanguages",
381
477
  "ParsingHistoryItem",
382
478
  "ParsingJob",
@@ -400,6 +496,7 @@ __all__ = [
400
496
  "PipelineCreateEmbeddingConfig_VertexaiEmbedding",
401
497
  "PipelineCreateTransformConfig",
402
498
  "PipelineDataSource",
499
+ "PipelineDataSourceComponent",
403
500
  "PipelineDataSourceCreate",
404
501
  "PipelineDataSourceCustomMetadataValue",
405
502
  "PipelineDeployment",
@@ -427,14 +524,38 @@ __all__ = [
427
524
  "Pooling",
428
525
  "PresetRetrievalParams",
429
526
  "PresignedUrl",
527
+ "ProgressEvent",
528
+ "ProgressEventStatus",
430
529
  "Project",
431
530
  "ProjectCreate",
432
531
  "PromptMixinPrompts",
433
532
  "PromptSpec",
434
533
  "PydanticProgramMode",
435
534
  "RelatedNodeInfo",
535
+ "RelatedNodeInfoNodeType",
536
+ "Report",
537
+ "ReportBlock",
538
+ "ReportBlockDependency",
539
+ "ReportCreateResponse",
540
+ "ReportEventItem",
541
+ "ReportEventItemEventData",
542
+ "ReportEventItemEventData_Progress",
543
+ "ReportEventItemEventData_ReportBlockUpdate",
544
+ "ReportEventItemEventData_ReportStateUpdate",
545
+ "ReportEventType",
546
+ "ReportMetadata",
547
+ "ReportPlan",
548
+ "ReportPlanBlock",
549
+ "ReportQuery",
550
+ "ReportResponse",
551
+ "ReportState",
552
+ "ReportStateEvent",
553
+ "ReportUpdateEvent",
436
554
  "RetrievalMode",
437
555
  "RetrieveResults",
556
+ "Retriever",
557
+ "RetrieverCreate",
558
+ "RetrieverPipeline",
438
559
  "Role",
439
560
  "SemanticChunkingConfig",
440
561
  "SentenceChunkingConfig",
@@ -442,21 +563,22 @@ __all__ = [
442
563
  "StatusEnum",
443
564
  "SupportedLlmModel",
444
565
  "SupportedLlmModelNames",
566
+ "TextBlock",
445
567
  "TextNode",
446
568
  "TextNodeRelationshipsValue",
447
569
  "TextNodeWithScore",
448
- "Token",
449
570
  "TokenChunkingConfig",
450
571
  "TokenTextSplitter",
451
572
  "TransformationCategoryNames",
452
573
  "Usage",
574
+ "UsageMetricResponse",
575
+ "UserJobRecord",
453
576
  "UserOrganization",
454
577
  "UserOrganizationCreate",
455
578
  "UserOrganizationDelete",
456
579
  "UserOrganizationRole",
457
580
  "ValidationError",
458
581
  "ValidationErrorLocItem",
459
- "Value",
460
582
  "VertexAiEmbeddingConfig",
461
583
  "VertexEmbeddingMode",
462
584
  "VertexTextEmbedding",
@@ -16,14 +16,14 @@ except ImportError:
16
16
  import pydantic # type: ignore
17
17
 
18
18
 
19
- class ChatMessage(pydantic.BaseModel):
19
+ class AppSchemaChatChatMessage(pydantic.BaseModel):
20
20
  id: str
21
21
  index: int = pydantic.Field(description="The index of the message in the chat.")
22
22
  annotations: typing.Optional[typing.List[MessageAnnotation]] = pydantic.Field(
23
23
  description="Retrieval annotations for the message."
24
24
  )
25
- role: MessageRole
26
- content: typing.Optional[str] = pydantic.Field(description="Text content of the generation")
25
+ role: MessageRole = pydantic.Field(description="The role of the message.")
26
+ content: typing.Optional[str]
27
27
  additional_kwargs: typing.Optional[typing.Dict[str, str]] = pydantic.Field(
28
28
  description="Additional arguments passed to the model"
29
29
  )
@@ -17,28 +17,22 @@ except ImportError:
17
17
  class AzureOpenAiEmbedding(pydantic.BaseModel):
18
18
  model_name: typing.Optional[str] = pydantic.Field(description="The name of the OpenAI embedding model.")
19
19
  embed_batch_size: typing.Optional[int] = pydantic.Field(description="The batch size for embedding calls.")
20
- num_workers: typing.Optional[int] = pydantic.Field(
21
- description="The number of workers to use for async embedding calls."
22
- )
20
+ num_workers: typing.Optional[int]
23
21
  additional_kwargs: typing.Optional[typing.Dict[str, typing.Any]] = pydantic.Field(
24
22
  description="Additional kwargs for the OpenAI API."
25
23
  )
26
- api_key: typing.Optional[str] = pydantic.Field(description="The OpenAI API key.")
24
+ api_key: typing.Optional[str]
27
25
  api_base: typing.Optional[str] = pydantic.Field(description="The base URL for Azure deployment.")
28
26
  api_version: typing.Optional[str] = pydantic.Field(description="The version for Azure OpenAI API.")
29
27
  max_retries: typing.Optional[int] = pydantic.Field(description="Maximum number of retries.")
30
28
  timeout: typing.Optional[float] = pydantic.Field(description="Timeout for each request.")
31
- default_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]] = pydantic.Field(
32
- description="The default headers for API requests."
33
- )
29
+ default_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]]
34
30
  reuse_client: typing.Optional[bool] = pydantic.Field(
35
31
  description="Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability."
36
32
  )
37
- dimensions: typing.Optional[int] = pydantic.Field(
38
- description="The number of dimensions on the output embedding vectors. Works only with v3 embedding models."
39
- )
40
- azure_endpoint: typing.Optional[str] = pydantic.Field(description="The Azure endpoint to use.")
41
- azure_deployment: typing.Optional[str] = pydantic.Field(description="The Azure deployment to use.")
33
+ dimensions: typing.Optional[int]
34
+ azure_endpoint: typing.Optional[str]
35
+ azure_deployment: typing.Optional[str]
42
36
  class_name: typing.Optional[str]
43
37
 
44
38
  def json(self, **kwargs: typing.Any) -> str:
@@ -19,12 +19,8 @@ class BasePromptTemplate(pydantic.BaseModel):
19
19
  template_vars: typing.List[str]
20
20
  kwargs: typing.Dict[str, str]
21
21
  output_parser: typing.Any
22
- template_var_mappings: typing.Optional[typing.Dict[str, typing.Any]] = pydantic.Field(
23
- description="Template variable mappings (Optional)."
24
- )
25
- function_mappings: typing.Optional[typing.Dict[str, typing.Optional[str]]] = pydantic.Field(
26
- description="Function mappings (Optional). This is a mapping from template variable names to functions that take in the current kwargs and return a string."
27
- )
22
+ template_var_mappings: typing.Optional[typing.Dict[str, typing.Any]]
23
+ function_mappings: typing.Optional[typing.Dict[str, typing.Optional[str]]]
28
24
 
29
25
  def json(self, **kwargs: typing.Any) -> str:
30
26
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -17,18 +17,12 @@ except ImportError:
17
17
  class BedrockEmbedding(pydantic.BaseModel):
18
18
  model_name: typing.Optional[str] = pydantic.Field(description="The modelId of the Bedrock model to use.")
19
19
  embed_batch_size: typing.Optional[int] = pydantic.Field(description="The batch size for embedding calls.")
20
- num_workers: typing.Optional[int] = pydantic.Field(
21
- description="The number of workers to use for async embedding calls."
22
- )
23
- profile_name: typing.Optional[str] = pydantic.Field(
24
- description="The name of aws profile to use. If not given, then the default profile is used."
25
- )
26
- aws_access_key_id: typing.Optional[str] = pydantic.Field(description="AWS Access Key ID to use")
27
- aws_secret_access_key: typing.Optional[str] = pydantic.Field(description="AWS Secret Access Key to use")
28
- aws_session_token: typing.Optional[str] = pydantic.Field(description="AWS Session Token to use")
29
- region_name: typing.Optional[str] = pydantic.Field(
30
- description="AWS region name to use. Uses region configured in AWS CLI if not passed"
31
- )
20
+ num_workers: typing.Optional[int]
21
+ profile_name: typing.Optional[str]
22
+ aws_access_key_id: typing.Optional[str]
23
+ aws_secret_access_key: typing.Optional[str]
24
+ aws_session_token: typing.Optional[str]
25
+ region_name: typing.Optional[str]
32
26
  max_retries: typing.Optional[int] = pydantic.Field(description="The maximum number of API retries.")
33
27
  timeout: typing.Optional[float] = pydantic.Field(
34
28
  description="The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts."
@@ -24,14 +24,12 @@ class CharacterSplitter(pydantic.BaseModel):
24
24
  )
25
25
  include_prev_next_rel: typing.Optional[bool] = pydantic.Field(description="Include prev/next node relationships.")
26
26
  callback_manager: typing.Optional[typing.Any]
27
- id_func: typing.Optional[str] = pydantic.Field(description="Function to generate node IDs.")
27
+ id_func: typing.Optional[str]
28
28
  chunk_size: typing.Optional[int] = pydantic.Field(description="The token chunk size for each chunk.")
29
29
  chunk_overlap: typing.Optional[int] = pydantic.Field(description="The token overlap of each chunk when splitting.")
30
30
  separator: typing.Optional[str] = pydantic.Field(description="Default separator for splitting into words")
31
31
  paragraph_separator: typing.Optional[str] = pydantic.Field(description="Separator between paragraphs.")
32
- secondary_chunking_regex: typing.Optional[str] = pydantic.Field(
33
- description="Backup regex for splitting into sentences."
34
- )
32
+ secondary_chunking_regex: typing.Optional[str]
35
33
  class_name: typing.Optional[str]
36
34
 
37
35
  def json(self, **kwargs: typing.Any) -> str:
@@ -0,0 +1,44 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+ from .llm_parameters import LlmParameters
8
+ from .preset_retrieval_params import PresetRetrievalParams
9
+
10
+ try:
11
+ import pydantic
12
+ if pydantic.__version__.startswith("1."):
13
+ raise ImportError
14
+ import pydantic.v1 as pydantic # type: ignore
15
+ except ImportError:
16
+ import pydantic # type: ignore
17
+
18
+
19
+ class ChatApp(pydantic.BaseModel):
20
+ """
21
+ Schema for a chat app
22
+ """
23
+
24
+ id: str
25
+ name: str
26
+ pipeline_id: str
27
+ project_id: str
28
+ llm_config: LlmParameters
29
+ retrieval_config: PresetRetrievalParams
30
+ created_at: dt.datetime
31
+ updated_at: dt.datetime
32
+
33
+ def json(self, **kwargs: typing.Any) -> str:
34
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
35
+ return super().json(**kwargs_with_defaults)
36
+
37
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
38
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
39
+ return super().dict(**kwargs_with_defaults)
40
+
41
+ class Config:
42
+ frozen = True
43
+ smart_union = True
44
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -0,0 +1,41 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+ from .llm_parameters import LlmParameters
8
+ from .preset_retrieval_params import PresetRetrievalParams
9
+
10
+ try:
11
+ import pydantic
12
+ if pydantic.__version__.startswith("1."):
13
+ raise ImportError
14
+ import pydantic.v1 as pydantic # type: ignore
15
+ except ImportError:
16
+ import pydantic # type: ignore
17
+
18
+
19
+ class ChatAppResponse(pydantic.BaseModel):
20
+ id: str
21
+ name: str
22
+ pipeline_id: str
23
+ project_id: str
24
+ llm_config: LlmParameters
25
+ retrieval_config: PresetRetrievalParams
26
+ created_at: dt.datetime
27
+ updated_at: dt.datetime
28
+ pipeline_name: str
29
+
30
+ def json(self, **kwargs: typing.Any) -> str:
31
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
32
+ return super().json(**kwargs_with_defaults)
33
+
34
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
35
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
36
+ return super().dict(**kwargs_with_defaults)
37
+
38
+ class Config:
39
+ frozen = True
40
+ smart_union = True
41
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -18,21 +18,13 @@ class CloudAzStorageBlobDataSource(pydantic.BaseModel):
18
18
  supports_access_control: typing.Optional[bool]
19
19
  container_name: str = pydantic.Field(description="The name of the Azure Storage Blob container to read from.")
20
20
  account_url: str = pydantic.Field(description="The Azure Storage Blob account URL to use for authentication.")
21
- blob: typing.Optional[str] = pydantic.Field(description="The blob name to read from.")
22
- prefix: typing.Optional[str] = pydantic.Field(
23
- description="The prefix of the Azure Storage Blob objects to read from."
24
- )
25
- account_name: typing.Optional[str] = pydantic.Field(
26
- description="The Azure Storage Blob account name to use for authentication."
27
- )
28
- account_key: typing.Optional[str] = pydantic.Field(
29
- description="The Azure Storage Blob account key to use for authentication."
30
- )
31
- tenant_id: typing.Optional[str] = pydantic.Field(description="The Azure AD tenant ID to use for authentication.")
32
- client_id: typing.Optional[str] = pydantic.Field(description="The Azure AD client ID to use for authentication.")
33
- client_secret: typing.Optional[str] = pydantic.Field(
34
- description="The Azure AD client secret to use for authentication."
35
- )
21
+ blob: typing.Optional[str]
22
+ prefix: typing.Optional[str]
23
+ account_name: typing.Optional[str]
24
+ account_key: typing.Optional[str]
25
+ tenant_id: typing.Optional[str]
26
+ client_id: typing.Optional[str]
27
+ client_secret: typing.Optional[str]
36
28
  class_name: typing.Optional[str]
37
29
 
38
30
  def json(self, **kwargs: typing.Any) -> str:
@@ -17,21 +17,15 @@ except ImportError:
17
17
 
18
18
  class CloudBoxDataSource(pydantic.BaseModel):
19
19
  supports_access_control: typing.Optional[bool]
20
- folder_id: typing.Optional[str] = pydantic.Field(description="The ID of the Box folder to read from.")
20
+ folder_id: typing.Optional[str]
21
21
  authentication_mechanism: BoxAuthMechanism = pydantic.Field(
22
22
  description="The type of authentication to use (Developer Token or CCG)"
23
23
  )
24
- developer_token: typing.Optional[str] = pydantic.Field(
25
- description="Developer token for authentication if authentication_mechanism is 'developer_token'."
26
- )
27
- client_id: typing.Optional[str] = pydantic.Field(
28
- description="Box API key used for identifying the application the user is authenticating with"
29
- )
30
- client_secret: typing.Optional[str] = pydantic.Field(description="Box API secret used for making auth requests.")
31
- user_id: typing.Optional[str] = pydantic.Field(description="Box User ID, if provided authenticates as user.")
32
- enterprise_id: typing.Optional[str] = pydantic.Field(
33
- description="Box Enterprise ID, if provided authenticates as service."
34
- )
24
+ developer_token: typing.Optional[str]
25
+ client_id: typing.Optional[str]
26
+ client_secret: typing.Optional[str]
27
+ user_id: typing.Optional[str]
28
+ enterprise_id: typing.Optional[str]
35
29
  class_name: typing.Optional[str]
36
30
 
37
31
  def json(self, **kwargs: typing.Any) -> str:
@@ -20,12 +20,12 @@ class CloudConfluenceDataSource(pydantic.BaseModel):
20
20
  authentication_mechanism: str = pydantic.Field(
21
21
  description="Type of Authentication for connecting to Confluence APIs."
22
22
  )
23
- user_name: typing.Optional[str] = pydantic.Field(description="The username to use for authentication.")
24
- api_token: typing.Optional[str] = pydantic.Field(description="The API token to use for authentication.")
25
- space_key: typing.Optional[str] = pydantic.Field(description="The space key to read from.")
26
- page_ids: typing.Optional[str] = pydantic.Field(description="The page IDs of the Confluence to read from.")
27
- cql: typing.Optional[str] = pydantic.Field(description="The CQL query to use for fetching pages.")
28
- label: typing.Optional[str] = pydantic.Field(description="The label to use for fetching pages.")
23
+ user_name: typing.Optional[str]
24
+ api_token: typing.Optional[str]
25
+ space_key: typing.Optional[str]
26
+ page_ids: typing.Optional[str]
27
+ cql: typing.Optional[str]
28
+ label: typing.Optional[str]
29
29
  class_name: typing.Optional[str]
30
30
 
31
31
  def json(self, **kwargs: typing.Any) -> str:
@@ -23,9 +23,7 @@ class CloudDocument(pydantic.BaseModel):
23
23
  metadata: typing.Dict[str, typing.Any]
24
24
  excluded_embed_metadata_keys: typing.Optional[typing.List[str]]
25
25
  excluded_llm_metadata_keys: typing.Optional[typing.List[str]]
26
- page_positions: typing.Optional[typing.List[int]] = pydantic.Field(
27
- description="indices in the CloudDocument.text where a new page begins. e.g. Second page starts at index specified by page_positions[1]."
28
- )
26
+ page_positions: typing.Optional[typing.List[int]]
29
27
  id: str
30
28
 
31
29
  def json(self, **kwargs: typing.Any) -> str:
@@ -23,9 +23,7 @@ class CloudDocumentCreate(pydantic.BaseModel):
23
23
  metadata: typing.Dict[str, typing.Any]
24
24
  excluded_embed_metadata_keys: typing.Optional[typing.List[str]]
25
25
  excluded_llm_metadata_keys: typing.Optional[typing.List[str]]
26
- page_positions: typing.Optional[typing.List[int]] = pydantic.Field(
27
- description="indices in the CloudDocument.text where a new page begins. e.g. Second page starts at index specified by page_positions[1]."
28
- )
26
+ page_positions: typing.Optional[typing.List[int]]
29
27
  id: typing.Optional[str]
30
28
 
31
29
  def json(self, **kwargs: typing.Any) -> str: