llama-cloud 0.1.24__py3-none-any.whl → 0.1.26__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +62 -2
- llama_cloud/resources/__init__.py +18 -1
- llama_cloud/resources/data_sources/__init__.py +2 -2
- llama_cloud/resources/data_sources/client.py +5 -5
- llama_cloud/resources/data_sources/types/__init__.py +2 -1
- llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py +7 -0
- llama_cloud/resources/files/__init__.py +3 -0
- llama_cloud/resources/files/client.py +15 -13
- llama_cloud/resources/files/types/__init__.py +7 -0
- llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py +7 -0
- llama_cloud/resources/files/types/file_create_permission_info_value.py +7 -0
- llama_cloud/resources/files/types/file_create_resource_info_value.py +5 -0
- llama_cloud/resources/llama_extract/__init__.py +8 -0
- llama_cloud/resources/llama_extract/types/__init__.py +8 -0
- llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py +4 -2
- llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py +7 -0
- llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema.py +4 -2
- llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema_zero_value.py +7 -0
- llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_override.py +4 -2
- llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_override_zero_value.py +7 -0
- llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema.py +4 -2
- llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema_zero_value.py +7 -0
- llama_cloud/resources/pipelines/__init__.py +4 -0
- llama_cloud/resources/pipelines/client.py +20 -9
- llama_cloud/resources/pipelines/types/__init__.py +4 -0
- llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py +7 -0
- llama_cloud/resources/pipelines/types/retrieval_params_search_filters_inference_schema_value.py +7 -0
- llama_cloud/types/__init__.py +44 -2
- llama_cloud/types/data_source.py +2 -2
- llama_cloud/types/data_source_create.py +2 -2
- llama_cloud/types/data_source_create_custom_metadata_value.py +7 -0
- llama_cloud/types/data_source_custom_metadata_value.py +7 -0
- llama_cloud/types/extract_agent.py +4 -2
- llama_cloud/types/extract_agent_data_schema_value.py +5 -0
- llama_cloud/types/extract_config.py +3 -0
- llama_cloud/types/extract_job_create_data_schema_override.py +4 -2
- llama_cloud/types/extract_job_create_data_schema_override_zero_value.py +7 -0
- llama_cloud/types/extract_resultset.py +2 -2
- llama_cloud/types/extract_resultset_data.py +4 -2
- llama_cloud/types/extract_resultset_data_item_value.py +7 -0
- llama_cloud/types/extract_resultset_data_zero_value.py +7 -0
- llama_cloud/types/extract_resultset_extraction_metadata_value.py +7 -0
- llama_cloud/types/extract_run.py +4 -3
- llama_cloud/types/extract_run_data.py +4 -2
- llama_cloud/types/extract_run_data_item_value.py +5 -0
- llama_cloud/types/extract_run_data_schema_value.py +5 -0
- llama_cloud/types/extract_run_data_zero_value.py +5 -0
- llama_cloud/types/extract_run_extraction_metadata_value.py +7 -0
- llama_cloud/types/extract_schema_validate_response.py +2 -2
- llama_cloud/types/extract_schema_validate_response_data_schema_value.py +7 -0
- llama_cloud/types/file.py +4 -3
- llama_cloud/types/file_permission_info_value.py +5 -0
- llama_cloud/types/file_resource_info_value.py +5 -0
- llama_cloud/types/metadata_filter.py +1 -1
- llama_cloud/types/pipeline_data_source.py +2 -2
- llama_cloud/types/pipeline_data_source_custom_metadata_value.py +7 -0
- llama_cloud/types/pipeline_file.py +8 -5
- llama_cloud/types/pipeline_file_config_hash_value.py +5 -0
- llama_cloud/types/pipeline_file_create.py +2 -2
- llama_cloud/types/pipeline_file_create_custom_metadata_value.py +7 -0
- llama_cloud/types/pipeline_file_custom_metadata_value.py +7 -0
- llama_cloud/types/pipeline_file_permission_info_value.py +7 -0
- llama_cloud/types/pipeline_file_resource_info_value.py +7 -0
- llama_cloud/types/preset_retrieval_params.py +6 -2
- llama_cloud/types/preset_retrieval_params_search_filters_inference_schema_value.py +7 -0
- {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.26.dist-info}/METADATA +3 -2
- {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.26.dist-info}/RECORD +69 -38
- {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.26.dist-info}/WHEEL +1 -1
- llama_cloud/types/json_type.py +0 -9
- {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.26.dist-info}/LICENSE +0 -0
|
@@ -17,7 +17,6 @@ from ...types.eval_execution_params import EvalExecutionParams
|
|
|
17
17
|
from ...types.file_count_by_status_response import FileCountByStatusResponse
|
|
18
18
|
from ...types.http_validation_error import HttpValidationError
|
|
19
19
|
from ...types.input_message import InputMessage
|
|
20
|
-
from ...types.json_type import JsonType
|
|
21
20
|
from ...types.llama_parse_parameters import LlamaParseParameters
|
|
22
21
|
from ...types.managed_ingestion_status_response import ManagedIngestionStatusResponse
|
|
23
22
|
from ...types.metadata_filters import MetadataFilters
|
|
@@ -37,8 +36,12 @@ from ...types.preset_retrieval_params import PresetRetrievalParams
|
|
|
37
36
|
from ...types.retrieval_mode import RetrievalMode
|
|
38
37
|
from ...types.retrieve_results import RetrieveResults
|
|
39
38
|
from ...types.text_node import TextNode
|
|
39
|
+
from .types.pipeline_file_update_custom_metadata_value import PipelineFileUpdateCustomMetadataValue
|
|
40
40
|
from .types.pipeline_update_embedding_config import PipelineUpdateEmbeddingConfig
|
|
41
41
|
from .types.pipeline_update_transform_config import PipelineUpdateTransformConfig
|
|
42
|
+
from .types.retrieval_params_search_filters_inference_schema_value import (
|
|
43
|
+
RetrievalParamsSearchFiltersInferenceSchemaValue,
|
|
44
|
+
)
|
|
42
45
|
|
|
43
46
|
try:
|
|
44
47
|
import pydantic
|
|
@@ -662,7 +665,9 @@ class PipelinesClient:
|
|
|
662
665
|
file_id: str,
|
|
663
666
|
pipeline_id: str,
|
|
664
667
|
*,
|
|
665
|
-
custom_metadata: typing.Optional[
|
|
668
|
+
custom_metadata: typing.Optional[
|
|
669
|
+
typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]
|
|
670
|
+
] = OMIT,
|
|
666
671
|
) -> PipelineFile:
|
|
667
672
|
"""
|
|
668
673
|
Update a file for a pipeline.
|
|
@@ -672,7 +677,7 @@ class PipelinesClient:
|
|
|
672
677
|
|
|
673
678
|
- pipeline_id: str.
|
|
674
679
|
|
|
675
|
-
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[
|
|
680
|
+
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]].
|
|
676
681
|
---
|
|
677
682
|
from llama_cloud.client import LlamaCloud
|
|
678
683
|
|
|
@@ -1039,7 +1044,9 @@ class PipelinesClient:
|
|
|
1039
1044
|
rerank_top_n: typing.Optional[int] = OMIT,
|
|
1040
1045
|
alpha: typing.Optional[float] = OMIT,
|
|
1041
1046
|
search_filters: typing.Optional[MetadataFilters] = OMIT,
|
|
1042
|
-
search_filters_inference_schema: typing.Optional[
|
|
1047
|
+
search_filters_inference_schema: typing.Optional[
|
|
1048
|
+
typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]
|
|
1049
|
+
] = OMIT,
|
|
1043
1050
|
files_top_k: typing.Optional[int] = OMIT,
|
|
1044
1051
|
retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
|
|
1045
1052
|
retrieve_image_nodes: typing.Optional[bool] = OMIT,
|
|
@@ -1072,7 +1079,7 @@ class PipelinesClient:
|
|
|
1072
1079
|
|
|
1073
1080
|
- search_filters: typing.Optional[MetadataFilters].
|
|
1074
1081
|
|
|
1075
|
-
- search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[
|
|
1082
|
+
- search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]].
|
|
1076
1083
|
|
|
1077
1084
|
- files_top_k: typing.Optional[int].
|
|
1078
1085
|
|
|
@@ -2297,7 +2304,9 @@ class AsyncPipelinesClient:
|
|
|
2297
2304
|
file_id: str,
|
|
2298
2305
|
pipeline_id: str,
|
|
2299
2306
|
*,
|
|
2300
|
-
custom_metadata: typing.Optional[
|
|
2307
|
+
custom_metadata: typing.Optional[
|
|
2308
|
+
typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]
|
|
2309
|
+
] = OMIT,
|
|
2301
2310
|
) -> PipelineFile:
|
|
2302
2311
|
"""
|
|
2303
2312
|
Update a file for a pipeline.
|
|
@@ -2307,7 +2316,7 @@ class AsyncPipelinesClient:
|
|
|
2307
2316
|
|
|
2308
2317
|
- pipeline_id: str.
|
|
2309
2318
|
|
|
2310
|
-
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[
|
|
2319
|
+
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]].
|
|
2311
2320
|
---
|
|
2312
2321
|
from llama_cloud.client import AsyncLlamaCloud
|
|
2313
2322
|
|
|
@@ -2676,7 +2685,9 @@ class AsyncPipelinesClient:
|
|
|
2676
2685
|
rerank_top_n: typing.Optional[int] = OMIT,
|
|
2677
2686
|
alpha: typing.Optional[float] = OMIT,
|
|
2678
2687
|
search_filters: typing.Optional[MetadataFilters] = OMIT,
|
|
2679
|
-
search_filters_inference_schema: typing.Optional[
|
|
2688
|
+
search_filters_inference_schema: typing.Optional[
|
|
2689
|
+
typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]
|
|
2690
|
+
] = OMIT,
|
|
2680
2691
|
files_top_k: typing.Optional[int] = OMIT,
|
|
2681
2692
|
retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
|
|
2682
2693
|
retrieve_image_nodes: typing.Optional[bool] = OMIT,
|
|
@@ -2709,7 +2720,7 @@ class AsyncPipelinesClient:
|
|
|
2709
2720
|
|
|
2710
2721
|
- search_filters: typing.Optional[MetadataFilters].
|
|
2711
2722
|
|
|
2712
|
-
- search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[
|
|
2723
|
+
- search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]].
|
|
2713
2724
|
|
|
2714
2725
|
- files_top_k: typing.Optional[int].
|
|
2715
2726
|
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
|
2
2
|
|
|
3
|
+
from .pipeline_file_update_custom_metadata_value import PipelineFileUpdateCustomMetadataValue
|
|
3
4
|
from .pipeline_update_embedding_config import (
|
|
4
5
|
PipelineUpdateEmbeddingConfig,
|
|
5
6
|
PipelineUpdateEmbeddingConfig_AzureEmbedding,
|
|
@@ -11,8 +12,10 @@ from .pipeline_update_embedding_config import (
|
|
|
11
12
|
PipelineUpdateEmbeddingConfig_VertexaiEmbedding,
|
|
12
13
|
)
|
|
13
14
|
from .pipeline_update_transform_config import PipelineUpdateTransformConfig
|
|
15
|
+
from .retrieval_params_search_filters_inference_schema_value import RetrievalParamsSearchFiltersInferenceSchemaValue
|
|
14
16
|
|
|
15
17
|
__all__ = [
|
|
18
|
+
"PipelineFileUpdateCustomMetadataValue",
|
|
16
19
|
"PipelineUpdateEmbeddingConfig",
|
|
17
20
|
"PipelineUpdateEmbeddingConfig_AzureEmbedding",
|
|
18
21
|
"PipelineUpdateEmbeddingConfig_BedrockEmbedding",
|
|
@@ -22,4 +25,5 @@ __all__ = [
|
|
|
22
25
|
"PipelineUpdateEmbeddingConfig_OpenaiEmbedding",
|
|
23
26
|
"PipelineUpdateEmbeddingConfig_VertexaiEmbedding",
|
|
24
27
|
"PipelineUpdateTransformConfig",
|
|
28
|
+
"RetrievalParamsSearchFiltersInferenceSchemaValue",
|
|
25
29
|
]
|
llama_cloud/types/__init__.py
CHANGED
|
@@ -72,6 +72,8 @@ from .data_source import DataSource
|
|
|
72
72
|
from .data_source_component import DataSourceComponent
|
|
73
73
|
from .data_source_create import DataSourceCreate
|
|
74
74
|
from .data_source_create_component import DataSourceCreateComponent
|
|
75
|
+
from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
|
|
76
|
+
from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
|
|
75
77
|
from .data_source_update_dispatcher_config import DataSourceUpdateDispatcherConfig
|
|
76
78
|
from .delete_params import DeleteParams
|
|
77
79
|
from .document_block import DocumentBlock
|
|
@@ -104,17 +106,27 @@ from .embedding_model_config_update_embedding_config import (
|
|
|
104
106
|
)
|
|
105
107
|
from .eval_execution_params import EvalExecutionParams
|
|
106
108
|
from .extract_agent import ExtractAgent
|
|
109
|
+
from .extract_agent_data_schema_value import ExtractAgentDataSchemaValue
|
|
107
110
|
from .extract_config import ExtractConfig
|
|
108
111
|
from .extract_job import ExtractJob
|
|
109
112
|
from .extract_job_create import ExtractJobCreate
|
|
110
113
|
from .extract_job_create_data_schema_override import ExtractJobCreateDataSchemaOverride
|
|
114
|
+
from .extract_job_create_data_schema_override_zero_value import ExtractJobCreateDataSchemaOverrideZeroValue
|
|
111
115
|
from .extract_mode import ExtractMode
|
|
112
116
|
from .extract_models import ExtractModels
|
|
113
117
|
from .extract_resultset import ExtractResultset
|
|
114
118
|
from .extract_resultset_data import ExtractResultsetData
|
|
119
|
+
from .extract_resultset_data_item_value import ExtractResultsetDataItemValue
|
|
120
|
+
from .extract_resultset_data_zero_value import ExtractResultsetDataZeroValue
|
|
121
|
+
from .extract_resultset_extraction_metadata_value import ExtractResultsetExtractionMetadataValue
|
|
115
122
|
from .extract_run import ExtractRun
|
|
116
123
|
from .extract_run_data import ExtractRunData
|
|
124
|
+
from .extract_run_data_item_value import ExtractRunDataItemValue
|
|
125
|
+
from .extract_run_data_schema_value import ExtractRunDataSchemaValue
|
|
126
|
+
from .extract_run_data_zero_value import ExtractRunDataZeroValue
|
|
127
|
+
from .extract_run_extraction_metadata_value import ExtractRunExtractionMetadataValue
|
|
117
128
|
from .extract_schema_validate_response import ExtractSchemaValidateResponse
|
|
129
|
+
from .extract_schema_validate_response_data_schema_value import ExtractSchemaValidateResponseDataSchemaValue
|
|
118
130
|
from .extract_state import ExtractState
|
|
119
131
|
from .extract_target import ExtractTarget
|
|
120
132
|
from .fail_page_mode import FailPageMode
|
|
@@ -122,6 +134,8 @@ from .file import File
|
|
|
122
134
|
from .file_count_by_status_response import FileCountByStatusResponse
|
|
123
135
|
from .file_id_presigned_url import FileIdPresignedUrl
|
|
124
136
|
from .file_parse_public import FileParsePublic
|
|
137
|
+
from .file_permission_info_value import FilePermissionInfoValue
|
|
138
|
+
from .file_resource_info_value import FileResourceInfoValue
|
|
125
139
|
from .filter_condition import FilterCondition
|
|
126
140
|
from .filter_operator import FilterOperator
|
|
127
141
|
from .free_credits_usage import FreeCreditsUsage
|
|
@@ -150,7 +164,6 @@ from .job_record_parameters import (
|
|
|
150
164
|
JobRecordParameters_PipelineManagedIngestion,
|
|
151
165
|
)
|
|
152
166
|
from .job_record_with_usage_metrics import JobRecordWithUsageMetrics
|
|
153
|
-
from .json_type import JsonType
|
|
154
167
|
from .l_lama_parse_transform_config import LLamaParseTransformConfig
|
|
155
168
|
from .legacy_parse_job_config import LegacyParseJobConfig
|
|
156
169
|
from .llama_extract_settings import LlamaExtractSettings
|
|
@@ -225,6 +238,7 @@ from .pipeline_create_transform_config import PipelineCreateTransformConfig
|
|
|
225
238
|
from .pipeline_data_source import PipelineDataSource
|
|
226
239
|
from .pipeline_data_source_component import PipelineDataSourceComponent
|
|
227
240
|
from .pipeline_data_source_create import PipelineDataSourceCreate
|
|
241
|
+
from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
|
|
228
242
|
from .pipeline_data_source_status import PipelineDataSourceStatus
|
|
229
243
|
from .pipeline_deployment import PipelineDeployment
|
|
230
244
|
from .pipeline_embedding_config import (
|
|
@@ -238,7 +252,12 @@ from .pipeline_embedding_config import (
|
|
|
238
252
|
PipelineEmbeddingConfig_VertexaiEmbedding,
|
|
239
253
|
)
|
|
240
254
|
from .pipeline_file import PipelineFile
|
|
255
|
+
from .pipeline_file_config_hash_value import PipelineFileConfigHashValue
|
|
241
256
|
from .pipeline_file_create import PipelineFileCreate
|
|
257
|
+
from .pipeline_file_create_custom_metadata_value import PipelineFileCreateCustomMetadataValue
|
|
258
|
+
from .pipeline_file_custom_metadata_value import PipelineFileCustomMetadataValue
|
|
259
|
+
from .pipeline_file_permission_info_value import PipelineFilePermissionInfoValue
|
|
260
|
+
from .pipeline_file_resource_info_value import PipelineFileResourceInfoValue
|
|
242
261
|
from .pipeline_file_status import PipelineFileStatus
|
|
243
262
|
from .pipeline_file_update_dispatcher_config import PipelineFileUpdateDispatcherConfig
|
|
244
263
|
from .pipeline_file_updater_config import PipelineFileUpdaterConfig
|
|
@@ -256,6 +275,9 @@ from .playground_session import PlaygroundSession
|
|
|
256
275
|
from .pooling import Pooling
|
|
257
276
|
from .preset_composite_retrieval_params import PresetCompositeRetrievalParams
|
|
258
277
|
from .preset_retrieval_params import PresetRetrievalParams
|
|
278
|
+
from .preset_retrieval_params_search_filters_inference_schema_value import (
|
|
279
|
+
PresetRetrievalParamsSearchFiltersInferenceSchemaValue,
|
|
280
|
+
)
|
|
259
281
|
from .presigned_url import PresignedUrl
|
|
260
282
|
from .progress_event import ProgressEvent
|
|
261
283
|
from .progress_event_status import ProgressEventStatus
|
|
@@ -390,6 +412,8 @@ __all__ = [
|
|
|
390
412
|
"DataSourceComponent",
|
|
391
413
|
"DataSourceCreate",
|
|
392
414
|
"DataSourceCreateComponent",
|
|
415
|
+
"DataSourceCreateCustomMetadataValue",
|
|
416
|
+
"DataSourceCustomMetadataValue",
|
|
393
417
|
"DataSourceUpdateDispatcherConfig",
|
|
394
418
|
"DeleteParams",
|
|
395
419
|
"DocumentBlock",
|
|
@@ -418,17 +442,27 @@ __all__ = [
|
|
|
418
442
|
"EmbeddingModelConfigUpdateEmbeddingConfig_VertexaiEmbedding",
|
|
419
443
|
"EvalExecutionParams",
|
|
420
444
|
"ExtractAgent",
|
|
445
|
+
"ExtractAgentDataSchemaValue",
|
|
421
446
|
"ExtractConfig",
|
|
422
447
|
"ExtractJob",
|
|
423
448
|
"ExtractJobCreate",
|
|
424
449
|
"ExtractJobCreateDataSchemaOverride",
|
|
450
|
+
"ExtractJobCreateDataSchemaOverrideZeroValue",
|
|
425
451
|
"ExtractMode",
|
|
426
452
|
"ExtractModels",
|
|
427
453
|
"ExtractResultset",
|
|
428
454
|
"ExtractResultsetData",
|
|
455
|
+
"ExtractResultsetDataItemValue",
|
|
456
|
+
"ExtractResultsetDataZeroValue",
|
|
457
|
+
"ExtractResultsetExtractionMetadataValue",
|
|
429
458
|
"ExtractRun",
|
|
430
459
|
"ExtractRunData",
|
|
460
|
+
"ExtractRunDataItemValue",
|
|
461
|
+
"ExtractRunDataSchemaValue",
|
|
462
|
+
"ExtractRunDataZeroValue",
|
|
463
|
+
"ExtractRunExtractionMetadataValue",
|
|
431
464
|
"ExtractSchemaValidateResponse",
|
|
465
|
+
"ExtractSchemaValidateResponseDataSchemaValue",
|
|
432
466
|
"ExtractState",
|
|
433
467
|
"ExtractTarget",
|
|
434
468
|
"FailPageMode",
|
|
@@ -436,6 +470,8 @@ __all__ = [
|
|
|
436
470
|
"FileCountByStatusResponse",
|
|
437
471
|
"FileIdPresignedUrl",
|
|
438
472
|
"FileParsePublic",
|
|
473
|
+
"FilePermissionInfoValue",
|
|
474
|
+
"FileResourceInfoValue",
|
|
439
475
|
"FilterCondition",
|
|
440
476
|
"FilterOperator",
|
|
441
477
|
"FreeCreditsUsage",
|
|
@@ -462,7 +498,6 @@ __all__ = [
|
|
|
462
498
|
"JobRecordParameters_PipelineFileUpdater",
|
|
463
499
|
"JobRecordParameters_PipelineManagedIngestion",
|
|
464
500
|
"JobRecordWithUsageMetrics",
|
|
465
|
-
"JsonType",
|
|
466
501
|
"LLamaParseTransformConfig",
|
|
467
502
|
"LegacyParseJobConfig",
|
|
468
503
|
"LlamaExtractSettings",
|
|
@@ -533,6 +568,7 @@ __all__ = [
|
|
|
533
568
|
"PipelineDataSource",
|
|
534
569
|
"PipelineDataSourceComponent",
|
|
535
570
|
"PipelineDataSourceCreate",
|
|
571
|
+
"PipelineDataSourceCustomMetadataValue",
|
|
536
572
|
"PipelineDataSourceStatus",
|
|
537
573
|
"PipelineDeployment",
|
|
538
574
|
"PipelineEmbeddingConfig",
|
|
@@ -544,7 +580,12 @@ __all__ = [
|
|
|
544
580
|
"PipelineEmbeddingConfig_OpenaiEmbedding",
|
|
545
581
|
"PipelineEmbeddingConfig_VertexaiEmbedding",
|
|
546
582
|
"PipelineFile",
|
|
583
|
+
"PipelineFileConfigHashValue",
|
|
547
584
|
"PipelineFileCreate",
|
|
585
|
+
"PipelineFileCreateCustomMetadataValue",
|
|
586
|
+
"PipelineFileCustomMetadataValue",
|
|
587
|
+
"PipelineFilePermissionInfoValue",
|
|
588
|
+
"PipelineFileResourceInfoValue",
|
|
548
589
|
"PipelineFileStatus",
|
|
549
590
|
"PipelineFileUpdateDispatcherConfig",
|
|
550
591
|
"PipelineFileUpdaterConfig",
|
|
@@ -560,6 +601,7 @@ __all__ = [
|
|
|
560
601
|
"Pooling",
|
|
561
602
|
"PresetCompositeRetrievalParams",
|
|
562
603
|
"PresetRetrievalParams",
|
|
604
|
+
"PresetRetrievalParamsSearchFiltersInferenceSchemaValue",
|
|
563
605
|
"PresignedUrl",
|
|
564
606
|
"ProgressEvent",
|
|
565
607
|
"ProgressEventStatus",
|
llama_cloud/types/data_source.py
CHANGED
|
@@ -6,7 +6,7 @@ import typing
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .configurable_data_source_names import ConfigurableDataSourceNames
|
|
8
8
|
from .data_source_component import DataSourceComponent
|
|
9
|
-
from .
|
|
9
|
+
from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
|
|
10
10
|
|
|
11
11
|
try:
|
|
12
12
|
import pydantic
|
|
@@ -27,7 +27,7 @@ class DataSource(pydantic.BaseModel):
|
|
|
27
27
|
updated_at: typing.Optional[dt.datetime]
|
|
28
28
|
name: str = pydantic.Field(description="The name of the data source.")
|
|
29
29
|
source_type: ConfigurableDataSourceNames
|
|
30
|
-
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[
|
|
30
|
+
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceCustomMetadataValue]]]
|
|
31
31
|
component: DataSourceComponent = pydantic.Field(description="Component that implements the data source")
|
|
32
32
|
version_metadata: typing.Optional[typing.Dict[str, typing.Any]]
|
|
33
33
|
project_id: str
|
|
@@ -6,7 +6,7 @@ import typing
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .configurable_data_source_names import ConfigurableDataSourceNames
|
|
8
8
|
from .data_source_create_component import DataSourceCreateComponent
|
|
9
|
-
from .
|
|
9
|
+
from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
|
|
10
10
|
|
|
11
11
|
try:
|
|
12
12
|
import pydantic
|
|
@@ -24,7 +24,7 @@ class DataSourceCreate(pydantic.BaseModel):
|
|
|
24
24
|
|
|
25
25
|
name: str = pydantic.Field(description="The name of the data source.")
|
|
26
26
|
source_type: ConfigurableDataSourceNames
|
|
27
|
-
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[
|
|
27
|
+
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceCreateCustomMetadataValue]]]
|
|
28
28
|
component: DataSourceCreateComponent = pydantic.Field(description="Component that implements the data source")
|
|
29
29
|
|
|
30
30
|
def json(self, **kwargs: typing.Any) -> str:
|
|
@@ -4,8 +4,8 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
+
from .extract_agent_data_schema_value import ExtractAgentDataSchemaValue
|
|
7
8
|
from .extract_config import ExtractConfig
|
|
8
|
-
from .json_type import JsonType
|
|
9
9
|
|
|
10
10
|
try:
|
|
11
11
|
import pydantic
|
|
@@ -24,7 +24,9 @@ class ExtractAgent(pydantic.BaseModel):
|
|
|
24
24
|
id: str = pydantic.Field(description="The id of the extraction agent.")
|
|
25
25
|
name: str = pydantic.Field(description="The name of the extraction agent.")
|
|
26
26
|
project_id: str = pydantic.Field(description="The ID of the project that the extraction agent belongs to.")
|
|
27
|
-
data_schema: typing.Dict[str, typing.Optional[
|
|
27
|
+
data_schema: typing.Dict[str, typing.Optional[ExtractAgentDataSchemaValue]] = pydantic.Field(
|
|
28
|
+
description="The schema of the data."
|
|
29
|
+
)
|
|
28
30
|
config: ExtractConfig = pydantic.Field(description="The configuration parameters for the extraction agent.")
|
|
29
31
|
created_at: typing.Optional[dt.datetime]
|
|
30
32
|
updated_at: typing.Optional[dt.datetime]
|
|
@@ -24,6 +24,9 @@ class ExtractConfig(pydantic.BaseModel):
|
|
|
24
24
|
|
|
25
25
|
extraction_target: typing.Optional[ExtractTarget] = pydantic.Field(description="The extraction target specified.")
|
|
26
26
|
extraction_mode: typing.Optional[ExtractMode] = pydantic.Field(description="The extraction mode specified.")
|
|
27
|
+
multimodal_fast_mode: typing.Optional[bool] = pydantic.Field(
|
|
28
|
+
description="Whether to use fast mode for multimodal extraction."
|
|
29
|
+
)
|
|
27
30
|
system_prompt: typing.Optional[str]
|
|
28
31
|
use_reasoning: typing.Optional[bool] = pydantic.Field(description="Whether to use reasoning for the extraction.")
|
|
29
32
|
cite_sources: typing.Optional[bool] = pydantic.Field(description="Whether to cite sources for the extraction.")
|
|
@@ -2,6 +2,8 @@
|
|
|
2
2
|
|
|
3
3
|
import typing
|
|
4
4
|
|
|
5
|
-
from .
|
|
5
|
+
from .extract_job_create_data_schema_override_zero_value import ExtractJobCreateDataSchemaOverrideZeroValue
|
|
6
6
|
|
|
7
|
-
ExtractJobCreateDataSchemaOverride = typing.Union[
|
|
7
|
+
ExtractJobCreateDataSchemaOverride = typing.Union[
|
|
8
|
+
typing.Dict[str, typing.Optional[ExtractJobCreateDataSchemaOverrideZeroValue]], str
|
|
9
|
+
]
|
|
@@ -5,7 +5,7 @@ import typing
|
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .extract_resultset_data import ExtractResultsetData
|
|
8
|
-
from .
|
|
8
|
+
from .extract_resultset_extraction_metadata_value import ExtractResultsetExtractionMetadataValue
|
|
9
9
|
|
|
10
10
|
try:
|
|
11
11
|
import pydantic
|
|
@@ -24,7 +24,7 @@ class ExtractResultset(pydantic.BaseModel):
|
|
|
24
24
|
run_id: str = pydantic.Field(description="The id of the extraction run")
|
|
25
25
|
extraction_agent_id: str = pydantic.Field(description="The id of the extraction agent")
|
|
26
26
|
data: typing.Optional[ExtractResultsetData] = pydantic.Field(description="The data extracted from the file")
|
|
27
|
-
extraction_metadata: typing.Dict[str, typing.Optional[
|
|
27
|
+
extraction_metadata: typing.Dict[str, typing.Optional[ExtractResultsetExtractionMetadataValue]] = pydantic.Field(
|
|
28
28
|
description="The metadata extracted from the file"
|
|
29
29
|
)
|
|
30
30
|
|
|
@@ -2,8 +2,10 @@
|
|
|
2
2
|
|
|
3
3
|
import typing
|
|
4
4
|
|
|
5
|
-
from .
|
|
5
|
+
from .extract_resultset_data_item_value import ExtractResultsetDataItemValue
|
|
6
|
+
from .extract_resultset_data_zero_value import ExtractResultsetDataZeroValue
|
|
6
7
|
|
|
7
8
|
ExtractResultsetData = typing.Union[
|
|
8
|
-
typing.Dict[str, typing.Optional[
|
|
9
|
+
typing.Dict[str, typing.Optional[ExtractResultsetDataZeroValue]],
|
|
10
|
+
typing.List[typing.Dict[str, typing.Optional[ExtractResultsetDataItemValue]]],
|
|
9
11
|
]
|
llama_cloud/types/extract_run.py
CHANGED
|
@@ -6,9 +6,10 @@ import typing
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .extract_config import ExtractConfig
|
|
8
8
|
from .extract_run_data import ExtractRunData
|
|
9
|
+
from .extract_run_data_schema_value import ExtractRunDataSchemaValue
|
|
10
|
+
from .extract_run_extraction_metadata_value import ExtractRunExtractionMetadataValue
|
|
9
11
|
from .extract_state import ExtractState
|
|
10
12
|
from .file import File
|
|
11
|
-
from .json_type import JsonType
|
|
12
13
|
|
|
13
14
|
try:
|
|
14
15
|
import pydantic
|
|
@@ -28,7 +29,7 @@ class ExtractRun(pydantic.BaseModel):
|
|
|
28
29
|
created_at: typing.Optional[dt.datetime]
|
|
29
30
|
updated_at: typing.Optional[dt.datetime]
|
|
30
31
|
extraction_agent_id: str = pydantic.Field(description="The id of the extraction agent")
|
|
31
|
-
data_schema: typing.Dict[str, typing.Optional[
|
|
32
|
+
data_schema: typing.Dict[str, typing.Optional[ExtractRunDataSchemaValue]] = pydantic.Field(
|
|
32
33
|
description="The schema used for extraction"
|
|
33
34
|
)
|
|
34
35
|
config: ExtractConfig = pydantic.Field(description="The config used for extraction")
|
|
@@ -37,7 +38,7 @@ class ExtractRun(pydantic.BaseModel):
|
|
|
37
38
|
error: typing.Optional[str]
|
|
38
39
|
job_id: typing.Optional[str]
|
|
39
40
|
data: typing.Optional[ExtractRunData] = pydantic.Field(description="The data extracted from the file")
|
|
40
|
-
extraction_metadata: typing.Optional[typing.Dict[str, typing.Optional[
|
|
41
|
+
extraction_metadata: typing.Optional[typing.Dict[str, typing.Optional[ExtractRunExtractionMetadataValue]]]
|
|
41
42
|
from_ui: bool = pydantic.Field(description="Whether this extraction run was triggered from the UI")
|
|
42
43
|
|
|
43
44
|
def json(self, **kwargs: typing.Any) -> str:
|
|
@@ -2,8 +2,10 @@
|
|
|
2
2
|
|
|
3
3
|
import typing
|
|
4
4
|
|
|
5
|
-
from .
|
|
5
|
+
from .extract_run_data_item_value import ExtractRunDataItemValue
|
|
6
|
+
from .extract_run_data_zero_value import ExtractRunDataZeroValue
|
|
6
7
|
|
|
7
8
|
ExtractRunData = typing.Union[
|
|
8
|
-
typing.Dict[str, typing.Optional[
|
|
9
|
+
typing.Dict[str, typing.Optional[ExtractRunDataZeroValue]],
|
|
10
|
+
typing.List[typing.Dict[str, typing.Optional[ExtractRunDataItemValue]]],
|
|
9
11
|
]
|
|
@@ -4,7 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .
|
|
7
|
+
from .extract_schema_validate_response_data_schema_value import ExtractSchemaValidateResponseDataSchemaValue
|
|
8
8
|
|
|
9
9
|
try:
|
|
10
10
|
import pydantic
|
|
@@ -16,7 +16,7 @@ except ImportError:
|
|
|
16
16
|
|
|
17
17
|
|
|
18
18
|
class ExtractSchemaValidateResponse(pydantic.BaseModel):
|
|
19
|
-
data_schema: typing.Dict[str, typing.Optional[
|
|
19
|
+
data_schema: typing.Dict[str, typing.Optional[ExtractSchemaValidateResponseDataSchemaValue]]
|
|
20
20
|
|
|
21
21
|
def json(self, **kwargs: typing.Any) -> str:
|
|
22
22
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
llama_cloud/types/file.py
CHANGED
|
@@ -4,7 +4,8 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .
|
|
7
|
+
from .file_permission_info_value import FilePermissionInfoValue
|
|
8
|
+
from .file_resource_info_value import FileResourceInfoValue
|
|
8
9
|
|
|
9
10
|
try:
|
|
10
11
|
import pydantic
|
|
@@ -29,8 +30,8 @@ class File(pydantic.BaseModel):
|
|
|
29
30
|
file_type: typing.Optional[str]
|
|
30
31
|
project_id: str = pydantic.Field(description="The ID of the project that the file belongs to")
|
|
31
32
|
last_modified_at: typing.Optional[dt.datetime]
|
|
32
|
-
resource_info: typing.Optional[typing.Dict[str, typing.Optional[
|
|
33
|
-
permission_info: typing.Optional[typing.Dict[str, typing.Optional[
|
|
33
|
+
resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileResourceInfoValue]]]
|
|
34
|
+
permission_info: typing.Optional[typing.Dict[str, typing.Optional[FilePermissionInfoValue]]]
|
|
34
35
|
data_source_id: typing.Optional[str]
|
|
35
36
|
|
|
36
37
|
def json(self, **kwargs: typing.Any) -> str:
|
|
@@ -20,7 +20,7 @@ class MetadataFilter(pydantic.BaseModel):
|
|
|
20
20
|
"""
|
|
21
21
|
Comprehensive metadata filter for vector stores to support more operators.
|
|
22
22
|
|
|
23
|
-
Value uses Strict
|
|
23
|
+
Value uses Strict types, as int, float and str are compatible types and were all
|
|
24
24
|
converted to string before.
|
|
25
25
|
|
|
26
26
|
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
|
|
@@ -5,8 +5,8 @@ import typing
|
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .configurable_data_source_names import ConfigurableDataSourceNames
|
|
8
|
-
from .json_type import JsonType
|
|
9
8
|
from .pipeline_data_source_component import PipelineDataSourceComponent
|
|
9
|
+
from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
|
|
10
10
|
from .pipeline_data_source_status import PipelineDataSourceStatus
|
|
11
11
|
|
|
12
12
|
try:
|
|
@@ -28,7 +28,7 @@ class PipelineDataSource(pydantic.BaseModel):
|
|
|
28
28
|
updated_at: typing.Optional[dt.datetime]
|
|
29
29
|
name: str = pydantic.Field(description="The name of the data source.")
|
|
30
30
|
source_type: ConfigurableDataSourceNames
|
|
31
|
-
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[
|
|
31
|
+
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineDataSourceCustomMetadataValue]]]
|
|
32
32
|
component: PipelineDataSourceComponent = pydantic.Field(description="Component that implements the data source")
|
|
33
33
|
version_metadata: typing.Optional[typing.Dict[str, typing.Any]]
|
|
34
34
|
project_id: str
|