llama-cloud 0.1.24__py3-none-any.whl → 0.1.25__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (69) hide show
  1. llama_cloud/__init__.py +62 -2
  2. llama_cloud/resources/__init__.py +18 -1
  3. llama_cloud/resources/data_sources/__init__.py +2 -2
  4. llama_cloud/resources/data_sources/client.py +5 -5
  5. llama_cloud/resources/data_sources/types/__init__.py +2 -1
  6. llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py +7 -0
  7. llama_cloud/resources/files/__init__.py +3 -0
  8. llama_cloud/resources/files/client.py +15 -13
  9. llama_cloud/resources/files/types/__init__.py +7 -0
  10. llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py +7 -0
  11. llama_cloud/resources/files/types/file_create_permission_info_value.py +7 -0
  12. llama_cloud/resources/files/types/file_create_resource_info_value.py +5 -0
  13. llama_cloud/resources/llama_extract/__init__.py +8 -0
  14. llama_cloud/resources/llama_extract/types/__init__.py +8 -0
  15. llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py +4 -2
  16. llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py +7 -0
  17. llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema.py +4 -2
  18. llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema_zero_value.py +7 -0
  19. llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_override.py +4 -2
  20. llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_override_zero_value.py +7 -0
  21. llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema.py +4 -2
  22. llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema_zero_value.py +7 -0
  23. llama_cloud/resources/pipelines/__init__.py +4 -0
  24. llama_cloud/resources/pipelines/client.py +20 -9
  25. llama_cloud/resources/pipelines/types/__init__.py +4 -0
  26. llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py +7 -0
  27. llama_cloud/resources/pipelines/types/retrieval_params_search_filters_inference_schema_value.py +7 -0
  28. llama_cloud/types/__init__.py +44 -2
  29. llama_cloud/types/data_source.py +2 -2
  30. llama_cloud/types/data_source_create.py +2 -2
  31. llama_cloud/types/data_source_create_custom_metadata_value.py +7 -0
  32. llama_cloud/types/data_source_custom_metadata_value.py +7 -0
  33. llama_cloud/types/extract_agent.py +4 -2
  34. llama_cloud/types/extract_agent_data_schema_value.py +5 -0
  35. llama_cloud/types/extract_config.py +3 -0
  36. llama_cloud/types/extract_job_create_data_schema_override.py +4 -2
  37. llama_cloud/types/extract_job_create_data_schema_override_zero_value.py +7 -0
  38. llama_cloud/types/extract_resultset.py +2 -2
  39. llama_cloud/types/extract_resultset_data.py +4 -2
  40. llama_cloud/types/extract_resultset_data_item_value.py +7 -0
  41. llama_cloud/types/extract_resultset_data_zero_value.py +7 -0
  42. llama_cloud/types/extract_resultset_extraction_metadata_value.py +7 -0
  43. llama_cloud/types/extract_run.py +4 -3
  44. llama_cloud/types/extract_run_data.py +4 -2
  45. llama_cloud/types/extract_run_data_item_value.py +5 -0
  46. llama_cloud/types/extract_run_data_schema_value.py +5 -0
  47. llama_cloud/types/extract_run_data_zero_value.py +5 -0
  48. llama_cloud/types/extract_run_extraction_metadata_value.py +7 -0
  49. llama_cloud/types/extract_schema_validate_response.py +2 -2
  50. llama_cloud/types/extract_schema_validate_response_data_schema_value.py +7 -0
  51. llama_cloud/types/file.py +4 -3
  52. llama_cloud/types/file_permission_info_value.py +5 -0
  53. llama_cloud/types/file_resource_info_value.py +5 -0
  54. llama_cloud/types/pipeline_data_source.py +2 -2
  55. llama_cloud/types/pipeline_data_source_custom_metadata_value.py +7 -0
  56. llama_cloud/types/pipeline_file.py +8 -5
  57. llama_cloud/types/pipeline_file_config_hash_value.py +5 -0
  58. llama_cloud/types/pipeline_file_create.py +2 -2
  59. llama_cloud/types/pipeline_file_create_custom_metadata_value.py +7 -0
  60. llama_cloud/types/pipeline_file_custom_metadata_value.py +7 -0
  61. llama_cloud/types/pipeline_file_permission_info_value.py +7 -0
  62. llama_cloud/types/pipeline_file_resource_info_value.py +7 -0
  63. llama_cloud/types/preset_retrieval_params.py +6 -2
  64. llama_cloud/types/preset_retrieval_params_search_filters_inference_schema_value.py +7 -0
  65. {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.25.dist-info}/METADATA +1 -1
  66. {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.25.dist-info}/RECORD +68 -37
  67. llama_cloud/types/json_type.py +0 -9
  68. {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.25.dist-info}/LICENSE +0 -0
  69. {llama_cloud-0.1.24.dist-info → llama_cloud-0.1.25.dist-info}/WHEEL +0 -0
@@ -17,7 +17,6 @@ from ...types.eval_execution_params import EvalExecutionParams
17
17
  from ...types.file_count_by_status_response import FileCountByStatusResponse
18
18
  from ...types.http_validation_error import HttpValidationError
19
19
  from ...types.input_message import InputMessage
20
- from ...types.json_type import JsonType
21
20
  from ...types.llama_parse_parameters import LlamaParseParameters
22
21
  from ...types.managed_ingestion_status_response import ManagedIngestionStatusResponse
23
22
  from ...types.metadata_filters import MetadataFilters
@@ -37,8 +36,12 @@ from ...types.preset_retrieval_params import PresetRetrievalParams
37
36
  from ...types.retrieval_mode import RetrievalMode
38
37
  from ...types.retrieve_results import RetrieveResults
39
38
  from ...types.text_node import TextNode
39
+ from .types.pipeline_file_update_custom_metadata_value import PipelineFileUpdateCustomMetadataValue
40
40
  from .types.pipeline_update_embedding_config import PipelineUpdateEmbeddingConfig
41
41
  from .types.pipeline_update_transform_config import PipelineUpdateTransformConfig
42
+ from .types.retrieval_params_search_filters_inference_schema_value import (
43
+ RetrievalParamsSearchFiltersInferenceSchemaValue,
44
+ )
42
45
 
43
46
  try:
44
47
  import pydantic
@@ -662,7 +665,9 @@ class PipelinesClient:
662
665
  file_id: str,
663
666
  pipeline_id: str,
664
667
  *,
665
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]] = OMIT,
668
+ custom_metadata: typing.Optional[
669
+ typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]
670
+ ] = OMIT,
666
671
  ) -> PipelineFile:
667
672
  """
668
673
  Update a file for a pipeline.
@@ -672,7 +677,7 @@ class PipelinesClient:
672
677
 
673
678
  - pipeline_id: str.
674
679
 
675
- - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]].
680
+ - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]].
676
681
  ---
677
682
  from llama_cloud.client import LlamaCloud
678
683
 
@@ -1039,7 +1044,9 @@ class PipelinesClient:
1039
1044
  rerank_top_n: typing.Optional[int] = OMIT,
1040
1045
  alpha: typing.Optional[float] = OMIT,
1041
1046
  search_filters: typing.Optional[MetadataFilters] = OMIT,
1042
- search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]] = OMIT,
1047
+ search_filters_inference_schema: typing.Optional[
1048
+ typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]
1049
+ ] = OMIT,
1043
1050
  files_top_k: typing.Optional[int] = OMIT,
1044
1051
  retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
1045
1052
  retrieve_image_nodes: typing.Optional[bool] = OMIT,
@@ -1072,7 +1079,7 @@ class PipelinesClient:
1072
1079
 
1073
1080
  - search_filters: typing.Optional[MetadataFilters].
1074
1081
 
1075
- - search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]].
1082
+ - search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]].
1076
1083
 
1077
1084
  - files_top_k: typing.Optional[int].
1078
1085
 
@@ -2297,7 +2304,9 @@ class AsyncPipelinesClient:
2297
2304
  file_id: str,
2298
2305
  pipeline_id: str,
2299
2306
  *,
2300
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]] = OMIT,
2307
+ custom_metadata: typing.Optional[
2308
+ typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]
2309
+ ] = OMIT,
2301
2310
  ) -> PipelineFile:
2302
2311
  """
2303
2312
  Update a file for a pipeline.
@@ -2307,7 +2316,7 @@ class AsyncPipelinesClient:
2307
2316
 
2308
2317
  - pipeline_id: str.
2309
2318
 
2310
- - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]].
2319
+ - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileUpdateCustomMetadataValue]]].
2311
2320
  ---
2312
2321
  from llama_cloud.client import AsyncLlamaCloud
2313
2322
 
@@ -2676,7 +2685,9 @@ class AsyncPipelinesClient:
2676
2685
  rerank_top_n: typing.Optional[int] = OMIT,
2677
2686
  alpha: typing.Optional[float] = OMIT,
2678
2687
  search_filters: typing.Optional[MetadataFilters] = OMIT,
2679
- search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]] = OMIT,
2688
+ search_filters_inference_schema: typing.Optional[
2689
+ typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]
2690
+ ] = OMIT,
2680
2691
  files_top_k: typing.Optional[int] = OMIT,
2681
2692
  retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
2682
2693
  retrieve_image_nodes: typing.Optional[bool] = OMIT,
@@ -2709,7 +2720,7 @@ class AsyncPipelinesClient:
2709
2720
 
2710
2721
  - search_filters: typing.Optional[MetadataFilters].
2711
2722
 
2712
- - search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]].
2723
+ - search_filters_inference_schema: typing.Optional[typing.Dict[str, typing.Optional[RetrievalParamsSearchFiltersInferenceSchemaValue]]].
2713
2724
 
2714
2725
  - files_top_k: typing.Optional[int].
2715
2726
 
@@ -1,5 +1,6 @@
1
1
  # This file was auto-generated by Fern from our API Definition.
2
2
 
3
+ from .pipeline_file_update_custom_metadata_value import PipelineFileUpdateCustomMetadataValue
3
4
  from .pipeline_update_embedding_config import (
4
5
  PipelineUpdateEmbeddingConfig,
5
6
  PipelineUpdateEmbeddingConfig_AzureEmbedding,
@@ -11,8 +12,10 @@ from .pipeline_update_embedding_config import (
11
12
  PipelineUpdateEmbeddingConfig_VertexaiEmbedding,
12
13
  )
13
14
  from .pipeline_update_transform_config import PipelineUpdateTransformConfig
15
+ from .retrieval_params_search_filters_inference_schema_value import RetrievalParamsSearchFiltersInferenceSchemaValue
14
16
 
15
17
  __all__ = [
18
+ "PipelineFileUpdateCustomMetadataValue",
16
19
  "PipelineUpdateEmbeddingConfig",
17
20
  "PipelineUpdateEmbeddingConfig_AzureEmbedding",
18
21
  "PipelineUpdateEmbeddingConfig_BedrockEmbedding",
@@ -22,4 +25,5 @@ __all__ = [
22
25
  "PipelineUpdateEmbeddingConfig_OpenaiEmbedding",
23
26
  "PipelineUpdateEmbeddingConfig_VertexaiEmbedding",
24
27
  "PipelineUpdateTransformConfig",
28
+ "RetrievalParamsSearchFiltersInferenceSchemaValue",
25
29
  ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ PipelineFileUpdateCustomMetadataValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ RetrievalParamsSearchFiltersInferenceSchemaValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -72,6 +72,8 @@ from .data_source import DataSource
72
72
  from .data_source_component import DataSourceComponent
73
73
  from .data_source_create import DataSourceCreate
74
74
  from .data_source_create_component import DataSourceCreateComponent
75
+ from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
76
+ from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
75
77
  from .data_source_update_dispatcher_config import DataSourceUpdateDispatcherConfig
76
78
  from .delete_params import DeleteParams
77
79
  from .document_block import DocumentBlock
@@ -104,17 +106,27 @@ from .embedding_model_config_update_embedding_config import (
104
106
  )
105
107
  from .eval_execution_params import EvalExecutionParams
106
108
  from .extract_agent import ExtractAgent
109
+ from .extract_agent_data_schema_value import ExtractAgentDataSchemaValue
107
110
  from .extract_config import ExtractConfig
108
111
  from .extract_job import ExtractJob
109
112
  from .extract_job_create import ExtractJobCreate
110
113
  from .extract_job_create_data_schema_override import ExtractJobCreateDataSchemaOverride
114
+ from .extract_job_create_data_schema_override_zero_value import ExtractJobCreateDataSchemaOverrideZeroValue
111
115
  from .extract_mode import ExtractMode
112
116
  from .extract_models import ExtractModels
113
117
  from .extract_resultset import ExtractResultset
114
118
  from .extract_resultset_data import ExtractResultsetData
119
+ from .extract_resultset_data_item_value import ExtractResultsetDataItemValue
120
+ from .extract_resultset_data_zero_value import ExtractResultsetDataZeroValue
121
+ from .extract_resultset_extraction_metadata_value import ExtractResultsetExtractionMetadataValue
115
122
  from .extract_run import ExtractRun
116
123
  from .extract_run_data import ExtractRunData
124
+ from .extract_run_data_item_value import ExtractRunDataItemValue
125
+ from .extract_run_data_schema_value import ExtractRunDataSchemaValue
126
+ from .extract_run_data_zero_value import ExtractRunDataZeroValue
127
+ from .extract_run_extraction_metadata_value import ExtractRunExtractionMetadataValue
117
128
  from .extract_schema_validate_response import ExtractSchemaValidateResponse
129
+ from .extract_schema_validate_response_data_schema_value import ExtractSchemaValidateResponseDataSchemaValue
118
130
  from .extract_state import ExtractState
119
131
  from .extract_target import ExtractTarget
120
132
  from .fail_page_mode import FailPageMode
@@ -122,6 +134,8 @@ from .file import File
122
134
  from .file_count_by_status_response import FileCountByStatusResponse
123
135
  from .file_id_presigned_url import FileIdPresignedUrl
124
136
  from .file_parse_public import FileParsePublic
137
+ from .file_permission_info_value import FilePermissionInfoValue
138
+ from .file_resource_info_value import FileResourceInfoValue
125
139
  from .filter_condition import FilterCondition
126
140
  from .filter_operator import FilterOperator
127
141
  from .free_credits_usage import FreeCreditsUsage
@@ -150,7 +164,6 @@ from .job_record_parameters import (
150
164
  JobRecordParameters_PipelineManagedIngestion,
151
165
  )
152
166
  from .job_record_with_usage_metrics import JobRecordWithUsageMetrics
153
- from .json_type import JsonType
154
167
  from .l_lama_parse_transform_config import LLamaParseTransformConfig
155
168
  from .legacy_parse_job_config import LegacyParseJobConfig
156
169
  from .llama_extract_settings import LlamaExtractSettings
@@ -225,6 +238,7 @@ from .pipeline_create_transform_config import PipelineCreateTransformConfig
225
238
  from .pipeline_data_source import PipelineDataSource
226
239
  from .pipeline_data_source_component import PipelineDataSourceComponent
227
240
  from .pipeline_data_source_create import PipelineDataSourceCreate
241
+ from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
228
242
  from .pipeline_data_source_status import PipelineDataSourceStatus
229
243
  from .pipeline_deployment import PipelineDeployment
230
244
  from .pipeline_embedding_config import (
@@ -238,7 +252,12 @@ from .pipeline_embedding_config import (
238
252
  PipelineEmbeddingConfig_VertexaiEmbedding,
239
253
  )
240
254
  from .pipeline_file import PipelineFile
255
+ from .pipeline_file_config_hash_value import PipelineFileConfigHashValue
241
256
  from .pipeline_file_create import PipelineFileCreate
257
+ from .pipeline_file_create_custom_metadata_value import PipelineFileCreateCustomMetadataValue
258
+ from .pipeline_file_custom_metadata_value import PipelineFileCustomMetadataValue
259
+ from .pipeline_file_permission_info_value import PipelineFilePermissionInfoValue
260
+ from .pipeline_file_resource_info_value import PipelineFileResourceInfoValue
242
261
  from .pipeline_file_status import PipelineFileStatus
243
262
  from .pipeline_file_update_dispatcher_config import PipelineFileUpdateDispatcherConfig
244
263
  from .pipeline_file_updater_config import PipelineFileUpdaterConfig
@@ -256,6 +275,9 @@ from .playground_session import PlaygroundSession
256
275
  from .pooling import Pooling
257
276
  from .preset_composite_retrieval_params import PresetCompositeRetrievalParams
258
277
  from .preset_retrieval_params import PresetRetrievalParams
278
+ from .preset_retrieval_params_search_filters_inference_schema_value import (
279
+ PresetRetrievalParamsSearchFiltersInferenceSchemaValue,
280
+ )
259
281
  from .presigned_url import PresignedUrl
260
282
  from .progress_event import ProgressEvent
261
283
  from .progress_event_status import ProgressEventStatus
@@ -390,6 +412,8 @@ __all__ = [
390
412
  "DataSourceComponent",
391
413
  "DataSourceCreate",
392
414
  "DataSourceCreateComponent",
415
+ "DataSourceCreateCustomMetadataValue",
416
+ "DataSourceCustomMetadataValue",
393
417
  "DataSourceUpdateDispatcherConfig",
394
418
  "DeleteParams",
395
419
  "DocumentBlock",
@@ -418,17 +442,27 @@ __all__ = [
418
442
  "EmbeddingModelConfigUpdateEmbeddingConfig_VertexaiEmbedding",
419
443
  "EvalExecutionParams",
420
444
  "ExtractAgent",
445
+ "ExtractAgentDataSchemaValue",
421
446
  "ExtractConfig",
422
447
  "ExtractJob",
423
448
  "ExtractJobCreate",
424
449
  "ExtractJobCreateDataSchemaOverride",
450
+ "ExtractJobCreateDataSchemaOverrideZeroValue",
425
451
  "ExtractMode",
426
452
  "ExtractModels",
427
453
  "ExtractResultset",
428
454
  "ExtractResultsetData",
455
+ "ExtractResultsetDataItemValue",
456
+ "ExtractResultsetDataZeroValue",
457
+ "ExtractResultsetExtractionMetadataValue",
429
458
  "ExtractRun",
430
459
  "ExtractRunData",
460
+ "ExtractRunDataItemValue",
461
+ "ExtractRunDataSchemaValue",
462
+ "ExtractRunDataZeroValue",
463
+ "ExtractRunExtractionMetadataValue",
431
464
  "ExtractSchemaValidateResponse",
465
+ "ExtractSchemaValidateResponseDataSchemaValue",
432
466
  "ExtractState",
433
467
  "ExtractTarget",
434
468
  "FailPageMode",
@@ -436,6 +470,8 @@ __all__ = [
436
470
  "FileCountByStatusResponse",
437
471
  "FileIdPresignedUrl",
438
472
  "FileParsePublic",
473
+ "FilePermissionInfoValue",
474
+ "FileResourceInfoValue",
439
475
  "FilterCondition",
440
476
  "FilterOperator",
441
477
  "FreeCreditsUsage",
@@ -462,7 +498,6 @@ __all__ = [
462
498
  "JobRecordParameters_PipelineFileUpdater",
463
499
  "JobRecordParameters_PipelineManagedIngestion",
464
500
  "JobRecordWithUsageMetrics",
465
- "JsonType",
466
501
  "LLamaParseTransformConfig",
467
502
  "LegacyParseJobConfig",
468
503
  "LlamaExtractSettings",
@@ -533,6 +568,7 @@ __all__ = [
533
568
  "PipelineDataSource",
534
569
  "PipelineDataSourceComponent",
535
570
  "PipelineDataSourceCreate",
571
+ "PipelineDataSourceCustomMetadataValue",
536
572
  "PipelineDataSourceStatus",
537
573
  "PipelineDeployment",
538
574
  "PipelineEmbeddingConfig",
@@ -544,7 +580,12 @@ __all__ = [
544
580
  "PipelineEmbeddingConfig_OpenaiEmbedding",
545
581
  "PipelineEmbeddingConfig_VertexaiEmbedding",
546
582
  "PipelineFile",
583
+ "PipelineFileConfigHashValue",
547
584
  "PipelineFileCreate",
585
+ "PipelineFileCreateCustomMetadataValue",
586
+ "PipelineFileCustomMetadataValue",
587
+ "PipelineFilePermissionInfoValue",
588
+ "PipelineFileResourceInfoValue",
548
589
  "PipelineFileStatus",
549
590
  "PipelineFileUpdateDispatcherConfig",
550
591
  "PipelineFileUpdaterConfig",
@@ -560,6 +601,7 @@ __all__ = [
560
601
  "Pooling",
561
602
  "PresetCompositeRetrievalParams",
562
603
  "PresetRetrievalParams",
604
+ "PresetRetrievalParamsSearchFiltersInferenceSchemaValue",
563
605
  "PresignedUrl",
564
606
  "ProgressEvent",
565
607
  "ProgressEventStatus",
@@ -6,7 +6,7 @@ import typing
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .configurable_data_source_names import ConfigurableDataSourceNames
8
8
  from .data_source_component import DataSourceComponent
9
- from .json_type import JsonType
9
+ from .data_source_custom_metadata_value import DataSourceCustomMetadataValue
10
10
 
11
11
  try:
12
12
  import pydantic
@@ -27,7 +27,7 @@ class DataSource(pydantic.BaseModel):
27
27
  updated_at: typing.Optional[dt.datetime]
28
28
  name: str = pydantic.Field(description="The name of the data source.")
29
29
  source_type: ConfigurableDataSourceNames
30
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]]
30
+ custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceCustomMetadataValue]]]
31
31
  component: DataSourceComponent = pydantic.Field(description="Component that implements the data source")
32
32
  version_metadata: typing.Optional[typing.Dict[str, typing.Any]]
33
33
  project_id: str
@@ -6,7 +6,7 @@ import typing
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .configurable_data_source_names import ConfigurableDataSourceNames
8
8
  from .data_source_create_component import DataSourceCreateComponent
9
- from .json_type import JsonType
9
+ from .data_source_create_custom_metadata_value import DataSourceCreateCustomMetadataValue
10
10
 
11
11
  try:
12
12
  import pydantic
@@ -24,7 +24,7 @@ class DataSourceCreate(pydantic.BaseModel):
24
24
 
25
25
  name: str = pydantic.Field(description="The name of the data source.")
26
26
  source_type: ConfigurableDataSourceNames
27
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]]
27
+ custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceCreateCustomMetadataValue]]]
28
28
  component: DataSourceCreateComponent = pydantic.Field(description="Component that implements the data source")
29
29
 
30
30
  def json(self, **kwargs: typing.Any) -> str:
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ DataSourceCreateCustomMetadataValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ DataSourceCustomMetadataValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -4,8 +4,8 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
+ from .extract_agent_data_schema_value import ExtractAgentDataSchemaValue
7
8
  from .extract_config import ExtractConfig
8
- from .json_type import JsonType
9
9
 
10
10
  try:
11
11
  import pydantic
@@ -24,7 +24,9 @@ class ExtractAgent(pydantic.BaseModel):
24
24
  id: str = pydantic.Field(description="The id of the extraction agent.")
25
25
  name: str = pydantic.Field(description="The name of the extraction agent.")
26
26
  project_id: str = pydantic.Field(description="The ID of the project that the extraction agent belongs to.")
27
- data_schema: typing.Dict[str, typing.Optional[JsonType]] = pydantic.Field(description="The schema of the data.")
27
+ data_schema: typing.Dict[str, typing.Optional[ExtractAgentDataSchemaValue]] = pydantic.Field(
28
+ description="The schema of the data."
29
+ )
28
30
  config: ExtractConfig = pydantic.Field(description="The configuration parameters for the extraction agent.")
29
31
  created_at: typing.Optional[dt.datetime]
30
32
  updated_at: typing.Optional[dt.datetime]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractAgentDataSchemaValue = typing.Union[typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool]
@@ -24,6 +24,9 @@ class ExtractConfig(pydantic.BaseModel):
24
24
 
25
25
  extraction_target: typing.Optional[ExtractTarget] = pydantic.Field(description="The extraction target specified.")
26
26
  extraction_mode: typing.Optional[ExtractMode] = pydantic.Field(description="The extraction mode specified.")
27
+ multimodal_fast_mode: typing.Optional[bool] = pydantic.Field(
28
+ description="Whether to use fast mode for multimodal extraction."
29
+ )
27
30
  system_prompt: typing.Optional[str]
28
31
  use_reasoning: typing.Optional[bool] = pydantic.Field(description="Whether to use reasoning for the extraction.")
29
32
  cite_sources: typing.Optional[bool] = pydantic.Field(description="Whether to cite sources for the extraction.")
@@ -2,6 +2,8 @@
2
2
 
3
3
  import typing
4
4
 
5
- from .json_type import JsonType
5
+ from .extract_job_create_data_schema_override_zero_value import ExtractJobCreateDataSchemaOverrideZeroValue
6
6
 
7
- ExtractJobCreateDataSchemaOverride = typing.Union[typing.Dict[str, typing.Optional[JsonType]], str]
7
+ ExtractJobCreateDataSchemaOverride = typing.Union[
8
+ typing.Dict[str, typing.Optional[ExtractJobCreateDataSchemaOverrideZeroValue]], str
9
+ ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractJobCreateDataSchemaOverrideZeroValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -5,7 +5,7 @@ import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .extract_resultset_data import ExtractResultsetData
8
- from .json_type import JsonType
8
+ from .extract_resultset_extraction_metadata_value import ExtractResultsetExtractionMetadataValue
9
9
 
10
10
  try:
11
11
  import pydantic
@@ -24,7 +24,7 @@ class ExtractResultset(pydantic.BaseModel):
24
24
  run_id: str = pydantic.Field(description="The id of the extraction run")
25
25
  extraction_agent_id: str = pydantic.Field(description="The id of the extraction agent")
26
26
  data: typing.Optional[ExtractResultsetData] = pydantic.Field(description="The data extracted from the file")
27
- extraction_metadata: typing.Dict[str, typing.Optional[JsonType]] = pydantic.Field(
27
+ extraction_metadata: typing.Dict[str, typing.Optional[ExtractResultsetExtractionMetadataValue]] = pydantic.Field(
28
28
  description="The metadata extracted from the file"
29
29
  )
30
30
 
@@ -2,8 +2,10 @@
2
2
 
3
3
  import typing
4
4
 
5
- from .json_type import JsonType
5
+ from .extract_resultset_data_item_value import ExtractResultsetDataItemValue
6
+ from .extract_resultset_data_zero_value import ExtractResultsetDataZeroValue
6
7
 
7
8
  ExtractResultsetData = typing.Union[
8
- typing.Dict[str, typing.Optional[JsonType]], typing.List[typing.Dict[str, typing.Optional[JsonType]]]
9
+ typing.Dict[str, typing.Optional[ExtractResultsetDataZeroValue]],
10
+ typing.List[typing.Dict[str, typing.Optional[ExtractResultsetDataItemValue]]],
9
11
  ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractResultsetDataItemValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractResultsetDataZeroValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractResultsetExtractionMetadataValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -6,9 +6,10 @@ import typing
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .extract_config import ExtractConfig
8
8
  from .extract_run_data import ExtractRunData
9
+ from .extract_run_data_schema_value import ExtractRunDataSchemaValue
10
+ from .extract_run_extraction_metadata_value import ExtractRunExtractionMetadataValue
9
11
  from .extract_state import ExtractState
10
12
  from .file import File
11
- from .json_type import JsonType
12
13
 
13
14
  try:
14
15
  import pydantic
@@ -28,7 +29,7 @@ class ExtractRun(pydantic.BaseModel):
28
29
  created_at: typing.Optional[dt.datetime]
29
30
  updated_at: typing.Optional[dt.datetime]
30
31
  extraction_agent_id: str = pydantic.Field(description="The id of the extraction agent")
31
- data_schema: typing.Dict[str, typing.Optional[JsonType]] = pydantic.Field(
32
+ data_schema: typing.Dict[str, typing.Optional[ExtractRunDataSchemaValue]] = pydantic.Field(
32
33
  description="The schema used for extraction"
33
34
  )
34
35
  config: ExtractConfig = pydantic.Field(description="The config used for extraction")
@@ -37,7 +38,7 @@ class ExtractRun(pydantic.BaseModel):
37
38
  error: typing.Optional[str]
38
39
  job_id: typing.Optional[str]
39
40
  data: typing.Optional[ExtractRunData] = pydantic.Field(description="The data extracted from the file")
40
- extraction_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]]
41
+ extraction_metadata: typing.Optional[typing.Dict[str, typing.Optional[ExtractRunExtractionMetadataValue]]]
41
42
  from_ui: bool = pydantic.Field(description="Whether this extraction run was triggered from the UI")
42
43
 
43
44
  def json(self, **kwargs: typing.Any) -> str:
@@ -2,8 +2,10 @@
2
2
 
3
3
  import typing
4
4
 
5
- from .json_type import JsonType
5
+ from .extract_run_data_item_value import ExtractRunDataItemValue
6
+ from .extract_run_data_zero_value import ExtractRunDataZeroValue
6
7
 
7
8
  ExtractRunData = typing.Union[
8
- typing.Dict[str, typing.Optional[JsonType]], typing.List[typing.Dict[str, typing.Optional[JsonType]]]
9
+ typing.Dict[str, typing.Optional[ExtractRunDataZeroValue]],
10
+ typing.List[typing.Dict[str, typing.Optional[ExtractRunDataItemValue]]],
9
11
  ]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractRunDataItemValue = typing.Union[typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractRunDataSchemaValue = typing.Union[typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractRunDataZeroValue = typing.Union[typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool]
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractRunExtractionMetadataValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
@@ -4,7 +4,7 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
- from .json_type import JsonType
7
+ from .extract_schema_validate_response_data_schema_value import ExtractSchemaValidateResponseDataSchemaValue
8
8
 
9
9
  try:
10
10
  import pydantic
@@ -16,7 +16,7 @@ except ImportError:
16
16
 
17
17
 
18
18
  class ExtractSchemaValidateResponse(pydantic.BaseModel):
19
- data_schema: typing.Dict[str, typing.Optional[JsonType]]
19
+ data_schema: typing.Dict[str, typing.Optional[ExtractSchemaValidateResponseDataSchemaValue]]
20
20
 
21
21
  def json(self, **kwargs: typing.Any) -> str:
22
22
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ ExtractSchemaValidateResponseDataSchemaValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]
llama_cloud/types/file.py CHANGED
@@ -4,7 +4,8 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
- from .json_type import JsonType
7
+ from .file_permission_info_value import FilePermissionInfoValue
8
+ from .file_resource_info_value import FileResourceInfoValue
8
9
 
9
10
  try:
10
11
  import pydantic
@@ -29,8 +30,8 @@ class File(pydantic.BaseModel):
29
30
  file_type: typing.Optional[str]
30
31
  project_id: str = pydantic.Field(description="The ID of the project that the file belongs to")
31
32
  last_modified_at: typing.Optional[dt.datetime]
32
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]]
33
- permission_info: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]]
33
+ resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileResourceInfoValue]]]
34
+ permission_info: typing.Optional[typing.Dict[str, typing.Optional[FilePermissionInfoValue]]]
34
35
  data_source_id: typing.Optional[str]
35
36
 
36
37
  def json(self, **kwargs: typing.Any) -> str:
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ FilePermissionInfoValue = typing.Union[typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool]
@@ -0,0 +1,5 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ FileResourceInfoValue = typing.Union[typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool]
@@ -5,8 +5,8 @@ import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .configurable_data_source_names import ConfigurableDataSourceNames
8
- from .json_type import JsonType
9
8
  from .pipeline_data_source_component import PipelineDataSourceComponent
9
+ from .pipeline_data_source_custom_metadata_value import PipelineDataSourceCustomMetadataValue
10
10
  from .pipeline_data_source_status import PipelineDataSourceStatus
11
11
 
12
12
  try:
@@ -28,7 +28,7 @@ class PipelineDataSource(pydantic.BaseModel):
28
28
  updated_at: typing.Optional[dt.datetime]
29
29
  name: str = pydantic.Field(description="The name of the data source.")
30
30
  source_type: ConfigurableDataSourceNames
31
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[JsonType]]]
31
+ custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineDataSourceCustomMetadataValue]]]
32
32
  component: PipelineDataSourceComponent = pydantic.Field(description="Component that implements the data source")
33
33
  version_metadata: typing.Optional[typing.Dict[str, typing.Any]]
34
34
  project_id: str
@@ -0,0 +1,7 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+
5
+ PipelineDataSourceCustomMetadataValue = typing.Union[
6
+ typing.Dict[str, typing.Any], typing.List[typing.Any], str, int, float, bool
7
+ ]