llama-cloud 0.1.39__py3-none-any.whl → 0.1.41__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_cloud/__init__.py +24 -72
- llama_cloud/client.py +2 -5
- llama_cloud/resources/__init__.py +0 -4
- llama_cloud/resources/admin/client.py +5 -5
- llama_cloud/resources/beta/client.py +1351 -335
- llama_cloud/resources/llama_extract/client.py +56 -0
- llama_cloud/resources/parsing/client.py +8 -0
- llama_cloud/resources/pipelines/client.py +37 -0
- llama_cloud/types/__init__.py +24 -72
- llama_cloud/types/{src_app_schema_chat_chat_message.py → chat_message.py} +1 -1
- llama_cloud/types/extract_config.py +8 -2
- llama_cloud/types/extract_job_create.py +2 -0
- llama_cloud/types/extract_job_create_priority.py +29 -0
- llama_cloud/types/extract_models.py +28 -28
- llama_cloud/types/job_names.py +0 -4
- llama_cloud/types/{document_block.py → llama_extract_feature_availability.py} +5 -6
- llama_cloud/types/llama_extract_mode_availability.py +4 -3
- llama_cloud/types/llama_extract_settings.py +1 -1
- llama_cloud/types/llama_parse_parameters.py +1 -0
- llama_cloud/types/{progress_event.py → parse_configuration.py} +12 -12
- llama_cloud/types/{llama_index_core_base_llms_types_chat_message.py → parse_configuration_create.py} +9 -7
- llama_cloud/types/{edit_suggestion.py → parse_configuration_filter.py} +8 -6
- llama_cloud/types/{report_update_event.py → parse_configuration_query_response.py} +6 -6
- llama_cloud/types/parse_job_config.py +1 -0
- llama_cloud/types/pipeline.py +4 -0
- llama_cloud/types/pipeline_create.py +2 -0
- llama_cloud/types/playground_session.py +2 -2
- llama_cloud/types/public_model_name.py +97 -0
- llama_cloud/types/{report_create_response.py → schema_generation_availability.py} +4 -2
- llama_cloud/types/schema_generation_availability_status.py +17 -0
- llama_cloud/types/{report_event_item.py → sparse_model_config.py} +10 -8
- llama_cloud/types/sparse_model_type.py +33 -0
- llama_cloud/types/webhook_configuration.py +1 -0
- llama_cloud-0.1.41.dist-info/METADATA +106 -0
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/RECORD +37 -56
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/WHEEL +1 -1
- llama_cloud/resources/reports/__init__.py +0 -5
- llama_cloud/resources/reports/client.py +0 -1230
- llama_cloud/resources/reports/types/__init__.py +0 -7
- llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +0 -25
- llama_cloud/types/audio_block.py +0 -34
- llama_cloud/types/edit_suggestion_blocks_item.py +0 -8
- llama_cloud/types/image_block.py +0 -35
- llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +0 -56
- llama_cloud/types/paginated_report_response.py +0 -35
- llama_cloud/types/progress_event_status.py +0 -33
- llama_cloud/types/report.py +0 -33
- llama_cloud/types/report_block.py +0 -35
- llama_cloud/types/report_block_dependency.py +0 -29
- llama_cloud/types/report_event_item_event_data.py +0 -45
- llama_cloud/types/report_event_type.py +0 -37
- llama_cloud/types/report_metadata.py +0 -43
- llama_cloud/types/report_plan.py +0 -36
- llama_cloud/types/report_plan_block.py +0 -36
- llama_cloud/types/report_query.py +0 -33
- llama_cloud/types/report_response.py +0 -41
- llama_cloud/types/report_state.py +0 -37
- llama_cloud/types/report_state_event.py +0 -38
- llama_cloud/types/text_block.py +0 -31
- llama_cloud-0.1.39.dist-info/METADATA +0 -32
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/LICENSE +0 -0
llama_cloud/__init__.py
CHANGED
|
@@ -16,7 +16,6 @@ from .types import (
|
|
|
16
16
|
AgentDeploymentList,
|
|
17
17
|
AgentDeploymentSummary,
|
|
18
18
|
AggregateGroup,
|
|
19
|
-
AudioBlock,
|
|
20
19
|
AutoTransformConfig,
|
|
21
20
|
AzureOpenAiEmbedding,
|
|
22
21
|
AzureOpenAiEmbeddingConfig,
|
|
@@ -36,6 +35,7 @@ from .types import (
|
|
|
36
35
|
ChatApp,
|
|
37
36
|
ChatAppResponse,
|
|
38
37
|
ChatData,
|
|
38
|
+
ChatMessage,
|
|
39
39
|
ChunkMode,
|
|
40
40
|
ClassificationResult,
|
|
41
41
|
ClassifierRule,
|
|
@@ -85,11 +85,8 @@ from .types import (
|
|
|
85
85
|
DataSourceReaderVersionMetadataReaderVersion,
|
|
86
86
|
DataSourceUpdateDispatcherConfig,
|
|
87
87
|
DeleteParams,
|
|
88
|
-
DocumentBlock,
|
|
89
88
|
DocumentChunkMode,
|
|
90
89
|
DocumentIngestionJobParams,
|
|
91
|
-
EditSuggestion,
|
|
92
|
-
EditSuggestionBlocksItem,
|
|
93
90
|
ElementSegmentationConfig,
|
|
94
91
|
EmbeddingModelConfig,
|
|
95
92
|
EmbeddingModelConfigEmbeddingConfig,
|
|
@@ -118,6 +115,7 @@ from .types import (
|
|
|
118
115
|
ExtractJobCreate,
|
|
119
116
|
ExtractJobCreateDataSchemaOverride,
|
|
120
117
|
ExtractJobCreateDataSchemaOverrideZeroValue,
|
|
118
|
+
ExtractJobCreatePriority,
|
|
121
119
|
ExtractMode,
|
|
122
120
|
ExtractModels,
|
|
123
121
|
ExtractResultset,
|
|
@@ -170,7 +168,6 @@ from .types import (
|
|
|
170
168
|
HuggingFaceInferenceApiEmbedding,
|
|
171
169
|
HuggingFaceInferenceApiEmbeddingConfig,
|
|
172
170
|
HuggingFaceInferenceApiEmbeddingToken,
|
|
173
|
-
ImageBlock,
|
|
174
171
|
IngestionErrorResponse,
|
|
175
172
|
InputMessage,
|
|
176
173
|
JobNameMapping,
|
|
@@ -190,15 +187,10 @@ from .types import (
|
|
|
190
187
|
LLamaParseTransformConfig,
|
|
191
188
|
LegacyParseJobConfig,
|
|
192
189
|
LicenseInfoResponse,
|
|
190
|
+
LlamaExtractFeatureAvailability,
|
|
193
191
|
LlamaExtractModeAvailability,
|
|
194
192
|
LlamaExtractModeAvailabilityStatus,
|
|
195
193
|
LlamaExtractSettings,
|
|
196
|
-
LlamaIndexCoreBaseLlmsTypesChatMessage,
|
|
197
|
-
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem,
|
|
198
|
-
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Audio,
|
|
199
|
-
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Document,
|
|
200
|
-
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image,
|
|
201
|
-
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text,
|
|
202
194
|
LlamaParseParameters,
|
|
203
195
|
LlamaParseParametersPriority,
|
|
204
196
|
LlamaParseSupportedFileExtensions,
|
|
@@ -233,11 +225,14 @@ from .types import (
|
|
|
233
225
|
PaginatedJobsHistoryWithMetrics,
|
|
234
226
|
PaginatedListCloudDocumentsResponse,
|
|
235
227
|
PaginatedListPipelineFilesResponse,
|
|
236
|
-
PaginatedReportResponse,
|
|
237
228
|
PaginatedResponseAgentData,
|
|
238
229
|
PaginatedResponseAggregateGroup,
|
|
239
230
|
PaginatedResponseClassifyJob,
|
|
240
231
|
PaginatedResponseQuotaConfiguration,
|
|
232
|
+
ParseConfiguration,
|
|
233
|
+
ParseConfigurationCreate,
|
|
234
|
+
ParseConfigurationFilter,
|
|
235
|
+
ParseConfigurationQueryResponse,
|
|
241
236
|
ParseJobConfig,
|
|
242
237
|
ParseJobConfigPriority,
|
|
243
238
|
ParsePlanLevel,
|
|
@@ -305,11 +300,10 @@ from .types import (
|
|
|
305
300
|
PresetRetrievalParams,
|
|
306
301
|
PresetRetrievalParamsSearchFiltersInferenceSchemaValue,
|
|
307
302
|
PresignedUrl,
|
|
308
|
-
ProgressEvent,
|
|
309
|
-
ProgressEventStatus,
|
|
310
303
|
Project,
|
|
311
304
|
ProjectCreate,
|
|
312
305
|
PromptConf,
|
|
306
|
+
PublicModelName,
|
|
313
307
|
QuotaConfiguration,
|
|
314
308
|
QuotaConfigurationConfigurationType,
|
|
315
309
|
QuotaConfigurationStatus,
|
|
@@ -320,40 +314,24 @@ from .types import (
|
|
|
320
314
|
RecurringCreditGrant,
|
|
321
315
|
RelatedNodeInfo,
|
|
322
316
|
RelatedNodeInfoNodeType,
|
|
323
|
-
Report,
|
|
324
|
-
ReportBlock,
|
|
325
|
-
ReportBlockDependency,
|
|
326
|
-
ReportCreateResponse,
|
|
327
|
-
ReportEventItem,
|
|
328
|
-
ReportEventItemEventData,
|
|
329
|
-
ReportEventItemEventData_Progress,
|
|
330
|
-
ReportEventItemEventData_ReportBlockUpdate,
|
|
331
|
-
ReportEventItemEventData_ReportStateUpdate,
|
|
332
|
-
ReportEventType,
|
|
333
|
-
ReportMetadata,
|
|
334
|
-
ReportPlan,
|
|
335
|
-
ReportPlanBlock,
|
|
336
|
-
ReportQuery,
|
|
337
|
-
ReportResponse,
|
|
338
|
-
ReportState,
|
|
339
|
-
ReportStateEvent,
|
|
340
|
-
ReportUpdateEvent,
|
|
341
317
|
RetrievalMode,
|
|
342
318
|
RetrieveResults,
|
|
343
319
|
Retriever,
|
|
344
320
|
RetrieverCreate,
|
|
345
321
|
RetrieverPipeline,
|
|
346
322
|
Role,
|
|
323
|
+
SchemaGenerationAvailability,
|
|
324
|
+
SchemaGenerationAvailabilityStatus,
|
|
347
325
|
SchemaRelaxMode,
|
|
348
326
|
SemanticChunkingConfig,
|
|
349
327
|
SentenceChunkingConfig,
|
|
350
|
-
|
|
328
|
+
SparseModelConfig,
|
|
329
|
+
SparseModelType,
|
|
351
330
|
StatusEnum,
|
|
352
331
|
StructMode,
|
|
353
332
|
StructParseConf,
|
|
354
333
|
SupportedLlmModel,
|
|
355
334
|
SupportedLlmModelNames,
|
|
356
|
-
TextBlock,
|
|
357
335
|
TextNode,
|
|
358
336
|
TextNodeRelationshipsValue,
|
|
359
337
|
TextNodeWithScore,
|
|
@@ -412,7 +390,6 @@ from .resources import (
|
|
|
412
390
|
PipelineUpdateEmbeddingConfig_VertexaiEmbedding,
|
|
413
391
|
PipelineUpdateTransformConfig,
|
|
414
392
|
RetrievalParamsSearchFiltersInferenceSchemaValue,
|
|
415
|
-
UpdateReportPlanApiV1ReportsReportIdPlanPatchRequestAction,
|
|
416
393
|
admin,
|
|
417
394
|
agent_deployments,
|
|
418
395
|
alpha,
|
|
@@ -430,7 +407,6 @@ from .resources import (
|
|
|
430
407
|
parsing,
|
|
431
408
|
pipelines,
|
|
432
409
|
projects,
|
|
433
|
-
reports,
|
|
434
410
|
retrievers,
|
|
435
411
|
users,
|
|
436
412
|
)
|
|
@@ -452,7 +428,6 @@ __all__ = [
|
|
|
452
428
|
"AgentDeploymentList",
|
|
453
429
|
"AgentDeploymentSummary",
|
|
454
430
|
"AggregateGroup",
|
|
455
|
-
"AudioBlock",
|
|
456
431
|
"AutoTransformConfig",
|
|
457
432
|
"AzureOpenAiEmbedding",
|
|
458
433
|
"AzureOpenAiEmbeddingConfig",
|
|
@@ -472,6 +447,7 @@ __all__ = [
|
|
|
472
447
|
"ChatApp",
|
|
473
448
|
"ChatAppResponse",
|
|
474
449
|
"ChatData",
|
|
450
|
+
"ChatMessage",
|
|
475
451
|
"ChunkMode",
|
|
476
452
|
"ClassificationResult",
|
|
477
453
|
"ClassifierRule",
|
|
@@ -524,11 +500,8 @@ __all__ = [
|
|
|
524
500
|
"DataSourceUpdateCustomMetadataValue",
|
|
525
501
|
"DataSourceUpdateDispatcherConfig",
|
|
526
502
|
"DeleteParams",
|
|
527
|
-
"DocumentBlock",
|
|
528
503
|
"DocumentChunkMode",
|
|
529
504
|
"DocumentIngestionJobParams",
|
|
530
|
-
"EditSuggestion",
|
|
531
|
-
"EditSuggestionBlocksItem",
|
|
532
505
|
"ElementSegmentationConfig",
|
|
533
506
|
"EmbeddingModelConfig",
|
|
534
507
|
"EmbeddingModelConfigCreateEmbeddingConfig",
|
|
@@ -571,6 +544,7 @@ __all__ = [
|
|
|
571
544
|
"ExtractJobCreateBatchDataSchemaOverrideZeroValue",
|
|
572
545
|
"ExtractJobCreateDataSchemaOverride",
|
|
573
546
|
"ExtractJobCreateDataSchemaOverrideZeroValue",
|
|
547
|
+
"ExtractJobCreatePriority",
|
|
574
548
|
"ExtractMode",
|
|
575
549
|
"ExtractModels",
|
|
576
550
|
"ExtractResultset",
|
|
@@ -628,7 +602,6 @@ __all__ = [
|
|
|
628
602
|
"HuggingFaceInferenceApiEmbedding",
|
|
629
603
|
"HuggingFaceInferenceApiEmbeddingConfig",
|
|
630
604
|
"HuggingFaceInferenceApiEmbeddingToken",
|
|
631
|
-
"ImageBlock",
|
|
632
605
|
"IngestionErrorResponse",
|
|
633
606
|
"InputMessage",
|
|
634
607
|
"JobNameMapping",
|
|
@@ -649,15 +622,10 @@ __all__ = [
|
|
|
649
622
|
"LegacyParseJobConfig",
|
|
650
623
|
"LicenseInfoResponse",
|
|
651
624
|
"LlamaCloudEnvironment",
|
|
625
|
+
"LlamaExtractFeatureAvailability",
|
|
652
626
|
"LlamaExtractModeAvailability",
|
|
653
627
|
"LlamaExtractModeAvailabilityStatus",
|
|
654
628
|
"LlamaExtractSettings",
|
|
655
|
-
"LlamaIndexCoreBaseLlmsTypesChatMessage",
|
|
656
|
-
"LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem",
|
|
657
|
-
"LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Audio",
|
|
658
|
-
"LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Document",
|
|
659
|
-
"LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image",
|
|
660
|
-
"LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text",
|
|
661
629
|
"LlamaParseParameters",
|
|
662
630
|
"LlamaParseParametersPriority",
|
|
663
631
|
"LlamaParseSupportedFileExtensions",
|
|
@@ -692,11 +660,14 @@ __all__ = [
|
|
|
692
660
|
"PaginatedJobsHistoryWithMetrics",
|
|
693
661
|
"PaginatedListCloudDocumentsResponse",
|
|
694
662
|
"PaginatedListPipelineFilesResponse",
|
|
695
|
-
"PaginatedReportResponse",
|
|
696
663
|
"PaginatedResponseAgentData",
|
|
697
664
|
"PaginatedResponseAggregateGroup",
|
|
698
665
|
"PaginatedResponseClassifyJob",
|
|
699
666
|
"PaginatedResponseQuotaConfiguration",
|
|
667
|
+
"ParseConfiguration",
|
|
668
|
+
"ParseConfigurationCreate",
|
|
669
|
+
"ParseConfigurationFilter",
|
|
670
|
+
"ParseConfigurationQueryResponse",
|
|
700
671
|
"ParseJobConfig",
|
|
701
672
|
"ParseJobConfigPriority",
|
|
702
673
|
"ParsePlanLevel",
|
|
@@ -774,11 +745,10 @@ __all__ = [
|
|
|
774
745
|
"PresetRetrievalParams",
|
|
775
746
|
"PresetRetrievalParamsSearchFiltersInferenceSchemaValue",
|
|
776
747
|
"PresignedUrl",
|
|
777
|
-
"ProgressEvent",
|
|
778
|
-
"ProgressEventStatus",
|
|
779
748
|
"Project",
|
|
780
749
|
"ProjectCreate",
|
|
781
750
|
"PromptConf",
|
|
751
|
+
"PublicModelName",
|
|
782
752
|
"QuotaConfiguration",
|
|
783
753
|
"QuotaConfigurationConfigurationType",
|
|
784
754
|
"QuotaConfigurationStatus",
|
|
@@ -789,24 +759,6 @@ __all__ = [
|
|
|
789
759
|
"RecurringCreditGrant",
|
|
790
760
|
"RelatedNodeInfo",
|
|
791
761
|
"RelatedNodeInfoNodeType",
|
|
792
|
-
"Report",
|
|
793
|
-
"ReportBlock",
|
|
794
|
-
"ReportBlockDependency",
|
|
795
|
-
"ReportCreateResponse",
|
|
796
|
-
"ReportEventItem",
|
|
797
|
-
"ReportEventItemEventData",
|
|
798
|
-
"ReportEventItemEventData_Progress",
|
|
799
|
-
"ReportEventItemEventData_ReportBlockUpdate",
|
|
800
|
-
"ReportEventItemEventData_ReportStateUpdate",
|
|
801
|
-
"ReportEventType",
|
|
802
|
-
"ReportMetadata",
|
|
803
|
-
"ReportPlan",
|
|
804
|
-
"ReportPlanBlock",
|
|
805
|
-
"ReportQuery",
|
|
806
|
-
"ReportResponse",
|
|
807
|
-
"ReportState",
|
|
808
|
-
"ReportStateEvent",
|
|
809
|
-
"ReportUpdateEvent",
|
|
810
762
|
"RetrievalMode",
|
|
811
763
|
"RetrievalParamsSearchFiltersInferenceSchemaValue",
|
|
812
764
|
"RetrieveResults",
|
|
@@ -814,22 +766,23 @@ __all__ = [
|
|
|
814
766
|
"RetrieverCreate",
|
|
815
767
|
"RetrieverPipeline",
|
|
816
768
|
"Role",
|
|
769
|
+
"SchemaGenerationAvailability",
|
|
770
|
+
"SchemaGenerationAvailabilityStatus",
|
|
817
771
|
"SchemaRelaxMode",
|
|
818
772
|
"SemanticChunkingConfig",
|
|
819
773
|
"SentenceChunkingConfig",
|
|
820
|
-
"
|
|
774
|
+
"SparseModelConfig",
|
|
775
|
+
"SparseModelType",
|
|
821
776
|
"StatusEnum",
|
|
822
777
|
"StructMode",
|
|
823
778
|
"StructParseConf",
|
|
824
779
|
"SupportedLlmModel",
|
|
825
780
|
"SupportedLlmModelNames",
|
|
826
|
-
"TextBlock",
|
|
827
781
|
"TextNode",
|
|
828
782
|
"TextNodeRelationshipsValue",
|
|
829
783
|
"TextNodeWithScore",
|
|
830
784
|
"TokenChunkingConfig",
|
|
831
785
|
"UnprocessableEntityError",
|
|
832
|
-
"UpdateReportPlanApiV1ReportsReportIdPlanPatchRequestAction",
|
|
833
786
|
"UpdateUserResponse",
|
|
834
787
|
"UsageAndPlan",
|
|
835
788
|
"UsageMetricResponse",
|
|
@@ -865,7 +818,6 @@ __all__ = [
|
|
|
865
818
|
"parsing",
|
|
866
819
|
"pipelines",
|
|
867
820
|
"projects",
|
|
868
|
-
"reports",
|
|
869
821
|
"retrievers",
|
|
870
822
|
"users",
|
|
871
823
|
]
|
llama_cloud/client.py
CHANGED
|
@@ -23,7 +23,6 @@ from .resources.organizations.client import AsyncOrganizationsClient, Organizati
|
|
|
23
23
|
from .resources.parsing.client import AsyncParsingClient, ParsingClient
|
|
24
24
|
from .resources.pipelines.client import AsyncPipelinesClient, PipelinesClient
|
|
25
25
|
from .resources.projects.client import AsyncProjectsClient, ProjectsClient
|
|
26
|
-
from .resources.reports.client import AsyncReportsClient, ReportsClient
|
|
27
26
|
from .resources.retrievers.client import AsyncRetrieversClient, RetrieversClient
|
|
28
27
|
from .resources.users.client import AsyncUsersClient, UsersClient
|
|
29
28
|
|
|
@@ -43,6 +42,7 @@ class LlamaCloud:
|
|
|
43
42
|
token=token,
|
|
44
43
|
httpx_client=httpx.Client(timeout=timeout) if httpx_client is None else httpx_client,
|
|
45
44
|
)
|
|
45
|
+
self.agent_deployments = AgentDeploymentsClient(client_wrapper=self._client_wrapper)
|
|
46
46
|
self.data_sinks = DataSinksClient(client_wrapper=self._client_wrapper)
|
|
47
47
|
self.data_sources = DataSourcesClient(client_wrapper=self._client_wrapper)
|
|
48
48
|
self.embedding_model_configs = EmbeddingModelConfigsClient(client_wrapper=self._client_wrapper)
|
|
@@ -55,12 +55,10 @@ class LlamaCloud:
|
|
|
55
55
|
self.evals = EvalsClient(client_wrapper=self._client_wrapper)
|
|
56
56
|
self.parsing = ParsingClient(client_wrapper=self._client_wrapper)
|
|
57
57
|
self.chat_apps = ChatAppsClient(client_wrapper=self._client_wrapper)
|
|
58
|
-
self.agent_deployments = AgentDeploymentsClient(client_wrapper=self._client_wrapper)
|
|
59
58
|
self.classifier = ClassifierClient(client_wrapper=self._client_wrapper)
|
|
60
59
|
self.admin = AdminClient(client_wrapper=self._client_wrapper)
|
|
61
60
|
self.users = UsersClient(client_wrapper=self._client_wrapper)
|
|
62
61
|
self.llama_extract = LlamaExtractClient(client_wrapper=self._client_wrapper)
|
|
63
|
-
self.reports = ReportsClient(client_wrapper=self._client_wrapper)
|
|
64
62
|
self.beta = BetaClient(client_wrapper=self._client_wrapper)
|
|
65
63
|
self.alpha = AlphaClient(client_wrapper=self._client_wrapper)
|
|
66
64
|
|
|
@@ -80,6 +78,7 @@ class AsyncLlamaCloud:
|
|
|
80
78
|
token=token,
|
|
81
79
|
httpx_client=httpx.AsyncClient(timeout=timeout) if httpx_client is None else httpx_client,
|
|
82
80
|
)
|
|
81
|
+
self.agent_deployments = AsyncAgentDeploymentsClient(client_wrapper=self._client_wrapper)
|
|
83
82
|
self.data_sinks = AsyncDataSinksClient(client_wrapper=self._client_wrapper)
|
|
84
83
|
self.data_sources = AsyncDataSourcesClient(client_wrapper=self._client_wrapper)
|
|
85
84
|
self.embedding_model_configs = AsyncEmbeddingModelConfigsClient(client_wrapper=self._client_wrapper)
|
|
@@ -92,12 +91,10 @@ class AsyncLlamaCloud:
|
|
|
92
91
|
self.evals = AsyncEvalsClient(client_wrapper=self._client_wrapper)
|
|
93
92
|
self.parsing = AsyncParsingClient(client_wrapper=self._client_wrapper)
|
|
94
93
|
self.chat_apps = AsyncChatAppsClient(client_wrapper=self._client_wrapper)
|
|
95
|
-
self.agent_deployments = AsyncAgentDeploymentsClient(client_wrapper=self._client_wrapper)
|
|
96
94
|
self.classifier = AsyncClassifierClient(client_wrapper=self._client_wrapper)
|
|
97
95
|
self.admin = AsyncAdminClient(client_wrapper=self._client_wrapper)
|
|
98
96
|
self.users = AsyncUsersClient(client_wrapper=self._client_wrapper)
|
|
99
97
|
self.llama_extract = AsyncLlamaExtractClient(client_wrapper=self._client_wrapper)
|
|
100
|
-
self.reports = AsyncReportsClient(client_wrapper=self._client_wrapper)
|
|
101
98
|
self.beta = AsyncBetaClient(client_wrapper=self._client_wrapper)
|
|
102
99
|
self.alpha = AsyncAlphaClient(client_wrapper=self._client_wrapper)
|
|
103
100
|
|
|
@@ -18,7 +18,6 @@ from . import (
|
|
|
18
18
|
parsing,
|
|
19
19
|
pipelines,
|
|
20
20
|
projects,
|
|
21
|
-
reports,
|
|
22
21
|
retrievers,
|
|
23
22
|
users,
|
|
24
23
|
)
|
|
@@ -60,7 +59,6 @@ from .pipelines import (
|
|
|
60
59
|
PipelineUpdateTransformConfig,
|
|
61
60
|
RetrievalParamsSearchFiltersInferenceSchemaValue,
|
|
62
61
|
)
|
|
63
|
-
from .reports import UpdateReportPlanApiV1ReportsReportIdPlanPatchRequestAction
|
|
64
62
|
|
|
65
63
|
__all__ = [
|
|
66
64
|
"DataSinkUpdateComponent",
|
|
@@ -96,7 +94,6 @@ __all__ = [
|
|
|
96
94
|
"PipelineUpdateEmbeddingConfig_VertexaiEmbedding",
|
|
97
95
|
"PipelineUpdateTransformConfig",
|
|
98
96
|
"RetrievalParamsSearchFiltersInferenceSchemaValue",
|
|
99
|
-
"UpdateReportPlanApiV1ReportsReportIdPlanPatchRequestAction",
|
|
100
97
|
"admin",
|
|
101
98
|
"agent_deployments",
|
|
102
99
|
"alpha",
|
|
@@ -114,7 +111,6 @@ __all__ = [
|
|
|
114
111
|
"parsing",
|
|
115
112
|
"pipelines",
|
|
116
113
|
"projects",
|
|
117
|
-
"reports",
|
|
118
114
|
"retrievers",
|
|
119
115
|
"users",
|
|
120
116
|
]
|
|
@@ -11,7 +11,7 @@ from ...errors.unprocessable_entity_error import UnprocessableEntityError
|
|
|
11
11
|
from ...types.file_store_info_response import FileStoreInfoResponse
|
|
12
12
|
from ...types.http_validation_error import HttpValidationError
|
|
13
13
|
from ...types.license_info_response import LicenseInfoResponse
|
|
14
|
-
from ...types.
|
|
14
|
+
from ...types.llama_extract_feature_availability import LlamaExtractFeatureAvailability
|
|
15
15
|
|
|
16
16
|
try:
|
|
17
17
|
import pydantic
|
|
@@ -80,7 +80,7 @@ class AdminClient:
|
|
|
80
80
|
raise ApiError(status_code=_response.status_code, body=_response.text)
|
|
81
81
|
raise ApiError(status_code=_response.status_code, body=_response_json)
|
|
82
82
|
|
|
83
|
-
def get_llamaextract_features(self) ->
|
|
83
|
+
def get_llamaextract_features(self) -> LlamaExtractFeatureAvailability:
|
|
84
84
|
"""
|
|
85
85
|
Get LlamaExtract feature availability based on available models.
|
|
86
86
|
|
|
@@ -99,7 +99,7 @@ class AdminClient:
|
|
|
99
99
|
timeout=60,
|
|
100
100
|
)
|
|
101
101
|
if 200 <= _response.status_code < 300:
|
|
102
|
-
return pydantic.parse_obj_as(
|
|
102
|
+
return pydantic.parse_obj_as(LlamaExtractFeatureAvailability, _response.json()) # type: ignore
|
|
103
103
|
if _response.status_code == 422:
|
|
104
104
|
raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
|
|
105
105
|
try:
|
|
@@ -167,7 +167,7 @@ class AsyncAdminClient:
|
|
|
167
167
|
raise ApiError(status_code=_response.status_code, body=_response.text)
|
|
168
168
|
raise ApiError(status_code=_response.status_code, body=_response_json)
|
|
169
169
|
|
|
170
|
-
async def get_llamaextract_features(self) ->
|
|
170
|
+
async def get_llamaextract_features(self) -> LlamaExtractFeatureAvailability:
|
|
171
171
|
"""
|
|
172
172
|
Get LlamaExtract feature availability based on available models.
|
|
173
173
|
|
|
@@ -186,7 +186,7 @@ class AsyncAdminClient:
|
|
|
186
186
|
timeout=60,
|
|
187
187
|
)
|
|
188
188
|
if 200 <= _response.status_code < 300:
|
|
189
|
-
return pydantic.parse_obj_as(
|
|
189
|
+
return pydantic.parse_obj_as(LlamaExtractFeatureAvailability, _response.json()) # type: ignore
|
|
190
190
|
if _response.status_code == 422:
|
|
191
191
|
raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
|
|
192
192
|
try:
|