llama-cloud 0.1.6__py3-none-any.whl → 0.1.7__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +140 -6
- llama_cloud/client.py +15 -0
- llama_cloud/environment.py +1 -1
- llama_cloud/resources/__init__.py +15 -0
- llama_cloud/{types/token.py → resources/chat_apps/__init__.py} +0 -3
- llama_cloud/resources/chat_apps/client.py +630 -0
- llama_cloud/resources/data_sinks/client.py +12 -12
- llama_cloud/resources/data_sources/client.py +14 -14
- llama_cloud/resources/embedding_model_configs/client.py +20 -76
- llama_cloud/resources/evals/client.py +26 -36
- llama_cloud/resources/extraction/client.py +32 -32
- llama_cloud/resources/files/client.py +40 -44
- llama_cloud/resources/jobs/__init__.py +2 -0
- llama_cloud/resources/jobs/client.py +148 -0
- llama_cloud/resources/llama_extract/__init__.py +5 -0
- llama_cloud/resources/llama_extract/client.py +1038 -0
- llama_cloud/resources/llama_extract/types/__init__.py +6 -0
- llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_value.py +7 -0
- llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema_value.py +7 -0
- llama_cloud/resources/organizations/client.py +66 -70
- llama_cloud/resources/parsing/client.py +448 -428
- llama_cloud/resources/pipelines/client.py +256 -344
- llama_cloud/resources/projects/client.py +34 -60
- llama_cloud/resources/reports/__init__.py +5 -0
- llama_cloud/resources/reports/client.py +1198 -0
- llama_cloud/resources/reports/types/__init__.py +7 -0
- llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +25 -0
- llama_cloud/resources/retrievers/__init__.py +2 -0
- llama_cloud/resources/retrievers/client.py +654 -0
- llama_cloud/types/__init__.py +128 -6
- llama_cloud/types/{chat_message.py → app_schema_chat_chat_message.py} +3 -3
- llama_cloud/types/azure_open_ai_embedding.py +6 -12
- llama_cloud/types/base_prompt_template.py +2 -6
- llama_cloud/types/bedrock_embedding.py +6 -12
- llama_cloud/types/character_splitter.py +2 -4
- llama_cloud/types/chat_app.py +44 -0
- llama_cloud/types/chat_app_response.py +41 -0
- llama_cloud/types/cloud_az_storage_blob_data_source.py +7 -15
- llama_cloud/types/cloud_box_data_source.py +6 -12
- llama_cloud/types/cloud_confluence_data_source.py +6 -6
- llama_cloud/types/cloud_document.py +1 -3
- llama_cloud/types/cloud_document_create.py +1 -3
- llama_cloud/types/cloud_jira_data_source.py +4 -6
- llama_cloud/types/cloud_notion_page_data_source.py +2 -2
- llama_cloud/types/cloud_one_drive_data_source.py +3 -5
- llama_cloud/types/cloud_postgres_vector_store.py +1 -0
- llama_cloud/types/cloud_s_3_data_source.py +4 -8
- llama_cloud/types/cloud_sharepoint_data_source.py +6 -8
- llama_cloud/types/cloud_slack_data_source.py +6 -6
- llama_cloud/types/code_splitter.py +1 -1
- llama_cloud/types/cohere_embedding.py +3 -7
- llama_cloud/types/composite_retrieval_mode.py +21 -0
- llama_cloud/types/composite_retrieval_result.py +38 -0
- llama_cloud/types/composite_retrieved_text_node.py +42 -0
- llama_cloud/types/data_sink.py +4 -4
- llama_cloud/types/data_sink_component.py +20 -0
- llama_cloud/types/data_source.py +5 -7
- llama_cloud/types/data_source_component.py +28 -0
- llama_cloud/types/data_source_create.py +1 -3
- llama_cloud/types/edit_suggestion.py +39 -0
- llama_cloud/types/embedding_model_config.py +2 -2
- llama_cloud/types/embedding_model_config_update.py +2 -4
- llama_cloud/types/eval_dataset.py +2 -2
- llama_cloud/types/eval_dataset_job_record.py +8 -13
- llama_cloud/types/eval_execution_params_override.py +2 -6
- llama_cloud/types/eval_question.py +2 -2
- llama_cloud/types/extract_agent.py +45 -0
- llama_cloud/types/extract_agent_data_schema_value.py +5 -0
- llama_cloud/types/extract_config.py +40 -0
- llama_cloud/types/extract_job.py +35 -0
- llama_cloud/types/extract_job_create.py +40 -0
- llama_cloud/types/extract_job_create_data_schema_override_value.py +7 -0
- llama_cloud/types/extract_mode.py +17 -0
- llama_cloud/types/extract_resultset.py +46 -0
- llama_cloud/types/extract_resultset_data.py +11 -0
- llama_cloud/types/extract_resultset_data_item_value.py +7 -0
- llama_cloud/types/extract_resultset_data_zero_value.py +7 -0
- llama_cloud/types/extract_resultset_extraction_metadata_value.py +7 -0
- llama_cloud/types/extraction_result.py +2 -2
- llama_cloud/types/extraction_schema.py +3 -5
- llama_cloud/types/file.py +9 -14
- llama_cloud/types/filter_condition.py +9 -1
- llama_cloud/types/filter_operator.py +6 -2
- llama_cloud/types/gemini_embedding.py +6 -10
- llama_cloud/types/hugging_face_inference_api_embedding.py +11 -27
- llama_cloud/types/hugging_face_inference_api_embedding_token.py +5 -0
- llama_cloud/types/image_block.py +35 -0
- llama_cloud/types/input_message.py +2 -4
- llama_cloud/types/job_names.py +89 -0
- llama_cloud/types/job_record.py +57 -0
- llama_cloud/types/job_record_with_usage_metrics.py +36 -0
- llama_cloud/types/llama_index_core_base_llms_types_chat_message.py +39 -0
- llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +33 -0
- llama_cloud/types/llama_parse_parameters.py +4 -0
- llama_cloud/types/llm.py +3 -4
- llama_cloud/types/llm_model_data.py +1 -0
- llama_cloud/types/llm_parameters.py +3 -5
- llama_cloud/types/local_eval.py +8 -10
- llama_cloud/types/local_eval_results.py +1 -1
- llama_cloud/types/managed_ingestion_status.py +4 -0
- llama_cloud/types/managed_ingestion_status_response.py +4 -5
- llama_cloud/types/markdown_element_node_parser.py +3 -5
- llama_cloud/types/markdown_node_parser.py +1 -1
- llama_cloud/types/metadata_filter.py +2 -2
- llama_cloud/types/metadata_filter_value.py +5 -0
- llama_cloud/types/metric_result.py +3 -3
- llama_cloud/types/node_parser.py +1 -1
- llama_cloud/types/object_type.py +4 -0
- llama_cloud/types/open_ai_embedding.py +6 -12
- llama_cloud/types/organization.py +7 -2
- llama_cloud/types/page_splitter_node_parser.py +2 -2
- llama_cloud/types/paginated_jobs_history_with_metrics.py +35 -0
- llama_cloud/types/paginated_report_response.py +35 -0
- llama_cloud/types/parse_plan_level.py +21 -0
- llama_cloud/types/permission.py +3 -3
- llama_cloud/types/pipeline.py +7 -17
- llama_cloud/types/pipeline_configuration_hashes.py +3 -3
- llama_cloud/types/pipeline_create.py +8 -16
- llama_cloud/types/pipeline_data_source.py +7 -13
- llama_cloud/types/pipeline_data_source_component.py +28 -0
- llama_cloud/types/pipeline_data_source_create.py +1 -3
- llama_cloud/types/pipeline_deployment.py +4 -4
- llama_cloud/types/pipeline_file.py +13 -24
- llama_cloud/types/pipeline_file_create.py +1 -3
- llama_cloud/types/playground_session.py +4 -4
- llama_cloud/types/preset_retrieval_params.py +8 -14
- llama_cloud/types/presigned_url.py +1 -3
- llama_cloud/types/progress_event.py +44 -0
- llama_cloud/types/progress_event_status.py +33 -0
- llama_cloud/types/project.py +2 -2
- llama_cloud/types/prompt_mixin_prompts.py +1 -1
- llama_cloud/types/prompt_spec.py +3 -5
- llama_cloud/types/related_node_info.py +2 -2
- llama_cloud/types/related_node_info_node_type.py +7 -0
- llama_cloud/types/report.py +33 -0
- llama_cloud/types/report_block.py +34 -0
- llama_cloud/types/report_block_dependency.py +29 -0
- llama_cloud/types/report_create_response.py +31 -0
- llama_cloud/types/report_event_item.py +40 -0
- llama_cloud/types/report_event_item_event_data.py +45 -0
- llama_cloud/types/report_event_type.py +37 -0
- llama_cloud/types/report_metadata.py +43 -0
- llama_cloud/types/report_plan.py +36 -0
- llama_cloud/types/report_plan_block.py +36 -0
- llama_cloud/types/report_query.py +33 -0
- llama_cloud/types/report_response.py +41 -0
- llama_cloud/types/report_state.py +37 -0
- llama_cloud/types/report_state_event.py +38 -0
- llama_cloud/types/report_update_event.py +38 -0
- llama_cloud/types/retrieve_results.py +1 -1
- llama_cloud/types/retriever.py +45 -0
- llama_cloud/types/retriever_create.py +37 -0
- llama_cloud/types/retriever_pipeline.py +37 -0
- llama_cloud/types/role.py +3 -3
- llama_cloud/types/sentence_splitter.py +2 -4
- llama_cloud/types/status_enum.py +4 -0
- llama_cloud/types/supported_llm_model_names.py +4 -0
- llama_cloud/types/text_block.py +31 -0
- llama_cloud/types/text_node.py +15 -8
- llama_cloud/types/token_text_splitter.py +1 -1
- llama_cloud/types/usage_metric_response.py +34 -0
- llama_cloud/types/user_job_record.py +32 -0
- llama_cloud/types/user_organization.py +5 -9
- llama_cloud/types/user_organization_create.py +4 -4
- llama_cloud/types/user_organization_delete.py +2 -2
- llama_cloud/types/user_organization_role.py +2 -2
- llama_cloud/types/vertex_text_embedding.py +5 -9
- {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7.dist-info}/METADATA +2 -1
- llama_cloud-0.1.7.dist-info/RECORD +310 -0
- llama_cloud/types/value.py +0 -5
- llama_cloud-0.1.6.dist-info/RECORD +0 -241
- {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7.dist-info}/LICENSE +0 -0
- {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7.dist-info}/WHEEL +0 -0
|
@@ -91,7 +91,7 @@ class DataSinksClient:
|
|
|
91
91
|
)
|
|
92
92
|
client.data_sinks.create_data_sink(
|
|
93
93
|
request=DataSinkCreate(
|
|
94
|
-
name="
|
|
94
|
+
name="string",
|
|
95
95
|
sink_type=ConfigurableDataSinkNames.PINECONE,
|
|
96
96
|
),
|
|
97
97
|
)
|
|
@@ -140,7 +140,7 @@ class DataSinksClient:
|
|
|
140
140
|
)
|
|
141
141
|
client.data_sinks.upsert_data_sink(
|
|
142
142
|
request=DataSinkCreate(
|
|
143
|
-
name="
|
|
143
|
+
name="string",
|
|
144
144
|
sink_type=ConfigurableDataSinkNames.PINECONE,
|
|
145
145
|
),
|
|
146
146
|
)
|
|
@@ -176,7 +176,7 @@ class DataSinksClient:
|
|
|
176
176
|
token="YOUR_TOKEN",
|
|
177
177
|
)
|
|
178
178
|
client.data_sinks.get_data_sink(
|
|
179
|
-
data_sink_id="
|
|
179
|
+
data_sink_id="string",
|
|
180
180
|
)
|
|
181
181
|
"""
|
|
182
182
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -209,7 +209,7 @@ class DataSinksClient:
|
|
|
209
209
|
Parameters:
|
|
210
210
|
- data_sink_id: str.
|
|
211
211
|
|
|
212
|
-
- name: typing.Optional[str].
|
|
212
|
+
- name: typing.Optional[str].
|
|
213
213
|
|
|
214
214
|
- sink_type: ConfigurableDataSinkNames.
|
|
215
215
|
|
|
@@ -222,7 +222,7 @@ class DataSinksClient:
|
|
|
222
222
|
token="YOUR_TOKEN",
|
|
223
223
|
)
|
|
224
224
|
client.data_sinks.update_data_sink(
|
|
225
|
-
data_sink_id="
|
|
225
|
+
data_sink_id="string",
|
|
226
226
|
sink_type=ConfigurableDataSinkNames.PINECONE,
|
|
227
227
|
)
|
|
228
228
|
"""
|
|
@@ -261,7 +261,7 @@ class DataSinksClient:
|
|
|
261
261
|
token="YOUR_TOKEN",
|
|
262
262
|
)
|
|
263
263
|
client.data_sinks.delete_data_sink(
|
|
264
|
-
data_sink_id="
|
|
264
|
+
data_sink_id="string",
|
|
265
265
|
)
|
|
266
266
|
"""
|
|
267
267
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -345,7 +345,7 @@ class AsyncDataSinksClient:
|
|
|
345
345
|
)
|
|
346
346
|
await client.data_sinks.create_data_sink(
|
|
347
347
|
request=DataSinkCreate(
|
|
348
|
-
name="
|
|
348
|
+
name="string",
|
|
349
349
|
sink_type=ConfigurableDataSinkNames.PINECONE,
|
|
350
350
|
),
|
|
351
351
|
)
|
|
@@ -394,7 +394,7 @@ class AsyncDataSinksClient:
|
|
|
394
394
|
)
|
|
395
395
|
await client.data_sinks.upsert_data_sink(
|
|
396
396
|
request=DataSinkCreate(
|
|
397
|
-
name="
|
|
397
|
+
name="string",
|
|
398
398
|
sink_type=ConfigurableDataSinkNames.PINECONE,
|
|
399
399
|
),
|
|
400
400
|
)
|
|
@@ -430,7 +430,7 @@ class AsyncDataSinksClient:
|
|
|
430
430
|
token="YOUR_TOKEN",
|
|
431
431
|
)
|
|
432
432
|
await client.data_sinks.get_data_sink(
|
|
433
|
-
data_sink_id="
|
|
433
|
+
data_sink_id="string",
|
|
434
434
|
)
|
|
435
435
|
"""
|
|
436
436
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -463,7 +463,7 @@ class AsyncDataSinksClient:
|
|
|
463
463
|
Parameters:
|
|
464
464
|
- data_sink_id: str.
|
|
465
465
|
|
|
466
|
-
- name: typing.Optional[str].
|
|
466
|
+
- name: typing.Optional[str].
|
|
467
467
|
|
|
468
468
|
- sink_type: ConfigurableDataSinkNames.
|
|
469
469
|
|
|
@@ -476,7 +476,7 @@ class AsyncDataSinksClient:
|
|
|
476
476
|
token="YOUR_TOKEN",
|
|
477
477
|
)
|
|
478
478
|
await client.data_sinks.update_data_sink(
|
|
479
|
-
data_sink_id="
|
|
479
|
+
data_sink_id="string",
|
|
480
480
|
sink_type=ConfigurableDataSinkNames.PINECONE,
|
|
481
481
|
)
|
|
482
482
|
"""
|
|
@@ -515,7 +515,7 @@ class AsyncDataSinksClient:
|
|
|
515
515
|
token="YOUR_TOKEN",
|
|
516
516
|
)
|
|
517
517
|
await client.data_sinks.delete_data_sink(
|
|
518
|
-
data_sink_id="
|
|
518
|
+
data_sink_id="string",
|
|
519
519
|
)
|
|
520
520
|
"""
|
|
521
521
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -93,7 +93,7 @@ class DataSourcesClient:
|
|
|
93
93
|
)
|
|
94
94
|
client.data_sources.create_data_source(
|
|
95
95
|
request=DataSourceCreate(
|
|
96
|
-
name="
|
|
96
|
+
name="string",
|
|
97
97
|
source_type=ConfigurableDataSourceNames.S_3,
|
|
98
98
|
),
|
|
99
99
|
)
|
|
@@ -142,7 +142,7 @@ class DataSourcesClient:
|
|
|
142
142
|
)
|
|
143
143
|
client.data_sources.upsert_data_source(
|
|
144
144
|
request=DataSourceCreate(
|
|
145
|
-
name="
|
|
145
|
+
name="string",
|
|
146
146
|
source_type=ConfigurableDataSourceNames.S_3,
|
|
147
147
|
),
|
|
148
148
|
)
|
|
@@ -178,7 +178,7 @@ class DataSourcesClient:
|
|
|
178
178
|
token="YOUR_TOKEN",
|
|
179
179
|
)
|
|
180
180
|
client.data_sources.get_data_source(
|
|
181
|
-
data_source_id="
|
|
181
|
+
data_source_id="string",
|
|
182
182
|
)
|
|
183
183
|
"""
|
|
184
184
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -212,11 +212,11 @@ class DataSourcesClient:
|
|
|
212
212
|
Parameters:
|
|
213
213
|
- data_source_id: str.
|
|
214
214
|
|
|
215
|
-
- name: typing.Optional[str].
|
|
215
|
+
- name: typing.Optional[str].
|
|
216
216
|
|
|
217
217
|
- source_type: ConfigurableDataSourceNames.
|
|
218
218
|
|
|
219
|
-
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]].
|
|
219
|
+
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]].
|
|
220
220
|
|
|
221
221
|
- component: typing.Optional[DataSourceUpdateComponent]. Component that implements the data source
|
|
222
222
|
---
|
|
@@ -227,7 +227,7 @@ class DataSourcesClient:
|
|
|
227
227
|
token="YOUR_TOKEN",
|
|
228
228
|
)
|
|
229
229
|
client.data_sources.update_data_source(
|
|
230
|
-
data_source_id="
|
|
230
|
+
data_source_id="string",
|
|
231
231
|
source_type=ConfigurableDataSourceNames.S_3,
|
|
232
232
|
)
|
|
233
233
|
"""
|
|
@@ -268,7 +268,7 @@ class DataSourcesClient:
|
|
|
268
268
|
token="YOUR_TOKEN",
|
|
269
269
|
)
|
|
270
270
|
client.data_sources.delete_data_source(
|
|
271
|
-
data_source_id="
|
|
271
|
+
data_source_id="string",
|
|
272
272
|
)
|
|
273
273
|
"""
|
|
274
274
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -353,7 +353,7 @@ class AsyncDataSourcesClient:
|
|
|
353
353
|
)
|
|
354
354
|
await client.data_sources.create_data_source(
|
|
355
355
|
request=DataSourceCreate(
|
|
356
|
-
name="
|
|
356
|
+
name="string",
|
|
357
357
|
source_type=ConfigurableDataSourceNames.S_3,
|
|
358
358
|
),
|
|
359
359
|
)
|
|
@@ -402,7 +402,7 @@ class AsyncDataSourcesClient:
|
|
|
402
402
|
)
|
|
403
403
|
await client.data_sources.upsert_data_source(
|
|
404
404
|
request=DataSourceCreate(
|
|
405
|
-
name="
|
|
405
|
+
name="string",
|
|
406
406
|
source_type=ConfigurableDataSourceNames.S_3,
|
|
407
407
|
),
|
|
408
408
|
)
|
|
@@ -438,7 +438,7 @@ class AsyncDataSourcesClient:
|
|
|
438
438
|
token="YOUR_TOKEN",
|
|
439
439
|
)
|
|
440
440
|
await client.data_sources.get_data_source(
|
|
441
|
-
data_source_id="
|
|
441
|
+
data_source_id="string",
|
|
442
442
|
)
|
|
443
443
|
"""
|
|
444
444
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -472,11 +472,11 @@ class AsyncDataSourcesClient:
|
|
|
472
472
|
Parameters:
|
|
473
473
|
- data_source_id: str.
|
|
474
474
|
|
|
475
|
-
- name: typing.Optional[str].
|
|
475
|
+
- name: typing.Optional[str].
|
|
476
476
|
|
|
477
477
|
- source_type: ConfigurableDataSourceNames.
|
|
478
478
|
|
|
479
|
-
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]].
|
|
479
|
+
- custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]].
|
|
480
480
|
|
|
481
481
|
- component: typing.Optional[DataSourceUpdateComponent]. Component that implements the data source
|
|
482
482
|
---
|
|
@@ -487,7 +487,7 @@ class AsyncDataSourcesClient:
|
|
|
487
487
|
token="YOUR_TOKEN",
|
|
488
488
|
)
|
|
489
489
|
await client.data_sources.update_data_source(
|
|
490
|
-
data_source_id="
|
|
490
|
+
data_source_id="string",
|
|
491
491
|
source_type=ConfigurableDataSourceNames.S_3,
|
|
492
492
|
)
|
|
493
493
|
"""
|
|
@@ -528,7 +528,7 @@ class AsyncDataSourcesClient:
|
|
|
528
528
|
token="YOUR_TOKEN",
|
|
529
529
|
)
|
|
530
530
|
await client.data_sources.delete_data_source(
|
|
531
|
-
data_source_id="
|
|
531
|
+
data_source_id="string",
|
|
532
532
|
)
|
|
533
533
|
"""
|
|
534
534
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -34,6 +34,15 @@ class EmbeddingModelConfigsClient:
|
|
|
34
34
|
"""
|
|
35
35
|
Parameters:
|
|
36
36
|
- project_id: str.
|
|
37
|
+
---
|
|
38
|
+
from llama_cloud.client import LlamaCloud
|
|
39
|
+
|
|
40
|
+
client = LlamaCloud(
|
|
41
|
+
token="YOUR_TOKEN",
|
|
42
|
+
)
|
|
43
|
+
client.embedding_model_configs.list_embedding_model_configs(
|
|
44
|
+
project_id="string",
|
|
45
|
+
)
|
|
37
46
|
"""
|
|
38
47
|
_response = self._client_wrapper.httpx_client.request(
|
|
39
48
|
"GET",
|
|
@@ -64,22 +73,6 @@ class EmbeddingModelConfigsClient:
|
|
|
64
73
|
- name: str. The name of the embedding model config.
|
|
65
74
|
|
|
66
75
|
- embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
|
|
67
|
-
---
|
|
68
|
-
from llama_cloud import (
|
|
69
|
-
EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding,
|
|
70
|
-
)
|
|
71
|
-
from llama_cloud.client import LlamaCloud
|
|
72
|
-
|
|
73
|
-
client = LlamaCloud(
|
|
74
|
-
token="YOUR_TOKEN",
|
|
75
|
-
)
|
|
76
|
-
client.embedding_model_configs.create_embedding_model_config(
|
|
77
|
-
project_id="project_id",
|
|
78
|
-
name="name",
|
|
79
|
-
embedding_config=EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding(
|
|
80
|
-
type="VERTEXAI_EMBEDDING",
|
|
81
|
-
),
|
|
82
|
-
)
|
|
83
76
|
"""
|
|
84
77
|
_response = self._client_wrapper.httpx_client.request(
|
|
85
78
|
"POST",
|
|
@@ -116,16 +109,6 @@ class EmbeddingModelConfigsClient:
|
|
|
116
109
|
- organization_id: typing.Optional[str].
|
|
117
110
|
|
|
118
111
|
- request: EmbeddingModelConfigUpdate.
|
|
119
|
-
---
|
|
120
|
-
from llama_cloud import EmbeddingModelConfigUpdate
|
|
121
|
-
from llama_cloud.client import LlamaCloud
|
|
122
|
-
|
|
123
|
-
client = LlamaCloud(
|
|
124
|
-
token="YOUR_TOKEN",
|
|
125
|
-
)
|
|
126
|
-
client.embedding_model_configs.upsert_embedding_model_config(
|
|
127
|
-
request=EmbeddingModelConfigUpdate(),
|
|
128
|
-
)
|
|
129
112
|
"""
|
|
130
113
|
_response = self._client_wrapper.httpx_client.request(
|
|
131
114
|
"PUT",
|
|
@@ -155,17 +138,6 @@ class EmbeddingModelConfigsClient:
|
|
|
155
138
|
- embedding_model_config_id: str.
|
|
156
139
|
|
|
157
140
|
- request: EmbeddingModelConfigUpdate.
|
|
158
|
-
---
|
|
159
|
-
from llama_cloud import EmbeddingModelConfigUpdate
|
|
160
|
-
from llama_cloud.client import LlamaCloud
|
|
161
|
-
|
|
162
|
-
client = LlamaCloud(
|
|
163
|
-
token="YOUR_TOKEN",
|
|
164
|
-
)
|
|
165
|
-
client.embedding_model_configs.update_embedding_model_config(
|
|
166
|
-
embedding_model_config_id="embedding_model_config_id",
|
|
167
|
-
request=EmbeddingModelConfigUpdate(),
|
|
168
|
-
)
|
|
169
141
|
"""
|
|
170
142
|
_response = self._client_wrapper.httpx_client.request(
|
|
171
143
|
"PUT",
|
|
@@ -199,7 +171,7 @@ class EmbeddingModelConfigsClient:
|
|
|
199
171
|
token="YOUR_TOKEN",
|
|
200
172
|
)
|
|
201
173
|
client.embedding_model_configs.delete_embedding_model_config(
|
|
202
|
-
embedding_model_config_id="
|
|
174
|
+
embedding_model_config_id="string",
|
|
203
175
|
)
|
|
204
176
|
"""
|
|
205
177
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -229,6 +201,15 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
229
201
|
"""
|
|
230
202
|
Parameters:
|
|
231
203
|
- project_id: str.
|
|
204
|
+
---
|
|
205
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
206
|
+
|
|
207
|
+
client = AsyncLlamaCloud(
|
|
208
|
+
token="YOUR_TOKEN",
|
|
209
|
+
)
|
|
210
|
+
await client.embedding_model_configs.list_embedding_model_configs(
|
|
211
|
+
project_id="string",
|
|
212
|
+
)
|
|
232
213
|
"""
|
|
233
214
|
_response = await self._client_wrapper.httpx_client.request(
|
|
234
215
|
"GET",
|
|
@@ -259,22 +240,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
259
240
|
- name: str. The name of the embedding model config.
|
|
260
241
|
|
|
261
242
|
- embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
|
|
262
|
-
---
|
|
263
|
-
from llama_cloud import (
|
|
264
|
-
EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding,
|
|
265
|
-
)
|
|
266
|
-
from llama_cloud.client import AsyncLlamaCloud
|
|
267
|
-
|
|
268
|
-
client = AsyncLlamaCloud(
|
|
269
|
-
token="YOUR_TOKEN",
|
|
270
|
-
)
|
|
271
|
-
await client.embedding_model_configs.create_embedding_model_config(
|
|
272
|
-
project_id="project_id",
|
|
273
|
-
name="name",
|
|
274
|
-
embedding_config=EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding(
|
|
275
|
-
type="VERTEXAI_EMBEDDING",
|
|
276
|
-
),
|
|
277
|
-
)
|
|
278
243
|
"""
|
|
279
244
|
_response = await self._client_wrapper.httpx_client.request(
|
|
280
245
|
"POST",
|
|
@@ -311,16 +276,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
311
276
|
- organization_id: typing.Optional[str].
|
|
312
277
|
|
|
313
278
|
- request: EmbeddingModelConfigUpdate.
|
|
314
|
-
---
|
|
315
|
-
from llama_cloud import EmbeddingModelConfigUpdate
|
|
316
|
-
from llama_cloud.client import AsyncLlamaCloud
|
|
317
|
-
|
|
318
|
-
client = AsyncLlamaCloud(
|
|
319
|
-
token="YOUR_TOKEN",
|
|
320
|
-
)
|
|
321
|
-
await client.embedding_model_configs.upsert_embedding_model_config(
|
|
322
|
-
request=EmbeddingModelConfigUpdate(),
|
|
323
|
-
)
|
|
324
279
|
"""
|
|
325
280
|
_response = await self._client_wrapper.httpx_client.request(
|
|
326
281
|
"PUT",
|
|
@@ -350,17 +305,6 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
350
305
|
- embedding_model_config_id: str.
|
|
351
306
|
|
|
352
307
|
- request: EmbeddingModelConfigUpdate.
|
|
353
|
-
---
|
|
354
|
-
from llama_cloud import EmbeddingModelConfigUpdate
|
|
355
|
-
from llama_cloud.client import AsyncLlamaCloud
|
|
356
|
-
|
|
357
|
-
client = AsyncLlamaCloud(
|
|
358
|
-
token="YOUR_TOKEN",
|
|
359
|
-
)
|
|
360
|
-
await client.embedding_model_configs.update_embedding_model_config(
|
|
361
|
-
embedding_model_config_id="embedding_model_config_id",
|
|
362
|
-
request=EmbeddingModelConfigUpdate(),
|
|
363
|
-
)
|
|
364
308
|
"""
|
|
365
309
|
_response = await self._client_wrapper.httpx_client.request(
|
|
366
310
|
"PUT",
|
|
@@ -394,7 +338,7 @@ class AsyncEmbeddingModelConfigsClient:
|
|
|
394
338
|
token="YOUR_TOKEN",
|
|
395
339
|
)
|
|
396
340
|
await client.embedding_model_configs.delete_embedding_model_config(
|
|
397
|
-
embedding_model_config_id="
|
|
341
|
+
embedding_model_config_id="string",
|
|
398
342
|
)
|
|
399
343
|
"""
|
|
400
344
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -43,7 +43,7 @@ class EvalsClient:
|
|
|
43
43
|
token="YOUR_TOKEN",
|
|
44
44
|
)
|
|
45
45
|
client.evals.get_dataset(
|
|
46
|
-
dataset_id="
|
|
46
|
+
dataset_id="string",
|
|
47
47
|
)
|
|
48
48
|
"""
|
|
49
49
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -77,8 +77,8 @@ class EvalsClient:
|
|
|
77
77
|
token="YOUR_TOKEN",
|
|
78
78
|
)
|
|
79
79
|
client.evals.update_dataset(
|
|
80
|
-
dataset_id="
|
|
81
|
-
name="
|
|
80
|
+
dataset_id="string",
|
|
81
|
+
name="string",
|
|
82
82
|
)
|
|
83
83
|
"""
|
|
84
84
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -111,7 +111,7 @@ class EvalsClient:
|
|
|
111
111
|
token="YOUR_TOKEN",
|
|
112
112
|
)
|
|
113
113
|
client.evals.delete_dataset(
|
|
114
|
-
dataset_id="
|
|
114
|
+
dataset_id="string",
|
|
115
115
|
)
|
|
116
116
|
"""
|
|
117
117
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -143,7 +143,7 @@ class EvalsClient:
|
|
|
143
143
|
token="YOUR_TOKEN",
|
|
144
144
|
)
|
|
145
145
|
client.evals.list_questions(
|
|
146
|
-
dataset_id="
|
|
146
|
+
dataset_id="string",
|
|
147
147
|
)
|
|
148
148
|
"""
|
|
149
149
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -180,9 +180,9 @@ class EvalsClient:
|
|
|
180
180
|
token="YOUR_TOKEN",
|
|
181
181
|
)
|
|
182
182
|
client.evals.create_question(
|
|
183
|
-
dataset_id="
|
|
183
|
+
dataset_id="string",
|
|
184
184
|
request=EvalQuestionCreate(
|
|
185
|
-
content="
|
|
185
|
+
content="string",
|
|
186
186
|
),
|
|
187
187
|
)
|
|
188
188
|
"""
|
|
@@ -216,19 +216,14 @@ class EvalsClient:
|
|
|
216
216
|
|
|
217
217
|
- request: typing.List[EvalQuestionCreate].
|
|
218
218
|
---
|
|
219
|
-
from llama_cloud import EvalQuestionCreate
|
|
220
219
|
from llama_cloud.client import LlamaCloud
|
|
221
220
|
|
|
222
221
|
client = LlamaCloud(
|
|
223
222
|
token="YOUR_TOKEN",
|
|
224
223
|
)
|
|
225
224
|
client.evals.create_questions(
|
|
226
|
-
dataset_id="
|
|
227
|
-
request=[
|
|
228
|
-
EvalQuestionCreate(
|
|
229
|
-
content="content",
|
|
230
|
-
)
|
|
231
|
-
],
|
|
225
|
+
dataset_id="string",
|
|
226
|
+
request=[],
|
|
232
227
|
)
|
|
233
228
|
"""
|
|
234
229
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -263,7 +258,7 @@ class EvalsClient:
|
|
|
263
258
|
token="YOUR_TOKEN",
|
|
264
259
|
)
|
|
265
260
|
client.evals.get_question(
|
|
266
|
-
question_id="
|
|
261
|
+
question_id="string",
|
|
267
262
|
)
|
|
268
263
|
"""
|
|
269
264
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -298,9 +293,9 @@ class EvalsClient:
|
|
|
298
293
|
token="YOUR_TOKEN",
|
|
299
294
|
)
|
|
300
295
|
client.evals.replace_question(
|
|
301
|
-
question_id="
|
|
296
|
+
question_id="string",
|
|
302
297
|
request=EvalQuestionCreate(
|
|
303
|
-
content="
|
|
298
|
+
content="string",
|
|
304
299
|
),
|
|
305
300
|
)
|
|
306
301
|
"""
|
|
@@ -334,7 +329,7 @@ class EvalsClient:
|
|
|
334
329
|
token="YOUR_TOKEN",
|
|
335
330
|
)
|
|
336
331
|
client.evals.delete_question(
|
|
337
|
-
question_id="
|
|
332
|
+
question_id="string",
|
|
338
333
|
)
|
|
339
334
|
"""
|
|
340
335
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -399,7 +394,7 @@ class AsyncEvalsClient:
|
|
|
399
394
|
token="YOUR_TOKEN",
|
|
400
395
|
)
|
|
401
396
|
await client.evals.get_dataset(
|
|
402
|
-
dataset_id="
|
|
397
|
+
dataset_id="string",
|
|
403
398
|
)
|
|
404
399
|
"""
|
|
405
400
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -433,8 +428,8 @@ class AsyncEvalsClient:
|
|
|
433
428
|
token="YOUR_TOKEN",
|
|
434
429
|
)
|
|
435
430
|
await client.evals.update_dataset(
|
|
436
|
-
dataset_id="
|
|
437
|
-
name="
|
|
431
|
+
dataset_id="string",
|
|
432
|
+
name="string",
|
|
438
433
|
)
|
|
439
434
|
"""
|
|
440
435
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -467,7 +462,7 @@ class AsyncEvalsClient:
|
|
|
467
462
|
token="YOUR_TOKEN",
|
|
468
463
|
)
|
|
469
464
|
await client.evals.delete_dataset(
|
|
470
|
-
dataset_id="
|
|
465
|
+
dataset_id="string",
|
|
471
466
|
)
|
|
472
467
|
"""
|
|
473
468
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -499,7 +494,7 @@ class AsyncEvalsClient:
|
|
|
499
494
|
token="YOUR_TOKEN",
|
|
500
495
|
)
|
|
501
496
|
await client.evals.list_questions(
|
|
502
|
-
dataset_id="
|
|
497
|
+
dataset_id="string",
|
|
503
498
|
)
|
|
504
499
|
"""
|
|
505
500
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -536,9 +531,9 @@ class AsyncEvalsClient:
|
|
|
536
531
|
token="YOUR_TOKEN",
|
|
537
532
|
)
|
|
538
533
|
await client.evals.create_question(
|
|
539
|
-
dataset_id="
|
|
534
|
+
dataset_id="string",
|
|
540
535
|
request=EvalQuestionCreate(
|
|
541
|
-
content="
|
|
536
|
+
content="string",
|
|
542
537
|
),
|
|
543
538
|
)
|
|
544
539
|
"""
|
|
@@ -572,19 +567,14 @@ class AsyncEvalsClient:
|
|
|
572
567
|
|
|
573
568
|
- request: typing.List[EvalQuestionCreate].
|
|
574
569
|
---
|
|
575
|
-
from llama_cloud import EvalQuestionCreate
|
|
576
570
|
from llama_cloud.client import AsyncLlamaCloud
|
|
577
571
|
|
|
578
572
|
client = AsyncLlamaCloud(
|
|
579
573
|
token="YOUR_TOKEN",
|
|
580
574
|
)
|
|
581
575
|
await client.evals.create_questions(
|
|
582
|
-
dataset_id="
|
|
583
|
-
request=[
|
|
584
|
-
EvalQuestionCreate(
|
|
585
|
-
content="content",
|
|
586
|
-
)
|
|
587
|
-
],
|
|
576
|
+
dataset_id="string",
|
|
577
|
+
request=[],
|
|
588
578
|
)
|
|
589
579
|
"""
|
|
590
580
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -619,7 +609,7 @@ class AsyncEvalsClient:
|
|
|
619
609
|
token="YOUR_TOKEN",
|
|
620
610
|
)
|
|
621
611
|
await client.evals.get_question(
|
|
622
|
-
question_id="
|
|
612
|
+
question_id="string",
|
|
623
613
|
)
|
|
624
614
|
"""
|
|
625
615
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -654,9 +644,9 @@ class AsyncEvalsClient:
|
|
|
654
644
|
token="YOUR_TOKEN",
|
|
655
645
|
)
|
|
656
646
|
await client.evals.replace_question(
|
|
657
|
-
question_id="
|
|
647
|
+
question_id="string",
|
|
658
648
|
request=EvalQuestionCreate(
|
|
659
|
-
content="
|
|
649
|
+
content="string",
|
|
660
650
|
),
|
|
661
651
|
)
|
|
662
652
|
"""
|
|
@@ -690,7 +680,7 @@ class AsyncEvalsClient:
|
|
|
690
680
|
token="YOUR_TOKEN",
|
|
691
681
|
)
|
|
692
682
|
await client.evals.delete_question(
|
|
693
|
-
question_id="
|
|
683
|
+
question_id="string",
|
|
694
684
|
)
|
|
695
685
|
"""
|
|
696
686
|
_response = await self._client_wrapper.httpx_client.request(
|