llama-cloud 0.1.6__py3-none-any.whl → 0.1.7__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (173) hide show
  1. llama_cloud/__init__.py +140 -6
  2. llama_cloud/client.py +15 -0
  3. llama_cloud/environment.py +1 -1
  4. llama_cloud/resources/__init__.py +15 -0
  5. llama_cloud/{types/token.py → resources/chat_apps/__init__.py} +0 -3
  6. llama_cloud/resources/chat_apps/client.py +630 -0
  7. llama_cloud/resources/data_sinks/client.py +12 -12
  8. llama_cloud/resources/data_sources/client.py +14 -14
  9. llama_cloud/resources/embedding_model_configs/client.py +20 -76
  10. llama_cloud/resources/evals/client.py +26 -36
  11. llama_cloud/resources/extraction/client.py +32 -32
  12. llama_cloud/resources/files/client.py +40 -44
  13. llama_cloud/resources/jobs/__init__.py +2 -0
  14. llama_cloud/resources/jobs/client.py +148 -0
  15. llama_cloud/resources/llama_extract/__init__.py +5 -0
  16. llama_cloud/resources/llama_extract/client.py +1038 -0
  17. llama_cloud/resources/llama_extract/types/__init__.py +6 -0
  18. llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_value.py +7 -0
  19. llama_cloud/resources/llama_extract/types/extract_agent_update_data_schema_value.py +7 -0
  20. llama_cloud/resources/organizations/client.py +66 -70
  21. llama_cloud/resources/parsing/client.py +448 -428
  22. llama_cloud/resources/pipelines/client.py +256 -344
  23. llama_cloud/resources/projects/client.py +34 -60
  24. llama_cloud/resources/reports/__init__.py +5 -0
  25. llama_cloud/resources/reports/client.py +1198 -0
  26. llama_cloud/resources/reports/types/__init__.py +7 -0
  27. llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +25 -0
  28. llama_cloud/resources/retrievers/__init__.py +2 -0
  29. llama_cloud/resources/retrievers/client.py +654 -0
  30. llama_cloud/types/__init__.py +128 -6
  31. llama_cloud/types/{chat_message.py → app_schema_chat_chat_message.py} +3 -3
  32. llama_cloud/types/azure_open_ai_embedding.py +6 -12
  33. llama_cloud/types/base_prompt_template.py +2 -6
  34. llama_cloud/types/bedrock_embedding.py +6 -12
  35. llama_cloud/types/character_splitter.py +2 -4
  36. llama_cloud/types/chat_app.py +44 -0
  37. llama_cloud/types/chat_app_response.py +41 -0
  38. llama_cloud/types/cloud_az_storage_blob_data_source.py +7 -15
  39. llama_cloud/types/cloud_box_data_source.py +6 -12
  40. llama_cloud/types/cloud_confluence_data_source.py +6 -6
  41. llama_cloud/types/cloud_document.py +1 -3
  42. llama_cloud/types/cloud_document_create.py +1 -3
  43. llama_cloud/types/cloud_jira_data_source.py +4 -6
  44. llama_cloud/types/cloud_notion_page_data_source.py +2 -2
  45. llama_cloud/types/cloud_one_drive_data_source.py +3 -5
  46. llama_cloud/types/cloud_postgres_vector_store.py +1 -0
  47. llama_cloud/types/cloud_s_3_data_source.py +4 -8
  48. llama_cloud/types/cloud_sharepoint_data_source.py +6 -8
  49. llama_cloud/types/cloud_slack_data_source.py +6 -6
  50. llama_cloud/types/code_splitter.py +1 -1
  51. llama_cloud/types/cohere_embedding.py +3 -7
  52. llama_cloud/types/composite_retrieval_mode.py +21 -0
  53. llama_cloud/types/composite_retrieval_result.py +38 -0
  54. llama_cloud/types/composite_retrieved_text_node.py +42 -0
  55. llama_cloud/types/data_sink.py +4 -4
  56. llama_cloud/types/data_sink_component.py +20 -0
  57. llama_cloud/types/data_source.py +5 -7
  58. llama_cloud/types/data_source_component.py +28 -0
  59. llama_cloud/types/data_source_create.py +1 -3
  60. llama_cloud/types/edit_suggestion.py +39 -0
  61. llama_cloud/types/embedding_model_config.py +2 -2
  62. llama_cloud/types/embedding_model_config_update.py +2 -4
  63. llama_cloud/types/eval_dataset.py +2 -2
  64. llama_cloud/types/eval_dataset_job_record.py +8 -13
  65. llama_cloud/types/eval_execution_params_override.py +2 -6
  66. llama_cloud/types/eval_question.py +2 -2
  67. llama_cloud/types/extract_agent.py +45 -0
  68. llama_cloud/types/extract_agent_data_schema_value.py +5 -0
  69. llama_cloud/types/extract_config.py +40 -0
  70. llama_cloud/types/extract_job.py +35 -0
  71. llama_cloud/types/extract_job_create.py +40 -0
  72. llama_cloud/types/extract_job_create_data_schema_override_value.py +7 -0
  73. llama_cloud/types/extract_mode.py +17 -0
  74. llama_cloud/types/extract_resultset.py +46 -0
  75. llama_cloud/types/extract_resultset_data.py +11 -0
  76. llama_cloud/types/extract_resultset_data_item_value.py +7 -0
  77. llama_cloud/types/extract_resultset_data_zero_value.py +7 -0
  78. llama_cloud/types/extract_resultset_extraction_metadata_value.py +7 -0
  79. llama_cloud/types/extraction_result.py +2 -2
  80. llama_cloud/types/extraction_schema.py +3 -5
  81. llama_cloud/types/file.py +9 -14
  82. llama_cloud/types/filter_condition.py +9 -1
  83. llama_cloud/types/filter_operator.py +6 -2
  84. llama_cloud/types/gemini_embedding.py +6 -10
  85. llama_cloud/types/hugging_face_inference_api_embedding.py +11 -27
  86. llama_cloud/types/hugging_face_inference_api_embedding_token.py +5 -0
  87. llama_cloud/types/image_block.py +35 -0
  88. llama_cloud/types/input_message.py +2 -4
  89. llama_cloud/types/job_names.py +89 -0
  90. llama_cloud/types/job_record.py +57 -0
  91. llama_cloud/types/job_record_with_usage_metrics.py +36 -0
  92. llama_cloud/types/llama_index_core_base_llms_types_chat_message.py +39 -0
  93. llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +33 -0
  94. llama_cloud/types/llama_parse_parameters.py +4 -0
  95. llama_cloud/types/llm.py +3 -4
  96. llama_cloud/types/llm_model_data.py +1 -0
  97. llama_cloud/types/llm_parameters.py +3 -5
  98. llama_cloud/types/local_eval.py +8 -10
  99. llama_cloud/types/local_eval_results.py +1 -1
  100. llama_cloud/types/managed_ingestion_status.py +4 -0
  101. llama_cloud/types/managed_ingestion_status_response.py +4 -5
  102. llama_cloud/types/markdown_element_node_parser.py +3 -5
  103. llama_cloud/types/markdown_node_parser.py +1 -1
  104. llama_cloud/types/metadata_filter.py +2 -2
  105. llama_cloud/types/metadata_filter_value.py +5 -0
  106. llama_cloud/types/metric_result.py +3 -3
  107. llama_cloud/types/node_parser.py +1 -1
  108. llama_cloud/types/object_type.py +4 -0
  109. llama_cloud/types/open_ai_embedding.py +6 -12
  110. llama_cloud/types/organization.py +7 -2
  111. llama_cloud/types/page_splitter_node_parser.py +2 -2
  112. llama_cloud/types/paginated_jobs_history_with_metrics.py +35 -0
  113. llama_cloud/types/paginated_report_response.py +35 -0
  114. llama_cloud/types/parse_plan_level.py +21 -0
  115. llama_cloud/types/permission.py +3 -3
  116. llama_cloud/types/pipeline.py +7 -17
  117. llama_cloud/types/pipeline_configuration_hashes.py +3 -3
  118. llama_cloud/types/pipeline_create.py +8 -16
  119. llama_cloud/types/pipeline_data_source.py +7 -13
  120. llama_cloud/types/pipeline_data_source_component.py +28 -0
  121. llama_cloud/types/pipeline_data_source_create.py +1 -3
  122. llama_cloud/types/pipeline_deployment.py +4 -4
  123. llama_cloud/types/pipeline_file.py +13 -24
  124. llama_cloud/types/pipeline_file_create.py +1 -3
  125. llama_cloud/types/playground_session.py +4 -4
  126. llama_cloud/types/preset_retrieval_params.py +8 -14
  127. llama_cloud/types/presigned_url.py +1 -3
  128. llama_cloud/types/progress_event.py +44 -0
  129. llama_cloud/types/progress_event_status.py +33 -0
  130. llama_cloud/types/project.py +2 -2
  131. llama_cloud/types/prompt_mixin_prompts.py +1 -1
  132. llama_cloud/types/prompt_spec.py +3 -5
  133. llama_cloud/types/related_node_info.py +2 -2
  134. llama_cloud/types/related_node_info_node_type.py +7 -0
  135. llama_cloud/types/report.py +33 -0
  136. llama_cloud/types/report_block.py +34 -0
  137. llama_cloud/types/report_block_dependency.py +29 -0
  138. llama_cloud/types/report_create_response.py +31 -0
  139. llama_cloud/types/report_event_item.py +40 -0
  140. llama_cloud/types/report_event_item_event_data.py +45 -0
  141. llama_cloud/types/report_event_type.py +37 -0
  142. llama_cloud/types/report_metadata.py +43 -0
  143. llama_cloud/types/report_plan.py +36 -0
  144. llama_cloud/types/report_plan_block.py +36 -0
  145. llama_cloud/types/report_query.py +33 -0
  146. llama_cloud/types/report_response.py +41 -0
  147. llama_cloud/types/report_state.py +37 -0
  148. llama_cloud/types/report_state_event.py +38 -0
  149. llama_cloud/types/report_update_event.py +38 -0
  150. llama_cloud/types/retrieve_results.py +1 -1
  151. llama_cloud/types/retriever.py +45 -0
  152. llama_cloud/types/retriever_create.py +37 -0
  153. llama_cloud/types/retriever_pipeline.py +37 -0
  154. llama_cloud/types/role.py +3 -3
  155. llama_cloud/types/sentence_splitter.py +2 -4
  156. llama_cloud/types/status_enum.py +4 -0
  157. llama_cloud/types/supported_llm_model_names.py +4 -0
  158. llama_cloud/types/text_block.py +31 -0
  159. llama_cloud/types/text_node.py +15 -8
  160. llama_cloud/types/token_text_splitter.py +1 -1
  161. llama_cloud/types/usage_metric_response.py +34 -0
  162. llama_cloud/types/user_job_record.py +32 -0
  163. llama_cloud/types/user_organization.py +5 -9
  164. llama_cloud/types/user_organization_create.py +4 -4
  165. llama_cloud/types/user_organization_delete.py +2 -2
  166. llama_cloud/types/user_organization_role.py +2 -2
  167. llama_cloud/types/vertex_text_embedding.py +5 -9
  168. {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7.dist-info}/METADATA +2 -1
  169. llama_cloud-0.1.7.dist-info/RECORD +310 -0
  170. llama_cloud/types/value.py +0 -5
  171. llama_cloud-0.1.6.dist-info/RECORD +0 -241
  172. {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7.dist-info}/LICENSE +0 -0
  173. {llama_cloud-0.1.6.dist-info → llama_cloud-0.1.7.dist-info}/WHEEL +0 -0
@@ -91,7 +91,7 @@ class DataSinksClient:
91
91
  )
92
92
  client.data_sinks.create_data_sink(
93
93
  request=DataSinkCreate(
94
- name="name",
94
+ name="string",
95
95
  sink_type=ConfigurableDataSinkNames.PINECONE,
96
96
  ),
97
97
  )
@@ -140,7 +140,7 @@ class DataSinksClient:
140
140
  )
141
141
  client.data_sinks.upsert_data_sink(
142
142
  request=DataSinkCreate(
143
- name="name",
143
+ name="string",
144
144
  sink_type=ConfigurableDataSinkNames.PINECONE,
145
145
  ),
146
146
  )
@@ -176,7 +176,7 @@ class DataSinksClient:
176
176
  token="YOUR_TOKEN",
177
177
  )
178
178
  client.data_sinks.get_data_sink(
179
- data_sink_id="data_sink_id",
179
+ data_sink_id="string",
180
180
  )
181
181
  """
182
182
  _response = self._client_wrapper.httpx_client.request(
@@ -209,7 +209,7 @@ class DataSinksClient:
209
209
  Parameters:
210
210
  - data_sink_id: str.
211
211
 
212
- - name: typing.Optional[str]. The name of the data sink.
212
+ - name: typing.Optional[str].
213
213
 
214
214
  - sink_type: ConfigurableDataSinkNames.
215
215
 
@@ -222,7 +222,7 @@ class DataSinksClient:
222
222
  token="YOUR_TOKEN",
223
223
  )
224
224
  client.data_sinks.update_data_sink(
225
- data_sink_id="data_sink_id",
225
+ data_sink_id="string",
226
226
  sink_type=ConfigurableDataSinkNames.PINECONE,
227
227
  )
228
228
  """
@@ -261,7 +261,7 @@ class DataSinksClient:
261
261
  token="YOUR_TOKEN",
262
262
  )
263
263
  client.data_sinks.delete_data_sink(
264
- data_sink_id="data_sink_id",
264
+ data_sink_id="string",
265
265
  )
266
266
  """
267
267
  _response = self._client_wrapper.httpx_client.request(
@@ -345,7 +345,7 @@ class AsyncDataSinksClient:
345
345
  )
346
346
  await client.data_sinks.create_data_sink(
347
347
  request=DataSinkCreate(
348
- name="name",
348
+ name="string",
349
349
  sink_type=ConfigurableDataSinkNames.PINECONE,
350
350
  ),
351
351
  )
@@ -394,7 +394,7 @@ class AsyncDataSinksClient:
394
394
  )
395
395
  await client.data_sinks.upsert_data_sink(
396
396
  request=DataSinkCreate(
397
- name="name",
397
+ name="string",
398
398
  sink_type=ConfigurableDataSinkNames.PINECONE,
399
399
  ),
400
400
  )
@@ -430,7 +430,7 @@ class AsyncDataSinksClient:
430
430
  token="YOUR_TOKEN",
431
431
  )
432
432
  await client.data_sinks.get_data_sink(
433
- data_sink_id="data_sink_id",
433
+ data_sink_id="string",
434
434
  )
435
435
  """
436
436
  _response = await self._client_wrapper.httpx_client.request(
@@ -463,7 +463,7 @@ class AsyncDataSinksClient:
463
463
  Parameters:
464
464
  - data_sink_id: str.
465
465
 
466
- - name: typing.Optional[str]. The name of the data sink.
466
+ - name: typing.Optional[str].
467
467
 
468
468
  - sink_type: ConfigurableDataSinkNames.
469
469
 
@@ -476,7 +476,7 @@ class AsyncDataSinksClient:
476
476
  token="YOUR_TOKEN",
477
477
  )
478
478
  await client.data_sinks.update_data_sink(
479
- data_sink_id="data_sink_id",
479
+ data_sink_id="string",
480
480
  sink_type=ConfigurableDataSinkNames.PINECONE,
481
481
  )
482
482
  """
@@ -515,7 +515,7 @@ class AsyncDataSinksClient:
515
515
  token="YOUR_TOKEN",
516
516
  )
517
517
  await client.data_sinks.delete_data_sink(
518
- data_sink_id="data_sink_id",
518
+ data_sink_id="string",
519
519
  )
520
520
  """
521
521
  _response = await self._client_wrapper.httpx_client.request(
@@ -93,7 +93,7 @@ class DataSourcesClient:
93
93
  )
94
94
  client.data_sources.create_data_source(
95
95
  request=DataSourceCreate(
96
- name="name",
96
+ name="string",
97
97
  source_type=ConfigurableDataSourceNames.S_3,
98
98
  ),
99
99
  )
@@ -142,7 +142,7 @@ class DataSourcesClient:
142
142
  )
143
143
  client.data_sources.upsert_data_source(
144
144
  request=DataSourceCreate(
145
- name="name",
145
+ name="string",
146
146
  source_type=ConfigurableDataSourceNames.S_3,
147
147
  ),
148
148
  )
@@ -178,7 +178,7 @@ class DataSourcesClient:
178
178
  token="YOUR_TOKEN",
179
179
  )
180
180
  client.data_sources.get_data_source(
181
- data_source_id="data_source_id",
181
+ data_source_id="string",
182
182
  )
183
183
  """
184
184
  _response = self._client_wrapper.httpx_client.request(
@@ -212,11 +212,11 @@ class DataSourcesClient:
212
212
  Parameters:
213
213
  - data_source_id: str.
214
214
 
215
- - name: typing.Optional[str]. The name of the data source.
215
+ - name: typing.Optional[str].
216
216
 
217
217
  - source_type: ConfigurableDataSourceNames.
218
218
 
219
- - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]]. Custom metadata that will be present on all data loaded from the data source
219
+ - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]].
220
220
 
221
221
  - component: typing.Optional[DataSourceUpdateComponent]. Component that implements the data source
222
222
  ---
@@ -227,7 +227,7 @@ class DataSourcesClient:
227
227
  token="YOUR_TOKEN",
228
228
  )
229
229
  client.data_sources.update_data_source(
230
- data_source_id="data_source_id",
230
+ data_source_id="string",
231
231
  source_type=ConfigurableDataSourceNames.S_3,
232
232
  )
233
233
  """
@@ -268,7 +268,7 @@ class DataSourcesClient:
268
268
  token="YOUR_TOKEN",
269
269
  )
270
270
  client.data_sources.delete_data_source(
271
- data_source_id="data_source_id",
271
+ data_source_id="string",
272
272
  )
273
273
  """
274
274
  _response = self._client_wrapper.httpx_client.request(
@@ -353,7 +353,7 @@ class AsyncDataSourcesClient:
353
353
  )
354
354
  await client.data_sources.create_data_source(
355
355
  request=DataSourceCreate(
356
- name="name",
356
+ name="string",
357
357
  source_type=ConfigurableDataSourceNames.S_3,
358
358
  ),
359
359
  )
@@ -402,7 +402,7 @@ class AsyncDataSourcesClient:
402
402
  )
403
403
  await client.data_sources.upsert_data_source(
404
404
  request=DataSourceCreate(
405
- name="name",
405
+ name="string",
406
406
  source_type=ConfigurableDataSourceNames.S_3,
407
407
  ),
408
408
  )
@@ -438,7 +438,7 @@ class AsyncDataSourcesClient:
438
438
  token="YOUR_TOKEN",
439
439
  )
440
440
  await client.data_sources.get_data_source(
441
- data_source_id="data_source_id",
441
+ data_source_id="string",
442
442
  )
443
443
  """
444
444
  _response = await self._client_wrapper.httpx_client.request(
@@ -472,11 +472,11 @@ class AsyncDataSourcesClient:
472
472
  Parameters:
473
473
  - data_source_id: str.
474
474
 
475
- - name: typing.Optional[str]. The name of the data source.
475
+ - name: typing.Optional[str].
476
476
 
477
477
  - source_type: ConfigurableDataSourceNames.
478
478
 
479
- - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]]. Custom metadata that will be present on all data loaded from the data source
479
+ - custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[DataSourceUpdateCustomMetadataValue]]].
480
480
 
481
481
  - component: typing.Optional[DataSourceUpdateComponent]. Component that implements the data source
482
482
  ---
@@ -487,7 +487,7 @@ class AsyncDataSourcesClient:
487
487
  token="YOUR_TOKEN",
488
488
  )
489
489
  await client.data_sources.update_data_source(
490
- data_source_id="data_source_id",
490
+ data_source_id="string",
491
491
  source_type=ConfigurableDataSourceNames.S_3,
492
492
  )
493
493
  """
@@ -528,7 +528,7 @@ class AsyncDataSourcesClient:
528
528
  token="YOUR_TOKEN",
529
529
  )
530
530
  await client.data_sources.delete_data_source(
531
- data_source_id="data_source_id",
531
+ data_source_id="string",
532
532
  )
533
533
  """
534
534
  _response = await self._client_wrapper.httpx_client.request(
@@ -34,6 +34,15 @@ class EmbeddingModelConfigsClient:
34
34
  """
35
35
  Parameters:
36
36
  - project_id: str.
37
+ ---
38
+ from llama_cloud.client import LlamaCloud
39
+
40
+ client = LlamaCloud(
41
+ token="YOUR_TOKEN",
42
+ )
43
+ client.embedding_model_configs.list_embedding_model_configs(
44
+ project_id="string",
45
+ )
37
46
  """
38
47
  _response = self._client_wrapper.httpx_client.request(
39
48
  "GET",
@@ -64,22 +73,6 @@ class EmbeddingModelConfigsClient:
64
73
  - name: str. The name of the embedding model config.
65
74
 
66
75
  - embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
67
- ---
68
- from llama_cloud import (
69
- EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding,
70
- )
71
- from llama_cloud.client import LlamaCloud
72
-
73
- client = LlamaCloud(
74
- token="YOUR_TOKEN",
75
- )
76
- client.embedding_model_configs.create_embedding_model_config(
77
- project_id="project_id",
78
- name="name",
79
- embedding_config=EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding(
80
- type="VERTEXAI_EMBEDDING",
81
- ),
82
- )
83
76
  """
84
77
  _response = self._client_wrapper.httpx_client.request(
85
78
  "POST",
@@ -116,16 +109,6 @@ class EmbeddingModelConfigsClient:
116
109
  - organization_id: typing.Optional[str].
117
110
 
118
111
  - request: EmbeddingModelConfigUpdate.
119
- ---
120
- from llama_cloud import EmbeddingModelConfigUpdate
121
- from llama_cloud.client import LlamaCloud
122
-
123
- client = LlamaCloud(
124
- token="YOUR_TOKEN",
125
- )
126
- client.embedding_model_configs.upsert_embedding_model_config(
127
- request=EmbeddingModelConfigUpdate(),
128
- )
129
112
  """
130
113
  _response = self._client_wrapper.httpx_client.request(
131
114
  "PUT",
@@ -155,17 +138,6 @@ class EmbeddingModelConfigsClient:
155
138
  - embedding_model_config_id: str.
156
139
 
157
140
  - request: EmbeddingModelConfigUpdate.
158
- ---
159
- from llama_cloud import EmbeddingModelConfigUpdate
160
- from llama_cloud.client import LlamaCloud
161
-
162
- client = LlamaCloud(
163
- token="YOUR_TOKEN",
164
- )
165
- client.embedding_model_configs.update_embedding_model_config(
166
- embedding_model_config_id="embedding_model_config_id",
167
- request=EmbeddingModelConfigUpdate(),
168
- )
169
141
  """
170
142
  _response = self._client_wrapper.httpx_client.request(
171
143
  "PUT",
@@ -199,7 +171,7 @@ class EmbeddingModelConfigsClient:
199
171
  token="YOUR_TOKEN",
200
172
  )
201
173
  client.embedding_model_configs.delete_embedding_model_config(
202
- embedding_model_config_id="embedding_model_config_id",
174
+ embedding_model_config_id="string",
203
175
  )
204
176
  """
205
177
  _response = self._client_wrapper.httpx_client.request(
@@ -229,6 +201,15 @@ class AsyncEmbeddingModelConfigsClient:
229
201
  """
230
202
  Parameters:
231
203
  - project_id: str.
204
+ ---
205
+ from llama_cloud.client import AsyncLlamaCloud
206
+
207
+ client = AsyncLlamaCloud(
208
+ token="YOUR_TOKEN",
209
+ )
210
+ await client.embedding_model_configs.list_embedding_model_configs(
211
+ project_id="string",
212
+ )
232
213
  """
233
214
  _response = await self._client_wrapper.httpx_client.request(
234
215
  "GET",
@@ -259,22 +240,6 @@ class AsyncEmbeddingModelConfigsClient:
259
240
  - name: str. The name of the embedding model config.
260
241
 
261
242
  - embedding_config: EmbeddingModelConfigCreateEmbeddingConfig. The embedding configuration for the embedding model config.
262
- ---
263
- from llama_cloud import (
264
- EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding,
265
- )
266
- from llama_cloud.client import AsyncLlamaCloud
267
-
268
- client = AsyncLlamaCloud(
269
- token="YOUR_TOKEN",
270
- )
271
- await client.embedding_model_configs.create_embedding_model_config(
272
- project_id="project_id",
273
- name="name",
274
- embedding_config=EmbeddingModelConfigCreateEmbeddingConfig_VertexaiEmbedding(
275
- type="VERTEXAI_EMBEDDING",
276
- ),
277
- )
278
243
  """
279
244
  _response = await self._client_wrapper.httpx_client.request(
280
245
  "POST",
@@ -311,16 +276,6 @@ class AsyncEmbeddingModelConfigsClient:
311
276
  - organization_id: typing.Optional[str].
312
277
 
313
278
  - request: EmbeddingModelConfigUpdate.
314
- ---
315
- from llama_cloud import EmbeddingModelConfigUpdate
316
- from llama_cloud.client import AsyncLlamaCloud
317
-
318
- client = AsyncLlamaCloud(
319
- token="YOUR_TOKEN",
320
- )
321
- await client.embedding_model_configs.upsert_embedding_model_config(
322
- request=EmbeddingModelConfigUpdate(),
323
- )
324
279
  """
325
280
  _response = await self._client_wrapper.httpx_client.request(
326
281
  "PUT",
@@ -350,17 +305,6 @@ class AsyncEmbeddingModelConfigsClient:
350
305
  - embedding_model_config_id: str.
351
306
 
352
307
  - request: EmbeddingModelConfigUpdate.
353
- ---
354
- from llama_cloud import EmbeddingModelConfigUpdate
355
- from llama_cloud.client import AsyncLlamaCloud
356
-
357
- client = AsyncLlamaCloud(
358
- token="YOUR_TOKEN",
359
- )
360
- await client.embedding_model_configs.update_embedding_model_config(
361
- embedding_model_config_id="embedding_model_config_id",
362
- request=EmbeddingModelConfigUpdate(),
363
- )
364
308
  """
365
309
  _response = await self._client_wrapper.httpx_client.request(
366
310
  "PUT",
@@ -394,7 +338,7 @@ class AsyncEmbeddingModelConfigsClient:
394
338
  token="YOUR_TOKEN",
395
339
  )
396
340
  await client.embedding_model_configs.delete_embedding_model_config(
397
- embedding_model_config_id="embedding_model_config_id",
341
+ embedding_model_config_id="string",
398
342
  )
399
343
  """
400
344
  _response = await self._client_wrapper.httpx_client.request(
@@ -43,7 +43,7 @@ class EvalsClient:
43
43
  token="YOUR_TOKEN",
44
44
  )
45
45
  client.evals.get_dataset(
46
- dataset_id="dataset_id",
46
+ dataset_id="string",
47
47
  )
48
48
  """
49
49
  _response = self._client_wrapper.httpx_client.request(
@@ -77,8 +77,8 @@ class EvalsClient:
77
77
  token="YOUR_TOKEN",
78
78
  )
79
79
  client.evals.update_dataset(
80
- dataset_id="dataset_id",
81
- name="name",
80
+ dataset_id="string",
81
+ name="string",
82
82
  )
83
83
  """
84
84
  _response = self._client_wrapper.httpx_client.request(
@@ -111,7 +111,7 @@ class EvalsClient:
111
111
  token="YOUR_TOKEN",
112
112
  )
113
113
  client.evals.delete_dataset(
114
- dataset_id="dataset_id",
114
+ dataset_id="string",
115
115
  )
116
116
  """
117
117
  _response = self._client_wrapper.httpx_client.request(
@@ -143,7 +143,7 @@ class EvalsClient:
143
143
  token="YOUR_TOKEN",
144
144
  )
145
145
  client.evals.list_questions(
146
- dataset_id="dataset_id",
146
+ dataset_id="string",
147
147
  )
148
148
  """
149
149
  _response = self._client_wrapper.httpx_client.request(
@@ -180,9 +180,9 @@ class EvalsClient:
180
180
  token="YOUR_TOKEN",
181
181
  )
182
182
  client.evals.create_question(
183
- dataset_id="dataset_id",
183
+ dataset_id="string",
184
184
  request=EvalQuestionCreate(
185
- content="content",
185
+ content="string",
186
186
  ),
187
187
  )
188
188
  """
@@ -216,19 +216,14 @@ class EvalsClient:
216
216
 
217
217
  - request: typing.List[EvalQuestionCreate].
218
218
  ---
219
- from llama_cloud import EvalQuestionCreate
220
219
  from llama_cloud.client import LlamaCloud
221
220
 
222
221
  client = LlamaCloud(
223
222
  token="YOUR_TOKEN",
224
223
  )
225
224
  client.evals.create_questions(
226
- dataset_id="dataset_id",
227
- request=[
228
- EvalQuestionCreate(
229
- content="content",
230
- )
231
- ],
225
+ dataset_id="string",
226
+ request=[],
232
227
  )
233
228
  """
234
229
  _response = self._client_wrapper.httpx_client.request(
@@ -263,7 +258,7 @@ class EvalsClient:
263
258
  token="YOUR_TOKEN",
264
259
  )
265
260
  client.evals.get_question(
266
- question_id="question_id",
261
+ question_id="string",
267
262
  )
268
263
  """
269
264
  _response = self._client_wrapper.httpx_client.request(
@@ -298,9 +293,9 @@ class EvalsClient:
298
293
  token="YOUR_TOKEN",
299
294
  )
300
295
  client.evals.replace_question(
301
- question_id="question_id",
296
+ question_id="string",
302
297
  request=EvalQuestionCreate(
303
- content="content",
298
+ content="string",
304
299
  ),
305
300
  )
306
301
  """
@@ -334,7 +329,7 @@ class EvalsClient:
334
329
  token="YOUR_TOKEN",
335
330
  )
336
331
  client.evals.delete_question(
337
- question_id="question_id",
332
+ question_id="string",
338
333
  )
339
334
  """
340
335
  _response = self._client_wrapper.httpx_client.request(
@@ -399,7 +394,7 @@ class AsyncEvalsClient:
399
394
  token="YOUR_TOKEN",
400
395
  )
401
396
  await client.evals.get_dataset(
402
- dataset_id="dataset_id",
397
+ dataset_id="string",
403
398
  )
404
399
  """
405
400
  _response = await self._client_wrapper.httpx_client.request(
@@ -433,8 +428,8 @@ class AsyncEvalsClient:
433
428
  token="YOUR_TOKEN",
434
429
  )
435
430
  await client.evals.update_dataset(
436
- dataset_id="dataset_id",
437
- name="name",
431
+ dataset_id="string",
432
+ name="string",
438
433
  )
439
434
  """
440
435
  _response = await self._client_wrapper.httpx_client.request(
@@ -467,7 +462,7 @@ class AsyncEvalsClient:
467
462
  token="YOUR_TOKEN",
468
463
  )
469
464
  await client.evals.delete_dataset(
470
- dataset_id="dataset_id",
465
+ dataset_id="string",
471
466
  )
472
467
  """
473
468
  _response = await self._client_wrapper.httpx_client.request(
@@ -499,7 +494,7 @@ class AsyncEvalsClient:
499
494
  token="YOUR_TOKEN",
500
495
  )
501
496
  await client.evals.list_questions(
502
- dataset_id="dataset_id",
497
+ dataset_id="string",
503
498
  )
504
499
  """
505
500
  _response = await self._client_wrapper.httpx_client.request(
@@ -536,9 +531,9 @@ class AsyncEvalsClient:
536
531
  token="YOUR_TOKEN",
537
532
  )
538
533
  await client.evals.create_question(
539
- dataset_id="dataset_id",
534
+ dataset_id="string",
540
535
  request=EvalQuestionCreate(
541
- content="content",
536
+ content="string",
542
537
  ),
543
538
  )
544
539
  """
@@ -572,19 +567,14 @@ class AsyncEvalsClient:
572
567
 
573
568
  - request: typing.List[EvalQuestionCreate].
574
569
  ---
575
- from llama_cloud import EvalQuestionCreate
576
570
  from llama_cloud.client import AsyncLlamaCloud
577
571
 
578
572
  client = AsyncLlamaCloud(
579
573
  token="YOUR_TOKEN",
580
574
  )
581
575
  await client.evals.create_questions(
582
- dataset_id="dataset_id",
583
- request=[
584
- EvalQuestionCreate(
585
- content="content",
586
- )
587
- ],
576
+ dataset_id="string",
577
+ request=[],
588
578
  )
589
579
  """
590
580
  _response = await self._client_wrapper.httpx_client.request(
@@ -619,7 +609,7 @@ class AsyncEvalsClient:
619
609
  token="YOUR_TOKEN",
620
610
  )
621
611
  await client.evals.get_question(
622
- question_id="question_id",
612
+ question_id="string",
623
613
  )
624
614
  """
625
615
  _response = await self._client_wrapper.httpx_client.request(
@@ -654,9 +644,9 @@ class AsyncEvalsClient:
654
644
  token="YOUR_TOKEN",
655
645
  )
656
646
  await client.evals.replace_question(
657
- question_id="question_id",
647
+ question_id="string",
658
648
  request=EvalQuestionCreate(
659
- content="content",
649
+ content="string",
660
650
  ),
661
651
  )
662
652
  """
@@ -690,7 +680,7 @@ class AsyncEvalsClient:
690
680
  token="YOUR_TOKEN",
691
681
  )
692
682
  await client.evals.delete_question(
693
- question_id="question_id",
683
+ question_id="string",
694
684
  )
695
685
  """
696
686
  _response = await self._client_wrapper.httpx_client.request(