llama-cloud 0.1.5__py3-none-any.whl → 0.1.6__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +12 -10
- llama_cloud/environment.py +1 -1
- llama_cloud/resources/__init__.py +2 -1
- llama_cloud/resources/data_sinks/client.py +14 -14
- llama_cloud/resources/data_sources/client.py +16 -16
- llama_cloud/resources/embedding_model_configs/client.py +80 -24
- llama_cloud/resources/evals/client.py +36 -26
- llama_cloud/resources/extraction/client.py +32 -32
- llama_cloud/resources/files/__init__.py +2 -2
- llama_cloud/resources/files/client.py +53 -28
- llama_cloud/resources/files/types/__init__.py +2 -1
- llama_cloud/resources/files/types/file_create_permission_info_value.py +7 -0
- llama_cloud/resources/organizations/client.py +60 -56
- llama_cloud/resources/parsing/client.py +555 -324
- llama_cloud/resources/pipelines/client.py +446 -302
- llama_cloud/resources/projects/client.py +270 -136
- llama_cloud/types/__init__.py +10 -10
- llama_cloud/types/azure_open_ai_embedding.py +12 -6
- llama_cloud/types/base_prompt_template.py +6 -2
- llama_cloud/types/bedrock_embedding.py +12 -6
- llama_cloud/types/character_splitter.py +4 -2
- llama_cloud/types/chat_message.py +1 -1
- llama_cloud/types/cloud_az_storage_blob_data_source.py +16 -7
- llama_cloud/types/cloud_box_data_source.py +13 -6
- llama_cloud/types/cloud_confluence_data_source.py +7 -6
- llama_cloud/types/cloud_document.py +3 -1
- llama_cloud/types/cloud_document_create.py +3 -1
- llama_cloud/types/cloud_google_drive_data_source.py +1 -0
- llama_cloud/types/cloud_jira_data_source.py +7 -4
- llama_cloud/types/cloud_notion_page_data_source.py +3 -2
- llama_cloud/types/cloud_one_drive_data_source.py +6 -3
- llama_cloud/types/cloud_s_3_data_source.py +9 -4
- llama_cloud/types/cloud_sharepoint_data_source.py +9 -6
- llama_cloud/types/cloud_slack_data_source.py +7 -6
- llama_cloud/types/code_splitter.py +1 -1
- llama_cloud/types/cohere_embedding.py +7 -3
- llama_cloud/types/data_sink.py +4 -4
- llama_cloud/types/data_sink_create.py +1 -1
- llama_cloud/types/data_source.py +7 -5
- llama_cloud/types/data_source_create.py +4 -2
- llama_cloud/types/embedding_model_config.py +2 -2
- llama_cloud/types/embedding_model_config_update.py +4 -2
- llama_cloud/types/eval_dataset.py +2 -2
- llama_cloud/types/eval_dataset_job_record.py +13 -7
- llama_cloud/types/eval_execution_params_override.py +6 -2
- llama_cloud/types/eval_question.py +2 -2
- llama_cloud/types/extraction_result.py +2 -2
- llama_cloud/types/extraction_schema.py +5 -3
- llama_cloud/types/file.py +15 -7
- llama_cloud/types/file_permission_info_value.py +5 -0
- llama_cloud/types/filter_operator.py +2 -2
- llama_cloud/types/gemini_embedding.py +10 -6
- llama_cloud/types/hugging_face_inference_api_embedding.py +27 -11
- llama_cloud/types/input_message.py +3 -1
- llama_cloud/types/job_name_mapping.py +4 -0
- llama_cloud/types/llama_parse_parameters.py +11 -0
- llama_cloud/types/llm.py +4 -2
- llama_cloud/types/llm_parameters.py +5 -2
- llama_cloud/types/local_eval.py +10 -8
- llama_cloud/types/local_eval_results.py +1 -1
- llama_cloud/types/managed_ingestion_status_response.py +5 -3
- llama_cloud/types/markdown_element_node_parser.py +5 -3
- llama_cloud/types/markdown_node_parser.py +1 -1
- llama_cloud/types/metadata_filter.py +2 -2
- llama_cloud/types/metric_result.py +3 -3
- llama_cloud/types/node_parser.py +1 -1
- llama_cloud/types/open_ai_embedding.py +12 -6
- llama_cloud/types/organization.py +2 -2
- llama_cloud/types/page_splitter_node_parser.py +2 -2
- llama_cloud/types/parsing_job_structured_result.py +32 -0
- llama_cloud/types/permission.py +3 -3
- llama_cloud/types/pipeline.py +17 -7
- llama_cloud/types/pipeline_configuration_hashes.py +3 -3
- llama_cloud/types/pipeline_create.py +15 -5
- llama_cloud/types/pipeline_data_source.py +13 -7
- llama_cloud/types/pipeline_data_source_create.py +3 -1
- llama_cloud/types/pipeline_deployment.py +4 -4
- llama_cloud/types/pipeline_file.py +25 -11
- llama_cloud/types/pipeline_file_create.py +3 -1
- llama_cloud/types/pipeline_file_permission_info_value.py +7 -0
- llama_cloud/types/playground_session.py +2 -2
- llama_cloud/types/preset_retrieval_params.py +14 -7
- llama_cloud/types/presigned_url.py +3 -1
- llama_cloud/types/project.py +2 -2
- llama_cloud/types/prompt_mixin_prompts.py +1 -1
- llama_cloud/types/prompt_spec.py +4 -2
- llama_cloud/types/role.py +3 -3
- llama_cloud/types/sentence_splitter.py +4 -2
- llama_cloud/types/text_node.py +3 -3
- llama_cloud/types/{hugging_face_inference_api_embedding_token.py → token.py} +1 -1
- llama_cloud/types/token_text_splitter.py +1 -1
- llama_cloud/types/user_organization.py +9 -5
- llama_cloud/types/user_organization_create.py +4 -4
- llama_cloud/types/user_organization_delete.py +2 -2
- llama_cloud/types/user_organization_role.py +2 -2
- llama_cloud/types/value.py +5 -0
- llama_cloud/types/vertex_text_embedding.py +9 -5
- {llama_cloud-0.1.5.dist-info → llama_cloud-0.1.6.dist-info}/METADATA +2 -1
- {llama_cloud-0.1.5.dist-info → llama_cloud-0.1.6.dist-info}/RECORD +101 -100
- {llama_cloud-0.1.5.dist-info → llama_cloud-0.1.6.dist-info}/WHEEL +1 -1
- llama_cloud/types/data_sink_component.py +0 -20
- llama_cloud/types/data_source_component.py +0 -28
- llama_cloud/types/metadata_filter_value.py +0 -5
- llama_cloud/types/pipeline_data_source_component.py +0 -28
- {llama_cloud-0.1.5.dist-info → llama_cloud-0.1.6.dist-info}/LICENSE +0 -0
|
@@ -15,6 +15,7 @@ from ...types.http_validation_error import HttpValidationError
|
|
|
15
15
|
from ...types.page_screenshot_metadata import PageScreenshotMetadata
|
|
16
16
|
from ...types.presigned_url import PresignedUrl
|
|
17
17
|
from .types.file_create_from_url_resource_info_value import FileCreateFromUrlResourceInfoValue
|
|
18
|
+
from .types.file_create_permission_info_value import FileCreatePermissionInfoValue
|
|
18
19
|
from .types.file_create_resource_info_value import FileCreateResourceInfoValue
|
|
19
20
|
|
|
20
21
|
try:
|
|
@@ -52,7 +53,7 @@ class FilesClient:
|
|
|
52
53
|
token="YOUR_TOKEN",
|
|
53
54
|
)
|
|
54
55
|
client.files.get_file(
|
|
55
|
-
id="
|
|
56
|
+
id="id",
|
|
56
57
|
)
|
|
57
58
|
"""
|
|
58
59
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -91,7 +92,7 @@ class FilesClient:
|
|
|
91
92
|
token="YOUR_TOKEN",
|
|
92
93
|
)
|
|
93
94
|
client.files.delete_file(
|
|
94
|
-
id="
|
|
95
|
+
id="id",
|
|
95
96
|
)
|
|
96
97
|
"""
|
|
97
98
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -162,6 +163,13 @@ class FilesClient:
|
|
|
162
163
|
- organization_id: typing.Optional[str].
|
|
163
164
|
|
|
164
165
|
- upload_file: typing.IO.
|
|
166
|
+
---
|
|
167
|
+
from llama_cloud.client import LlamaCloud
|
|
168
|
+
|
|
169
|
+
client = LlamaCloud(
|
|
170
|
+
token="YOUR_TOKEN",
|
|
171
|
+
)
|
|
172
|
+
client.files.upload_file()
|
|
165
173
|
"""
|
|
166
174
|
_response = self._client_wrapper.httpx_client.request(
|
|
167
175
|
"POST",
|
|
@@ -191,6 +199,7 @@ class FilesClient:
|
|
|
191
199
|
file_size: typing.Optional[int] = OMIT,
|
|
192
200
|
last_modified_at: typing.Optional[dt.datetime] = OMIT,
|
|
193
201
|
resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateResourceInfoValue]]] = OMIT,
|
|
202
|
+
permission_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreatePermissionInfoValue]]] = OMIT,
|
|
194
203
|
data_source_id: typing.Optional[str] = OMIT,
|
|
195
204
|
) -> PresignedUrl:
|
|
196
205
|
"""
|
|
@@ -203,13 +212,15 @@ class FilesClient:
|
|
|
203
212
|
|
|
204
213
|
- name: str.
|
|
205
214
|
|
|
206
|
-
- file_size: typing.Optional[int].
|
|
215
|
+
- file_size: typing.Optional[int]. Size of the file in bytes
|
|
216
|
+
|
|
217
|
+
- last_modified_at: typing.Optional[dt.datetime]. The last modified time of the file
|
|
207
218
|
|
|
208
|
-
-
|
|
219
|
+
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateResourceInfoValue]]]. Resource information for the file
|
|
209
220
|
|
|
210
|
-
-
|
|
221
|
+
- permission_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreatePermissionInfoValue]]]. Permission information for the file
|
|
211
222
|
|
|
212
|
-
- data_source_id: typing.Optional[str].
|
|
223
|
+
- data_source_id: typing.Optional[str]. The ID of the data source that the file belongs to
|
|
213
224
|
---
|
|
214
225
|
from llama_cloud.client import LlamaCloud
|
|
215
226
|
|
|
@@ -217,7 +228,7 @@ class FilesClient:
|
|
|
217
228
|
token="YOUR_TOKEN",
|
|
218
229
|
)
|
|
219
230
|
client.files.generate_presigned_url(
|
|
220
|
-
name="
|
|
231
|
+
name="name",
|
|
221
232
|
)
|
|
222
233
|
"""
|
|
223
234
|
_request: typing.Dict[str, typing.Any] = {"name": name}
|
|
@@ -227,6 +238,8 @@ class FilesClient:
|
|
|
227
238
|
_request["last_modified_at"] = last_modified_at
|
|
228
239
|
if resource_info is not OMIT:
|
|
229
240
|
_request["resource_info"] = resource_info
|
|
241
|
+
if permission_info is not OMIT:
|
|
242
|
+
_request["permission_info"] = permission_info
|
|
230
243
|
if data_source_id is not OMIT:
|
|
231
244
|
_request["data_source_id"] = data_source_id
|
|
232
245
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -307,15 +320,15 @@ class FilesClient:
|
|
|
307
320
|
|
|
308
321
|
- url: str. URL of the file to download
|
|
309
322
|
|
|
310
|
-
- proxy_url: typing.Optional[str].
|
|
323
|
+
- proxy_url: typing.Optional[str]. URL of the proxy server to use for downloading the file
|
|
311
324
|
|
|
312
|
-
- request_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]].
|
|
325
|
+
- request_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]]. Headers to include in the request when downloading the file
|
|
313
326
|
|
|
314
327
|
- verify_ssl: typing.Optional[bool]. Whether to verify the SSL certificate when downloading the file
|
|
315
328
|
|
|
316
329
|
- follow_redirects: typing.Optional[bool]. Whether to follow redirects when downloading the file
|
|
317
330
|
|
|
318
|
-
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateFromUrlResourceInfoValue]]].
|
|
331
|
+
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateFromUrlResourceInfoValue]]]. Resource information for the file
|
|
319
332
|
---
|
|
320
333
|
from llama_cloud.client import LlamaCloud
|
|
321
334
|
|
|
@@ -323,7 +336,7 @@ class FilesClient:
|
|
|
323
336
|
token="YOUR_TOKEN",
|
|
324
337
|
)
|
|
325
338
|
client.files.upload_file_from_url(
|
|
326
|
-
url="
|
|
339
|
+
url="url",
|
|
327
340
|
)
|
|
328
341
|
"""
|
|
329
342
|
_request: typing.Dict[str, typing.Any] = {"url": url}
|
|
@@ -376,7 +389,7 @@ class FilesClient:
|
|
|
376
389
|
token="YOUR_TOKEN",
|
|
377
390
|
)
|
|
378
391
|
client.files.read_file_content(
|
|
379
|
-
id="
|
|
392
|
+
id="id",
|
|
380
393
|
)
|
|
381
394
|
"""
|
|
382
395
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -415,7 +428,7 @@ class FilesClient:
|
|
|
415
428
|
token="YOUR_TOKEN",
|
|
416
429
|
)
|
|
417
430
|
client.files.list_file_page_screenshots(
|
|
418
|
-
id="
|
|
431
|
+
id="id",
|
|
419
432
|
)
|
|
420
433
|
"""
|
|
421
434
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -461,7 +474,7 @@ class FilesClient:
|
|
|
461
474
|
token="YOUR_TOKEN",
|
|
462
475
|
)
|
|
463
476
|
client.files.get_file_page_screenshot(
|
|
464
|
-
id="
|
|
477
|
+
id="id",
|
|
465
478
|
page_index=1,
|
|
466
479
|
)
|
|
467
480
|
"""
|
|
@@ -508,7 +521,7 @@ class AsyncFilesClient:
|
|
|
508
521
|
token="YOUR_TOKEN",
|
|
509
522
|
)
|
|
510
523
|
await client.files.get_file(
|
|
511
|
-
id="
|
|
524
|
+
id="id",
|
|
512
525
|
)
|
|
513
526
|
"""
|
|
514
527
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -547,7 +560,7 @@ class AsyncFilesClient:
|
|
|
547
560
|
token="YOUR_TOKEN",
|
|
548
561
|
)
|
|
549
562
|
await client.files.delete_file(
|
|
550
|
-
id="
|
|
563
|
+
id="id",
|
|
551
564
|
)
|
|
552
565
|
"""
|
|
553
566
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -618,6 +631,13 @@ class AsyncFilesClient:
|
|
|
618
631
|
- organization_id: typing.Optional[str].
|
|
619
632
|
|
|
620
633
|
- upload_file: typing.IO.
|
|
634
|
+
---
|
|
635
|
+
from llama_cloud.client import AsyncLlamaCloud
|
|
636
|
+
|
|
637
|
+
client = AsyncLlamaCloud(
|
|
638
|
+
token="YOUR_TOKEN",
|
|
639
|
+
)
|
|
640
|
+
await client.files.upload_file()
|
|
621
641
|
"""
|
|
622
642
|
_response = await self._client_wrapper.httpx_client.request(
|
|
623
643
|
"POST",
|
|
@@ -647,6 +667,7 @@ class AsyncFilesClient:
|
|
|
647
667
|
file_size: typing.Optional[int] = OMIT,
|
|
648
668
|
last_modified_at: typing.Optional[dt.datetime] = OMIT,
|
|
649
669
|
resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateResourceInfoValue]]] = OMIT,
|
|
670
|
+
permission_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreatePermissionInfoValue]]] = OMIT,
|
|
650
671
|
data_source_id: typing.Optional[str] = OMIT,
|
|
651
672
|
) -> PresignedUrl:
|
|
652
673
|
"""
|
|
@@ -659,13 +680,15 @@ class AsyncFilesClient:
|
|
|
659
680
|
|
|
660
681
|
- name: str.
|
|
661
682
|
|
|
662
|
-
- file_size: typing.Optional[int].
|
|
683
|
+
- file_size: typing.Optional[int]. Size of the file in bytes
|
|
684
|
+
|
|
685
|
+
- last_modified_at: typing.Optional[dt.datetime]. The last modified time of the file
|
|
663
686
|
|
|
664
|
-
-
|
|
687
|
+
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateResourceInfoValue]]]. Resource information for the file
|
|
665
688
|
|
|
666
|
-
-
|
|
689
|
+
- permission_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreatePermissionInfoValue]]]. Permission information for the file
|
|
667
690
|
|
|
668
|
-
- data_source_id: typing.Optional[str].
|
|
691
|
+
- data_source_id: typing.Optional[str]. The ID of the data source that the file belongs to
|
|
669
692
|
---
|
|
670
693
|
from llama_cloud.client import AsyncLlamaCloud
|
|
671
694
|
|
|
@@ -673,7 +696,7 @@ class AsyncFilesClient:
|
|
|
673
696
|
token="YOUR_TOKEN",
|
|
674
697
|
)
|
|
675
698
|
await client.files.generate_presigned_url(
|
|
676
|
-
name="
|
|
699
|
+
name="name",
|
|
677
700
|
)
|
|
678
701
|
"""
|
|
679
702
|
_request: typing.Dict[str, typing.Any] = {"name": name}
|
|
@@ -683,6 +706,8 @@ class AsyncFilesClient:
|
|
|
683
706
|
_request["last_modified_at"] = last_modified_at
|
|
684
707
|
if resource_info is not OMIT:
|
|
685
708
|
_request["resource_info"] = resource_info
|
|
709
|
+
if permission_info is not OMIT:
|
|
710
|
+
_request["permission_info"] = permission_info
|
|
686
711
|
if data_source_id is not OMIT:
|
|
687
712
|
_request["data_source_id"] = data_source_id
|
|
688
713
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -763,15 +788,15 @@ class AsyncFilesClient:
|
|
|
763
788
|
|
|
764
789
|
- url: str. URL of the file to download
|
|
765
790
|
|
|
766
|
-
- proxy_url: typing.Optional[str].
|
|
791
|
+
- proxy_url: typing.Optional[str]. URL of the proxy server to use for downloading the file
|
|
767
792
|
|
|
768
|
-
- request_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]].
|
|
793
|
+
- request_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]]. Headers to include in the request when downloading the file
|
|
769
794
|
|
|
770
795
|
- verify_ssl: typing.Optional[bool]. Whether to verify the SSL certificate when downloading the file
|
|
771
796
|
|
|
772
797
|
- follow_redirects: typing.Optional[bool]. Whether to follow redirects when downloading the file
|
|
773
798
|
|
|
774
|
-
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateFromUrlResourceInfoValue]]].
|
|
799
|
+
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateFromUrlResourceInfoValue]]]. Resource information for the file
|
|
775
800
|
---
|
|
776
801
|
from llama_cloud.client import AsyncLlamaCloud
|
|
777
802
|
|
|
@@ -779,7 +804,7 @@ class AsyncFilesClient:
|
|
|
779
804
|
token="YOUR_TOKEN",
|
|
780
805
|
)
|
|
781
806
|
await client.files.upload_file_from_url(
|
|
782
|
-
url="
|
|
807
|
+
url="url",
|
|
783
808
|
)
|
|
784
809
|
"""
|
|
785
810
|
_request: typing.Dict[str, typing.Any] = {"url": url}
|
|
@@ -832,7 +857,7 @@ class AsyncFilesClient:
|
|
|
832
857
|
token="YOUR_TOKEN",
|
|
833
858
|
)
|
|
834
859
|
await client.files.read_file_content(
|
|
835
|
-
id="
|
|
860
|
+
id="id",
|
|
836
861
|
)
|
|
837
862
|
"""
|
|
838
863
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -871,7 +896,7 @@ class AsyncFilesClient:
|
|
|
871
896
|
token="YOUR_TOKEN",
|
|
872
897
|
)
|
|
873
898
|
await client.files.list_file_page_screenshots(
|
|
874
|
-
id="
|
|
899
|
+
id="id",
|
|
875
900
|
)
|
|
876
901
|
"""
|
|
877
902
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -917,7 +942,7 @@ class AsyncFilesClient:
|
|
|
917
942
|
token="YOUR_TOKEN",
|
|
918
943
|
)
|
|
919
944
|
await client.files.get_file_page_screenshot(
|
|
920
|
-
id="
|
|
945
|
+
id="id",
|
|
921
946
|
page_index=1,
|
|
922
947
|
)
|
|
923
948
|
"""
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
|
2
2
|
|
|
3
3
|
from .file_create_from_url_resource_info_value import FileCreateFromUrlResourceInfoValue
|
|
4
|
+
from .file_create_permission_info_value import FileCreatePermissionInfoValue
|
|
4
5
|
from .file_create_resource_info_value import FileCreateResourceInfoValue
|
|
5
6
|
|
|
6
|
-
__all__ = ["FileCreateFromUrlResourceInfoValue", "FileCreateResourceInfoValue"]
|
|
7
|
+
__all__ = ["FileCreateFromUrlResourceInfoValue", "FileCreatePermissionInfoValue", "FileCreateResourceInfoValue"]
|
|
@@ -79,7 +79,7 @@ class OrganizationsClient:
|
|
|
79
79
|
)
|
|
80
80
|
client.organizations.create_organization(
|
|
81
81
|
request=OrganizationCreate(
|
|
82
|
-
name="
|
|
82
|
+
name="name",
|
|
83
83
|
),
|
|
84
84
|
)
|
|
85
85
|
"""
|
|
@@ -115,7 +115,7 @@ class OrganizationsClient:
|
|
|
115
115
|
)
|
|
116
116
|
client.organizations.upsert_organization(
|
|
117
117
|
request=OrganizationCreate(
|
|
118
|
-
name="
|
|
118
|
+
name="name",
|
|
119
119
|
),
|
|
120
120
|
)
|
|
121
121
|
"""
|
|
@@ -177,7 +177,7 @@ class OrganizationsClient:
|
|
|
177
177
|
token="YOUR_TOKEN",
|
|
178
178
|
)
|
|
179
179
|
client.organizations.set_default_organization(
|
|
180
|
-
organization_id="
|
|
180
|
+
organization_id="organization_id",
|
|
181
181
|
)
|
|
182
182
|
"""
|
|
183
183
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -210,7 +210,7 @@ class OrganizationsClient:
|
|
|
210
210
|
token="YOUR_TOKEN",
|
|
211
211
|
)
|
|
212
212
|
client.organizations.get_organization(
|
|
213
|
-
organization_id="
|
|
213
|
+
organization_id="organization_id",
|
|
214
214
|
)
|
|
215
215
|
"""
|
|
216
216
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -236,7 +236,7 @@ class OrganizationsClient:
|
|
|
236
236
|
Parameters:
|
|
237
237
|
- organization_id: str.
|
|
238
238
|
|
|
239
|
-
- name: typing.Optional[str].
|
|
239
|
+
- name: typing.Optional[str]. A name for the organization.
|
|
240
240
|
---
|
|
241
241
|
from llama_cloud.client import LlamaCloud
|
|
242
242
|
|
|
@@ -244,7 +244,7 @@ class OrganizationsClient:
|
|
|
244
244
|
token="YOUR_TOKEN",
|
|
245
245
|
)
|
|
246
246
|
client.organizations.update_organization(
|
|
247
|
-
organization_id="
|
|
247
|
+
organization_id="organization_id",
|
|
248
248
|
)
|
|
249
249
|
"""
|
|
250
250
|
_request: typing.Dict[str, typing.Any] = {}
|
|
@@ -280,7 +280,7 @@ class OrganizationsClient:
|
|
|
280
280
|
token="YOUR_TOKEN",
|
|
281
281
|
)
|
|
282
282
|
client.organizations.delete_organization(
|
|
283
|
-
organization_id="
|
|
283
|
+
organization_id="organization_id",
|
|
284
284
|
)
|
|
285
285
|
"""
|
|
286
286
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -344,7 +344,7 @@ class OrganizationsClient:
|
|
|
344
344
|
token="YOUR_TOKEN",
|
|
345
345
|
)
|
|
346
346
|
client.organizations.list_organization_users(
|
|
347
|
-
organization_id="
|
|
347
|
+
organization_id="organization_id",
|
|
348
348
|
)
|
|
349
349
|
"""
|
|
350
350
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -376,14 +376,15 @@ class OrganizationsClient:
|
|
|
376
376
|
|
|
377
377
|
- request: typing.List[UserOrganizationCreate].
|
|
378
378
|
---
|
|
379
|
+
from llama_cloud import UserOrganizationCreate
|
|
379
380
|
from llama_cloud.client import LlamaCloud
|
|
380
381
|
|
|
381
382
|
client = LlamaCloud(
|
|
382
383
|
token="YOUR_TOKEN",
|
|
383
384
|
)
|
|
384
385
|
client.organizations.add_users_to_organization(
|
|
385
|
-
organization_id="
|
|
386
|
-
request=[],
|
|
386
|
+
organization_id="organization_id",
|
|
387
|
+
request=[UserOrganizationCreate()],
|
|
387
388
|
)
|
|
388
389
|
"""
|
|
389
390
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -420,8 +421,8 @@ class OrganizationsClient:
|
|
|
420
421
|
token="YOUR_TOKEN",
|
|
421
422
|
)
|
|
422
423
|
client.organizations.remove_users_from_organization(
|
|
423
|
-
organization_id="
|
|
424
|
-
member_user_id="
|
|
424
|
+
organization_id="organization_id",
|
|
425
|
+
member_user_id="member_user_id",
|
|
425
426
|
)
|
|
426
427
|
"""
|
|
427
428
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -454,14 +455,15 @@ class OrganizationsClient:
|
|
|
454
455
|
|
|
455
456
|
- request: typing.List[UserOrganizationDelete].
|
|
456
457
|
---
|
|
458
|
+
from llama_cloud import UserOrganizationDelete
|
|
457
459
|
from llama_cloud.client import LlamaCloud
|
|
458
460
|
|
|
459
461
|
client = LlamaCloud(
|
|
460
462
|
token="YOUR_TOKEN",
|
|
461
463
|
)
|
|
462
464
|
client.organizations.batch_remove_users_from_organization(
|
|
463
|
-
organization_id="
|
|
464
|
-
request=[],
|
|
465
|
+
organization_id="organization_id",
|
|
466
|
+
request=[UserOrganizationDelete()],
|
|
465
467
|
)
|
|
466
468
|
"""
|
|
467
469
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -496,7 +498,7 @@ class OrganizationsClient:
|
|
|
496
498
|
token="YOUR_TOKEN",
|
|
497
499
|
)
|
|
498
500
|
client.organizations.list_roles(
|
|
499
|
-
organization_id="
|
|
501
|
+
organization_id="organization_id",
|
|
500
502
|
)
|
|
501
503
|
"""
|
|
502
504
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -530,7 +532,7 @@ class OrganizationsClient:
|
|
|
530
532
|
token="YOUR_TOKEN",
|
|
531
533
|
)
|
|
532
534
|
client.organizations.get_user_role(
|
|
533
|
-
organization_id="
|
|
535
|
+
organization_id="organization_id",
|
|
534
536
|
)
|
|
535
537
|
"""
|
|
536
538
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -572,10 +574,10 @@ class OrganizationsClient:
|
|
|
572
574
|
token="YOUR_TOKEN",
|
|
573
575
|
)
|
|
574
576
|
client.organizations.assign_role_to_user_in_organization(
|
|
575
|
-
organization_id="
|
|
576
|
-
user_id="
|
|
577
|
-
user_organization_role_create_organization_id="
|
|
578
|
-
role_id="
|
|
577
|
+
organization_id="organization_id",
|
|
578
|
+
user_id="user_id",
|
|
579
|
+
user_organization_role_create_organization_id="organization_id",
|
|
580
|
+
role_id="role_id",
|
|
579
581
|
)
|
|
580
582
|
"""
|
|
581
583
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -618,8 +620,8 @@ class OrganizationsClient:
|
|
|
618
620
|
token="YOUR_TOKEN",
|
|
619
621
|
)
|
|
620
622
|
client.organizations.list_projects_by_user(
|
|
621
|
-
organization_id="
|
|
622
|
-
user_id="
|
|
623
|
+
organization_id="organization_id",
|
|
624
|
+
user_id="user_id",
|
|
623
625
|
)
|
|
624
626
|
"""
|
|
625
627
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -658,9 +660,9 @@ class OrganizationsClient:
|
|
|
658
660
|
token="YOUR_TOKEN",
|
|
659
661
|
)
|
|
660
662
|
client.organizations.add_user_to_project(
|
|
661
|
-
organization_id="
|
|
662
|
-
user_id="
|
|
663
|
-
project_id="
|
|
663
|
+
organization_id="organization_id",
|
|
664
|
+
user_id="user_id",
|
|
665
|
+
project_id="project_id",
|
|
664
666
|
)
|
|
665
667
|
"""
|
|
666
668
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -700,9 +702,9 @@ class OrganizationsClient:
|
|
|
700
702
|
token="YOUR_TOKEN",
|
|
701
703
|
)
|
|
702
704
|
client.organizations.remove_user_from_project(
|
|
703
|
-
organization_id="
|
|
704
|
-
user_id="
|
|
705
|
-
project_id="
|
|
705
|
+
organization_id="organization_id",
|
|
706
|
+
user_id="user_id",
|
|
707
|
+
project_id="project_id",
|
|
706
708
|
)
|
|
707
709
|
"""
|
|
708
710
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -772,7 +774,7 @@ class AsyncOrganizationsClient:
|
|
|
772
774
|
)
|
|
773
775
|
await client.organizations.create_organization(
|
|
774
776
|
request=OrganizationCreate(
|
|
775
|
-
name="
|
|
777
|
+
name="name",
|
|
776
778
|
),
|
|
777
779
|
)
|
|
778
780
|
"""
|
|
@@ -808,7 +810,7 @@ class AsyncOrganizationsClient:
|
|
|
808
810
|
)
|
|
809
811
|
await client.organizations.upsert_organization(
|
|
810
812
|
request=OrganizationCreate(
|
|
811
|
-
name="
|
|
813
|
+
name="name",
|
|
812
814
|
),
|
|
813
815
|
)
|
|
814
816
|
"""
|
|
@@ -870,7 +872,7 @@ class AsyncOrganizationsClient:
|
|
|
870
872
|
token="YOUR_TOKEN",
|
|
871
873
|
)
|
|
872
874
|
await client.organizations.set_default_organization(
|
|
873
|
-
organization_id="
|
|
875
|
+
organization_id="organization_id",
|
|
874
876
|
)
|
|
875
877
|
"""
|
|
876
878
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -903,7 +905,7 @@ class AsyncOrganizationsClient:
|
|
|
903
905
|
token="YOUR_TOKEN",
|
|
904
906
|
)
|
|
905
907
|
await client.organizations.get_organization(
|
|
906
|
-
organization_id="
|
|
908
|
+
organization_id="organization_id",
|
|
907
909
|
)
|
|
908
910
|
"""
|
|
909
911
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -929,7 +931,7 @@ class AsyncOrganizationsClient:
|
|
|
929
931
|
Parameters:
|
|
930
932
|
- organization_id: str.
|
|
931
933
|
|
|
932
|
-
- name: typing.Optional[str].
|
|
934
|
+
- name: typing.Optional[str]. A name for the organization.
|
|
933
935
|
---
|
|
934
936
|
from llama_cloud.client import AsyncLlamaCloud
|
|
935
937
|
|
|
@@ -937,7 +939,7 @@ class AsyncOrganizationsClient:
|
|
|
937
939
|
token="YOUR_TOKEN",
|
|
938
940
|
)
|
|
939
941
|
await client.organizations.update_organization(
|
|
940
|
-
organization_id="
|
|
942
|
+
organization_id="organization_id",
|
|
941
943
|
)
|
|
942
944
|
"""
|
|
943
945
|
_request: typing.Dict[str, typing.Any] = {}
|
|
@@ -973,7 +975,7 @@ class AsyncOrganizationsClient:
|
|
|
973
975
|
token="YOUR_TOKEN",
|
|
974
976
|
)
|
|
975
977
|
await client.organizations.delete_organization(
|
|
976
|
-
organization_id="
|
|
978
|
+
organization_id="organization_id",
|
|
977
979
|
)
|
|
978
980
|
"""
|
|
979
981
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1037,7 +1039,7 @@ class AsyncOrganizationsClient:
|
|
|
1037
1039
|
token="YOUR_TOKEN",
|
|
1038
1040
|
)
|
|
1039
1041
|
await client.organizations.list_organization_users(
|
|
1040
|
-
organization_id="
|
|
1042
|
+
organization_id="organization_id",
|
|
1041
1043
|
)
|
|
1042
1044
|
"""
|
|
1043
1045
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1069,14 +1071,15 @@ class AsyncOrganizationsClient:
|
|
|
1069
1071
|
|
|
1070
1072
|
- request: typing.List[UserOrganizationCreate].
|
|
1071
1073
|
---
|
|
1074
|
+
from llama_cloud import UserOrganizationCreate
|
|
1072
1075
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1073
1076
|
|
|
1074
1077
|
client = AsyncLlamaCloud(
|
|
1075
1078
|
token="YOUR_TOKEN",
|
|
1076
1079
|
)
|
|
1077
1080
|
await client.organizations.add_users_to_organization(
|
|
1078
|
-
organization_id="
|
|
1079
|
-
request=[],
|
|
1081
|
+
organization_id="organization_id",
|
|
1082
|
+
request=[UserOrganizationCreate()],
|
|
1080
1083
|
)
|
|
1081
1084
|
"""
|
|
1082
1085
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1113,8 +1116,8 @@ class AsyncOrganizationsClient:
|
|
|
1113
1116
|
token="YOUR_TOKEN",
|
|
1114
1117
|
)
|
|
1115
1118
|
await client.organizations.remove_users_from_organization(
|
|
1116
|
-
organization_id="
|
|
1117
|
-
member_user_id="
|
|
1119
|
+
organization_id="organization_id",
|
|
1120
|
+
member_user_id="member_user_id",
|
|
1118
1121
|
)
|
|
1119
1122
|
"""
|
|
1120
1123
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1147,14 +1150,15 @@ class AsyncOrganizationsClient:
|
|
|
1147
1150
|
|
|
1148
1151
|
- request: typing.List[UserOrganizationDelete].
|
|
1149
1152
|
---
|
|
1153
|
+
from llama_cloud import UserOrganizationDelete
|
|
1150
1154
|
from llama_cloud.client import AsyncLlamaCloud
|
|
1151
1155
|
|
|
1152
1156
|
client = AsyncLlamaCloud(
|
|
1153
1157
|
token="YOUR_TOKEN",
|
|
1154
1158
|
)
|
|
1155
1159
|
await client.organizations.batch_remove_users_from_organization(
|
|
1156
|
-
organization_id="
|
|
1157
|
-
request=[],
|
|
1160
|
+
organization_id="organization_id",
|
|
1161
|
+
request=[UserOrganizationDelete()],
|
|
1158
1162
|
)
|
|
1159
1163
|
"""
|
|
1160
1164
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1189,7 +1193,7 @@ class AsyncOrganizationsClient:
|
|
|
1189
1193
|
token="YOUR_TOKEN",
|
|
1190
1194
|
)
|
|
1191
1195
|
await client.organizations.list_roles(
|
|
1192
|
-
organization_id="
|
|
1196
|
+
organization_id="organization_id",
|
|
1193
1197
|
)
|
|
1194
1198
|
"""
|
|
1195
1199
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1223,7 +1227,7 @@ class AsyncOrganizationsClient:
|
|
|
1223
1227
|
token="YOUR_TOKEN",
|
|
1224
1228
|
)
|
|
1225
1229
|
await client.organizations.get_user_role(
|
|
1226
|
-
organization_id="
|
|
1230
|
+
organization_id="organization_id",
|
|
1227
1231
|
)
|
|
1228
1232
|
"""
|
|
1229
1233
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1265,10 +1269,10 @@ class AsyncOrganizationsClient:
|
|
|
1265
1269
|
token="YOUR_TOKEN",
|
|
1266
1270
|
)
|
|
1267
1271
|
await client.organizations.assign_role_to_user_in_organization(
|
|
1268
|
-
organization_id="
|
|
1269
|
-
user_id="
|
|
1270
|
-
user_organization_role_create_organization_id="
|
|
1271
|
-
role_id="
|
|
1272
|
+
organization_id="organization_id",
|
|
1273
|
+
user_id="user_id",
|
|
1274
|
+
user_organization_role_create_organization_id="organization_id",
|
|
1275
|
+
role_id="role_id",
|
|
1272
1276
|
)
|
|
1273
1277
|
"""
|
|
1274
1278
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1311,8 +1315,8 @@ class AsyncOrganizationsClient:
|
|
|
1311
1315
|
token="YOUR_TOKEN",
|
|
1312
1316
|
)
|
|
1313
1317
|
await client.organizations.list_projects_by_user(
|
|
1314
|
-
organization_id="
|
|
1315
|
-
user_id="
|
|
1318
|
+
organization_id="organization_id",
|
|
1319
|
+
user_id="user_id",
|
|
1316
1320
|
)
|
|
1317
1321
|
"""
|
|
1318
1322
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1351,9 +1355,9 @@ class AsyncOrganizationsClient:
|
|
|
1351
1355
|
token="YOUR_TOKEN",
|
|
1352
1356
|
)
|
|
1353
1357
|
await client.organizations.add_user_to_project(
|
|
1354
|
-
organization_id="
|
|
1355
|
-
user_id="
|
|
1356
|
-
project_id="
|
|
1358
|
+
organization_id="organization_id",
|
|
1359
|
+
user_id="user_id",
|
|
1360
|
+
project_id="project_id",
|
|
1357
1361
|
)
|
|
1358
1362
|
"""
|
|
1359
1363
|
_response = await self._client_wrapper.httpx_client.request(
|
|
@@ -1393,9 +1397,9 @@ class AsyncOrganizationsClient:
|
|
|
1393
1397
|
token="YOUR_TOKEN",
|
|
1394
1398
|
)
|
|
1395
1399
|
await client.organizations.remove_user_from_project(
|
|
1396
|
-
organization_id="
|
|
1397
|
-
user_id="
|
|
1398
|
-
project_id="
|
|
1400
|
+
organization_id="organization_id",
|
|
1401
|
+
user_id="user_id",
|
|
1402
|
+
project_id="project_id",
|
|
1399
1403
|
)
|
|
1400
1404
|
"""
|
|
1401
1405
|
_response = await self._client_wrapper.httpx_client.request(
|