llama-cloud 0.0.14__py3-none-any.whl → 0.0.16__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (35) hide show
  1. llama_cloud/__init__.py +20 -0
  2. llama_cloud/resources/__init__.py +2 -0
  3. llama_cloud/resources/files/client.py +159 -0
  4. llama_cloud/resources/parsing/client.py +40 -0
  5. llama_cloud/resources/pipelines/__init__.py +2 -0
  6. llama_cloud/resources/pipelines/client.py +188 -2
  7. llama_cloud/resources/pipelines/types/__init__.py +2 -0
  8. llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py +11 -0
  9. llama_cloud/types/__init__.py +18 -0
  10. llama_cloud/types/cloud_az_storage_blob_data_source.py +1 -2
  11. llama_cloud/types/cloud_postgres_vector_store.py +6 -8
  12. llama_cloud/types/configurable_transformation_names.py +4 -0
  13. llama_cloud/types/configured_transformation_item_component_one.py +2 -0
  14. llama_cloud/types/extend_vertex_text_embedding.py +58 -0
  15. llama_cloud/types/llama_parse_parameters.py +3 -1
  16. llama_cloud/types/llm_model_data.py +1 -0
  17. llama_cloud/types/llm_parameters.py +4 -1
  18. llama_cloud/types/page_screenshot_metadata.py +33 -0
  19. llama_cloud/types/page_screenshot_node_with_score.py +38 -0
  20. llama_cloud/types/pipeline.py +4 -0
  21. llama_cloud/types/pipeline_configuration_hashes.py +37 -0
  22. llama_cloud/types/pipeline_create_embedding_config.py +11 -0
  23. llama_cloud/types/pipeline_data_source.py +7 -0
  24. llama_cloud/types/pipeline_data_source_create.py +3 -0
  25. llama_cloud/types/pipeline_embedding_config.py +11 -0
  26. llama_cloud/types/pipeline_file.py +4 -0
  27. llama_cloud/types/pipeline_file_config_hash_value.py +5 -0
  28. llama_cloud/types/preset_retrieval_params.py +1 -0
  29. llama_cloud/types/retrieve_results.py +4 -0
  30. llama_cloud/types/vertex_ai_embedding_config.py +34 -0
  31. llama_cloud/types/vertex_embedding_mode.py +45 -0
  32. {llama_cloud-0.0.14.dist-info → llama_cloud-0.0.16.dist-info}/METADATA +1 -1
  33. {llama_cloud-0.0.14.dist-info → llama_cloud-0.0.16.dist-info}/RECORD +35 -28
  34. {llama_cloud-0.0.14.dist-info → llama_cloud-0.0.16.dist-info}/LICENSE +0 -0
  35. {llama_cloud-0.0.14.dist-info → llama_cloud-0.0.16.dist-info}/WHEEL +0 -0
llama_cloud/__init__.py CHANGED
@@ -79,6 +79,7 @@ from .types import (
79
79
  EvalQuestion,
80
80
  EvalQuestionCreate,
81
81
  EvalQuestionResult,
82
+ ExtendVertexTextEmbedding,
82
83
  ExtractionJob,
83
84
  ExtractionResult,
84
85
  ExtractionResultDataValue,
@@ -124,6 +125,8 @@ from .types import (
124
125
  OpenAiEmbeddingConfig,
125
126
  Organization,
126
127
  OrganizationCreate,
128
+ PageScreenshotMetadata,
129
+ PageScreenshotNodeWithScore,
127
130
  PageSegmentationConfig,
128
131
  PageSplitterNodeParser,
129
132
  ParserLanguages,
@@ -134,6 +137,7 @@ from .types import (
134
137
  ParsingJobTextResult,
135
138
  ParsingUsage,
136
139
  Pipeline,
140
+ PipelineConfigurationHashes,
137
141
  PipelineCreate,
138
142
  PipelineCreateEmbeddingConfig,
139
143
  PipelineCreateEmbeddingConfig_AzureEmbedding,
@@ -142,6 +146,7 @@ from .types import (
142
146
  PipelineCreateEmbeddingConfig_GeminiEmbedding,
143
147
  PipelineCreateEmbeddingConfig_HuggingfaceApiEmbedding,
144
148
  PipelineCreateEmbeddingConfig_OpenaiEmbedding,
149
+ PipelineCreateEmbeddingConfig_VertexaiEmbedding,
145
150
  PipelineCreateTransformConfig,
146
151
  PipelineCreateTransformConfig_Advanced,
147
152
  PipelineCreateTransformConfig_Auto,
@@ -158,7 +163,9 @@ from .types import (
158
163
  PipelineEmbeddingConfig_GeminiEmbedding,
159
164
  PipelineEmbeddingConfig_HuggingfaceApiEmbedding,
160
165
  PipelineEmbeddingConfig_OpenaiEmbedding,
166
+ PipelineEmbeddingConfig_VertexaiEmbedding,
161
167
  PipelineFile,
168
+ PipelineFileConfigHashValue,
162
169
  PipelineFileCreate,
163
170
  PipelineFileCreateCustomMetadataValue,
164
171
  PipelineFileCustomMetadataValue,
@@ -197,6 +204,8 @@ from .types import (
197
204
  UserOrganizationDelete,
198
205
  ValidationError,
199
206
  ValidationErrorLocItem,
207
+ VertexAiEmbeddingConfig,
208
+ VertexEmbeddingMode,
200
209
  )
201
210
  from .errors import UnprocessableEntityError
202
211
  from .resources import (
@@ -216,6 +225,7 @@ from .resources import (
216
225
  PipelineUpdateEmbeddingConfig_GeminiEmbedding,
217
226
  PipelineUpdateEmbeddingConfig_HuggingfaceApiEmbedding,
218
227
  PipelineUpdateEmbeddingConfig_OpenaiEmbedding,
228
+ PipelineUpdateEmbeddingConfig_VertexaiEmbedding,
219
229
  PipelineUpdateTransformConfig,
220
230
  PipelineUpdateTransformConfig_Advanced,
221
231
  PipelineUpdateTransformConfig_Auto,
@@ -317,6 +327,7 @@ __all__ = [
317
327
  "EvalQuestion",
318
328
  "EvalQuestionCreate",
319
329
  "EvalQuestionResult",
330
+ "ExtendVertexTextEmbedding",
320
331
  "ExtractionJob",
321
332
  "ExtractionResult",
322
333
  "ExtractionResultDataValue",
@@ -366,6 +377,8 @@ __all__ = [
366
377
  "OpenAiEmbeddingConfig",
367
378
  "Organization",
368
379
  "OrganizationCreate",
380
+ "PageScreenshotMetadata",
381
+ "PageScreenshotNodeWithScore",
369
382
  "PageSegmentationConfig",
370
383
  "PageSplitterNodeParser",
371
384
  "ParserLanguages",
@@ -376,6 +389,7 @@ __all__ = [
376
389
  "ParsingJobTextResult",
377
390
  "ParsingUsage",
378
391
  "Pipeline",
392
+ "PipelineConfigurationHashes",
379
393
  "PipelineCreate",
380
394
  "PipelineCreateEmbeddingConfig",
381
395
  "PipelineCreateEmbeddingConfig_AzureEmbedding",
@@ -384,6 +398,7 @@ __all__ = [
384
398
  "PipelineCreateEmbeddingConfig_GeminiEmbedding",
385
399
  "PipelineCreateEmbeddingConfig_HuggingfaceApiEmbedding",
386
400
  "PipelineCreateEmbeddingConfig_OpenaiEmbedding",
401
+ "PipelineCreateEmbeddingConfig_VertexaiEmbedding",
387
402
  "PipelineCreateTransformConfig",
388
403
  "PipelineCreateTransformConfig_Advanced",
389
404
  "PipelineCreateTransformConfig_Auto",
@@ -400,7 +415,9 @@ __all__ = [
400
415
  "PipelineEmbeddingConfig_GeminiEmbedding",
401
416
  "PipelineEmbeddingConfig_HuggingfaceApiEmbedding",
402
417
  "PipelineEmbeddingConfig_OpenaiEmbedding",
418
+ "PipelineEmbeddingConfig_VertexaiEmbedding",
403
419
  "PipelineFile",
420
+ "PipelineFileConfigHashValue",
404
421
  "PipelineFileCreate",
405
422
  "PipelineFileCreateCustomMetadataValue",
406
423
  "PipelineFileCustomMetadataValue",
@@ -417,6 +434,7 @@ __all__ = [
417
434
  "PipelineUpdateEmbeddingConfig_GeminiEmbedding",
418
435
  "PipelineUpdateEmbeddingConfig_HuggingfaceApiEmbedding",
419
436
  "PipelineUpdateEmbeddingConfig_OpenaiEmbedding",
437
+ "PipelineUpdateEmbeddingConfig_VertexaiEmbedding",
420
438
  "PipelineUpdateTransformConfig",
421
439
  "PipelineUpdateTransformConfig_Advanced",
422
440
  "PipelineUpdateTransformConfig_Auto",
@@ -451,6 +469,8 @@ __all__ = [
451
469
  "UserOrganizationDelete",
452
470
  "ValidationError",
453
471
  "ValidationErrorLocItem",
472
+ "VertexAiEmbeddingConfig",
473
+ "VertexEmbeddingMode",
454
474
  "auth",
455
475
  "component_definitions",
456
476
  "data_sinks",
@@ -26,6 +26,7 @@ from .pipelines import (
26
26
  PipelineUpdateEmbeddingConfig_GeminiEmbedding,
27
27
  PipelineUpdateEmbeddingConfig_HuggingfaceApiEmbedding,
28
28
  PipelineUpdateEmbeddingConfig_OpenaiEmbedding,
29
+ PipelineUpdateEmbeddingConfig_VertexaiEmbedding,
29
30
  PipelineUpdateTransformConfig,
30
31
  PipelineUpdateTransformConfig_Advanced,
31
32
  PipelineUpdateTransformConfig_Auto,
@@ -48,6 +49,7 @@ __all__ = [
48
49
  "PipelineUpdateEmbeddingConfig_GeminiEmbedding",
49
50
  "PipelineUpdateEmbeddingConfig_HuggingfaceApiEmbedding",
50
51
  "PipelineUpdateEmbeddingConfig_OpenaiEmbedding",
52
+ "PipelineUpdateEmbeddingConfig_VertexaiEmbedding",
51
53
  "PipelineUpdateTransformConfig",
52
54
  "PipelineUpdateTransformConfig_Advanced",
53
55
  "PipelineUpdateTransformConfig_Auto",
@@ -12,6 +12,7 @@ from ...core.remove_none_from_dict import remove_none_from_dict
12
12
  from ...errors.unprocessable_entity_error import UnprocessableEntityError
13
13
  from ...types.file import File
14
14
  from ...types.http_validation_error import HttpValidationError
15
+ from ...types.page_screenshot_metadata import PageScreenshotMetadata
15
16
  from ...types.presigned_url import PresignedUrl
16
17
  from .types.file_create_resource_info_value import FileCreateResourceInfoValue
17
18
 
@@ -288,6 +289,85 @@ class FilesClient:
288
289
  raise ApiError(status_code=_response.status_code, body=_response.text)
289
290
  raise ApiError(status_code=_response.status_code, body=_response_json)
290
291
 
292
+ def list_file_page_screenshots(
293
+ self, id: str, *, project_id: typing.Optional[str] = None
294
+ ) -> typing.List[PageScreenshotMetadata]:
295
+ """
296
+ List metadata for all screenshots of pages from a file.
297
+
298
+ Parameters:
299
+ - id: str.
300
+
301
+ - project_id: typing.Optional[str].
302
+ ---
303
+ from llama_cloud.client import LlamaCloud
304
+
305
+ client = LlamaCloud(
306
+ token="YOUR_TOKEN",
307
+ )
308
+ client.files.list_file_page_screenshots(
309
+ id="string",
310
+ )
311
+ """
312
+ _response = self._client_wrapper.httpx_client.request(
313
+ "GET",
314
+ urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/files/{id}/page_screenshots"),
315
+ params=remove_none_from_dict({"project_id": project_id}),
316
+ headers=self._client_wrapper.get_headers(),
317
+ timeout=60,
318
+ )
319
+ if 200 <= _response.status_code < 300:
320
+ return pydantic.parse_obj_as(typing.List[PageScreenshotMetadata], _response.json()) # type: ignore
321
+ if _response.status_code == 422:
322
+ raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
323
+ try:
324
+ _response_json = _response.json()
325
+ except JSONDecodeError:
326
+ raise ApiError(status_code=_response.status_code, body=_response.text)
327
+ raise ApiError(status_code=_response.status_code, body=_response_json)
328
+
329
+ def get_file_page_screenshot(
330
+ self, id: str, page_index: int, *, project_id: typing.Optional[str] = None
331
+ ) -> typing.Any:
332
+ """
333
+ Get screenshot of a page from a file.
334
+
335
+ Parameters:
336
+ - id: str.
337
+
338
+ - page_index: int.
339
+
340
+ - project_id: typing.Optional[str].
341
+ ---
342
+ from llama_cloud.client import LlamaCloud
343
+
344
+ client = LlamaCloud(
345
+ token="YOUR_TOKEN",
346
+ )
347
+ client.files.get_file_page_screenshot(
348
+ id="string",
349
+ page_index=1,
350
+ )
351
+ """
352
+ _response = self._client_wrapper.httpx_client.request(
353
+ "GET",
354
+ urllib.parse.urljoin(
355
+ f"{self._client_wrapper.get_base_url()}/", f"api/v1/files/{id}/page_screenshots/{page_index}"
356
+ ),
357
+ params=remove_none_from_dict({"project_id": project_id}),
358
+ headers=self._client_wrapper.get_headers(),
359
+ timeout=60,
360
+ )
361
+ if 200 <= _response.status_code < 300:
362
+ return pydantic.parse_obj_as(typing.Any, _response.json()) # type: ignore
363
+ if _response.status_code == 422:
364
+ raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
365
+ try:
366
+ _response_json = _response.json()
367
+ except JSONDecodeError:
368
+ raise ApiError(status_code=_response.status_code, body=_response.text)
369
+ raise ApiError(status_code=_response.status_code, body=_response_json)
370
+
291
371
 
292
372
  class AsyncFilesClient:
293
373
  def __init__(self, *, client_wrapper: AsyncClientWrapper):
@@ -549,3 +629,82 @@ class AsyncFilesClient:
549
629
  except JSONDecodeError:
550
630
  raise ApiError(status_code=_response.status_code, body=_response.text)
551
631
  raise ApiError(status_code=_response.status_code, body=_response_json)
632
+
633
+ async def list_file_page_screenshots(
634
+ self, id: str, *, project_id: typing.Optional[str] = None
635
+ ) -> typing.List[PageScreenshotMetadata]:
636
+ """
637
+ List metadata for all screenshots of pages from a file.
638
+
639
+ Parameters:
640
+ - id: str.
641
+
642
+ - project_id: typing.Optional[str].
643
+ ---
644
+ from llama_cloud.client import AsyncLlamaCloud
645
+
646
+ client = AsyncLlamaCloud(
647
+ token="YOUR_TOKEN",
648
+ )
649
+ await client.files.list_file_page_screenshots(
650
+ id="string",
651
+ )
652
+ """
653
+ _response = await self._client_wrapper.httpx_client.request(
654
+ "GET",
655
+ urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/files/{id}/page_screenshots"),
656
+ params=remove_none_from_dict({"project_id": project_id}),
657
+ headers=self._client_wrapper.get_headers(),
658
+ timeout=60,
659
+ )
660
+ if 200 <= _response.status_code < 300:
661
+ return pydantic.parse_obj_as(typing.List[PageScreenshotMetadata], _response.json()) # type: ignore
662
+ if _response.status_code == 422:
663
+ raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
664
+ try:
665
+ _response_json = _response.json()
666
+ except JSONDecodeError:
667
+ raise ApiError(status_code=_response.status_code, body=_response.text)
668
+ raise ApiError(status_code=_response.status_code, body=_response_json)
669
+
670
+ async def get_file_page_screenshot(
671
+ self, id: str, page_index: int, *, project_id: typing.Optional[str] = None
672
+ ) -> typing.Any:
673
+ """
674
+ Get screenshot of a page from a file.
675
+
676
+ Parameters:
677
+ - id: str.
678
+
679
+ - page_index: int.
680
+
681
+ - project_id: typing.Optional[str].
682
+ ---
683
+ from llama_cloud.client import AsyncLlamaCloud
684
+
685
+ client = AsyncLlamaCloud(
686
+ token="YOUR_TOKEN",
687
+ )
688
+ await client.files.get_file_page_screenshot(
689
+ id="string",
690
+ page_index=1,
691
+ )
692
+ """
693
+ _response = await self._client_wrapper.httpx_client.request(
694
+ "GET",
695
+ urllib.parse.urljoin(
696
+ f"{self._client_wrapper.get_base_url()}/", f"api/v1/files/{id}/page_screenshots/{page_index}"
697
+ ),
698
+ params=remove_none_from_dict({"project_id": project_id}),
699
+ headers=self._client_wrapper.get_headers(),
700
+ timeout=60,
701
+ )
702
+ if 200 <= _response.status_code < 300:
703
+ return pydantic.parse_obj_as(typing.Any, _response.json()) # type: ignore
704
+ if _response.status_code == 422:
705
+ raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
706
+ try:
707
+ _response_json = _response.json()
708
+ except JSONDecodeError:
709
+ raise ApiError(status_code=_response.status_code, body=_response.text)
710
+ raise ApiError(status_code=_response.status_code, body=_response_json)
@@ -118,7 +118,12 @@ class ParsingClient:
118
118
  vendor_multimodal_api_key: str,
119
119
  page_prefix: str,
120
120
  page_suffix: str,
121
+ webhook_url: str,
121
122
  take_screenshot: bool,
123
+ disable_ocr: bool,
124
+ disable_reconstruction: bool,
125
+ input_s_3_path: str,
126
+ output_s_3_path_prefix: str,
122
127
  file: typing.IO,
123
128
  ) -> ParsingJob:
124
129
  """
@@ -159,8 +164,18 @@ class ParsingClient:
159
164
 
160
165
  - page_suffix: str.
161
166
 
167
+ - webhook_url: str.
168
+
162
169
  - take_screenshot: bool.
163
170
 
171
+ - disable_ocr: bool.
172
+
173
+ - disable_reconstruction: bool.
174
+
175
+ - input_s_3_path: str.
176
+
177
+ - output_s_3_path_prefix: str.
178
+
164
179
  - file: typing.IO.
165
180
  """
166
181
  _response = self._client_wrapper.httpx_client.request(
@@ -185,7 +200,12 @@ class ParsingClient:
185
200
  "vendor_multimodal_api_key": vendor_multimodal_api_key,
186
201
  "page_prefix": page_prefix,
187
202
  "page_suffix": page_suffix,
203
+ "webhook_url": webhook_url,
188
204
  "take_screenshot": take_screenshot,
205
+ "disable_ocr": disable_ocr,
206
+ "disable_reconstruction": disable_reconstruction,
207
+ "input_s3_path": input_s_3_path,
208
+ "output_s3_path_prefix": output_s_3_path_prefix,
189
209
  }
190
210
  ),
191
211
  files={"file": file},
@@ -647,7 +667,12 @@ class AsyncParsingClient:
647
667
  vendor_multimodal_api_key: str,
648
668
  page_prefix: str,
649
669
  page_suffix: str,
670
+ webhook_url: str,
650
671
  take_screenshot: bool,
672
+ disable_ocr: bool,
673
+ disable_reconstruction: bool,
674
+ input_s_3_path: str,
675
+ output_s_3_path_prefix: str,
651
676
  file: typing.IO,
652
677
  ) -> ParsingJob:
653
678
  """
@@ -688,8 +713,18 @@ class AsyncParsingClient:
688
713
 
689
714
  - page_suffix: str.
690
715
 
716
+ - webhook_url: str.
717
+
691
718
  - take_screenshot: bool.
692
719
 
720
+ - disable_ocr: bool.
721
+
722
+ - disable_reconstruction: bool.
723
+
724
+ - input_s_3_path: str.
725
+
726
+ - output_s_3_path_prefix: str.
727
+
693
728
  - file: typing.IO.
694
729
  """
695
730
  _response = await self._client_wrapper.httpx_client.request(
@@ -714,7 +749,12 @@ class AsyncParsingClient:
714
749
  "vendor_multimodal_api_key": vendor_multimodal_api_key,
715
750
  "page_prefix": page_prefix,
716
751
  "page_suffix": page_suffix,
752
+ "webhook_url": webhook_url,
717
753
  "take_screenshot": take_screenshot,
754
+ "disable_ocr": disable_ocr,
755
+ "disable_reconstruction": disable_reconstruction,
756
+ "input_s3_path": input_s_3_path,
757
+ "output_s3_path_prefix": output_s_3_path_prefix,
718
758
  }
719
759
  ),
720
760
  files={"file": file},
@@ -9,6 +9,7 @@ from .types import (
9
9
  PipelineUpdateEmbeddingConfig_GeminiEmbedding,
10
10
  PipelineUpdateEmbeddingConfig_HuggingfaceApiEmbedding,
11
11
  PipelineUpdateEmbeddingConfig_OpenaiEmbedding,
12
+ PipelineUpdateEmbeddingConfig_VertexaiEmbedding,
12
13
  PipelineUpdateTransformConfig,
13
14
  PipelineUpdateTransformConfig_Advanced,
14
15
  PipelineUpdateTransformConfig_Auto,
@@ -23,6 +24,7 @@ __all__ = [
23
24
  "PipelineUpdateEmbeddingConfig_GeminiEmbedding",
24
25
  "PipelineUpdateEmbeddingConfig_HuggingfaceApiEmbedding",
25
26
  "PipelineUpdateEmbeddingConfig_OpenaiEmbedding",
27
+ "PipelineUpdateEmbeddingConfig_VertexaiEmbedding",
26
28
  "PipelineUpdateTransformConfig",
27
29
  "PipelineUpdateTransformConfig_Advanced",
28
30
  "PipelineUpdateTransformConfig_Auto",