llama-cloud 0.1.23__py3-none-any.whl → 0.1.25__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (31) hide show
  1. llama_cloud/__init__.py +4 -8
  2. llama_cloud/client.py +0 -3
  3. llama_cloud/resources/__init__.py +0 -2
  4. llama_cloud/resources/files/client.py +5 -4
  5. llama_cloud/resources/jobs/client.py +8 -0
  6. llama_cloud/resources/llama_extract/client.py +92 -24
  7. llama_cloud/resources/organizations/client.py +14 -4
  8. llama_cloud/resources/parsing/client.py +8 -0
  9. llama_cloud/resources/pipelines/client.py +20 -0
  10. llama_cloud/types/__init__.py +4 -6
  11. llama_cloud/types/composite_retrieval_result.py +5 -1
  12. llama_cloud/types/extract_config.py +3 -0
  13. llama_cloud/types/extract_models.py +20 -8
  14. llama_cloud/types/{llm_config_result.py → file_id_presigned_url.py} +9 -5
  15. llama_cloud/types/legacy_parse_job_config.py +1 -0
  16. llama_cloud/types/llama_extract_settings.py +3 -1
  17. llama_cloud/types/llama_parse_parameters.py +1 -0
  18. llama_cloud/types/page_figure_metadata.py +1 -0
  19. llama_cloud/types/{llm_configs_response.py → page_figure_node_with_score.py} +9 -4
  20. llama_cloud/types/parse_job_config.py +1 -0
  21. llama_cloud/types/preset_retrieval_params.py +6 -0
  22. llama_cloud/types/retrieve_results.py +5 -1
  23. llama_cloud/types/supported_llm_model_names.py +12 -4
  24. llama_cloud/types/user_organization_delete.py +1 -0
  25. {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/METADATA +1 -1
  26. {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/RECORD +28 -31
  27. llama_cloud/resources/admin/__init__.py +0 -2
  28. llama_cloud/resources/admin/client.py +0 -78
  29. llama_cloud/types/llm_config_result_llm_type.py +0 -33
  30. {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/LICENSE +0 -0
  31. {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/WHEEL +0 -0
@@ -1050,6 +1050,8 @@ class PipelinesClient:
1050
1050
  files_top_k: typing.Optional[int] = OMIT,
1051
1051
  retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
1052
1052
  retrieve_image_nodes: typing.Optional[bool] = OMIT,
1053
+ retrieve_page_screenshot_nodes: typing.Optional[bool] = OMIT,
1054
+ retrieve_page_figure_nodes: typing.Optional[bool] = OMIT,
1053
1055
  query: str,
1054
1056
  class_name: typing.Optional[str] = OMIT,
1055
1057
  ) -> RetrieveResults:
@@ -1085,6 +1087,10 @@ class PipelinesClient:
1085
1087
 
1086
1088
  - retrieve_image_nodes: typing.Optional[bool]. Whether to retrieve image nodes.
1087
1089
 
1090
+ - retrieve_page_screenshot_nodes: typing.Optional[bool]. Whether to retrieve page screenshot nodes.
1091
+
1092
+ - retrieve_page_figure_nodes: typing.Optional[bool]. Whether to retrieve page figure nodes.
1093
+
1088
1094
  - query: str. The query to retrieve against.
1089
1095
 
1090
1096
  - class_name: typing.Optional[str].
@@ -1128,6 +1134,10 @@ class PipelinesClient:
1128
1134
  _request["retrieval_mode"] = retrieval_mode
1129
1135
  if retrieve_image_nodes is not OMIT:
1130
1136
  _request["retrieve_image_nodes"] = retrieve_image_nodes
1137
+ if retrieve_page_screenshot_nodes is not OMIT:
1138
+ _request["retrieve_page_screenshot_nodes"] = retrieve_page_screenshot_nodes
1139
+ if retrieve_page_figure_nodes is not OMIT:
1140
+ _request["retrieve_page_figure_nodes"] = retrieve_page_figure_nodes
1131
1141
  if class_name is not OMIT:
1132
1142
  _request["class_name"] = class_name
1133
1143
  _response = self._client_wrapper.httpx_client.request(
@@ -2681,6 +2691,8 @@ class AsyncPipelinesClient:
2681
2691
  files_top_k: typing.Optional[int] = OMIT,
2682
2692
  retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
2683
2693
  retrieve_image_nodes: typing.Optional[bool] = OMIT,
2694
+ retrieve_page_screenshot_nodes: typing.Optional[bool] = OMIT,
2695
+ retrieve_page_figure_nodes: typing.Optional[bool] = OMIT,
2684
2696
  query: str,
2685
2697
  class_name: typing.Optional[str] = OMIT,
2686
2698
  ) -> RetrieveResults:
@@ -2716,6 +2728,10 @@ class AsyncPipelinesClient:
2716
2728
 
2717
2729
  - retrieve_image_nodes: typing.Optional[bool]. Whether to retrieve image nodes.
2718
2730
 
2731
+ - retrieve_page_screenshot_nodes: typing.Optional[bool]. Whether to retrieve page screenshot nodes.
2732
+
2733
+ - retrieve_page_figure_nodes: typing.Optional[bool]. Whether to retrieve page figure nodes.
2734
+
2719
2735
  - query: str. The query to retrieve against.
2720
2736
 
2721
2737
  - class_name: typing.Optional[str].
@@ -2759,6 +2775,10 @@ class AsyncPipelinesClient:
2759
2775
  _request["retrieval_mode"] = retrieval_mode
2760
2776
  if retrieve_image_nodes is not OMIT:
2761
2777
  _request["retrieve_image_nodes"] = retrieve_image_nodes
2778
+ if retrieve_page_screenshot_nodes is not OMIT:
2779
+ _request["retrieve_page_screenshot_nodes"] = retrieve_page_screenshot_nodes
2780
+ if retrieve_page_figure_nodes is not OMIT:
2781
+ _request["retrieve_page_figure_nodes"] = retrieve_page_figure_nodes
2762
2782
  if class_name is not OMIT:
2763
2783
  _request["class_name"] = class_name
2764
2784
  _response = await self._client_wrapper.httpx_client.request(
@@ -132,6 +132,7 @@ from .extract_target import ExtractTarget
132
132
  from .fail_page_mode import FailPageMode
133
133
  from .file import File
134
134
  from .file_count_by_status_response import FileCountByStatusResponse
135
+ from .file_id_presigned_url import FileIdPresignedUrl
135
136
  from .file_parse_public import FileParsePublic
136
137
  from .file_permission_info_value import FilePermissionInfoValue
137
138
  from .file_resource_info_value import FileResourceInfoValue
@@ -176,9 +177,6 @@ from .llama_index_core_base_llms_types_chat_message_blocks_item import (
176
177
  )
177
178
  from .llama_parse_parameters import LlamaParseParameters
178
179
  from .llama_parse_supported_file_extensions import LlamaParseSupportedFileExtensions
179
- from .llm_config_result import LlmConfigResult
180
- from .llm_config_result_llm_type import LlmConfigResultLlmType
181
- from .llm_configs_response import LlmConfigsResponse
182
180
  from .llm_model_data import LlmModelData
183
181
  from .llm_parameters import LlmParameters
184
182
  from .load_files_job_config import LoadFilesJobConfig
@@ -199,6 +197,7 @@ from .open_ai_embedding_config import OpenAiEmbeddingConfig
199
197
  from .organization import Organization
200
198
  from .organization_create import OrganizationCreate
201
199
  from .page_figure_metadata import PageFigureMetadata
200
+ from .page_figure_node_with_score import PageFigureNodeWithScore
202
201
  from .page_screenshot_metadata import PageScreenshotMetadata
203
202
  from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
204
203
  from .page_segmentation_config import PageSegmentationConfig
@@ -469,6 +468,7 @@ __all__ = [
469
468
  "FailPageMode",
470
469
  "File",
471
470
  "FileCountByStatusResponse",
471
+ "FileIdPresignedUrl",
472
472
  "FileParsePublic",
473
473
  "FilePermissionInfoValue",
474
474
  "FileResourceInfoValue",
@@ -509,9 +509,6 @@ __all__ = [
509
509
  "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text",
510
510
  "LlamaParseParameters",
511
511
  "LlamaParseSupportedFileExtensions",
512
- "LlmConfigResult",
513
- "LlmConfigResultLlmType",
514
- "LlmConfigsResponse",
515
512
  "LlmModelData",
516
513
  "LlmParameters",
517
514
  "LoadFilesJobConfig",
@@ -532,6 +529,7 @@ __all__ = [
532
529
  "Organization",
533
530
  "OrganizationCreate",
534
531
  "PageFigureMetadata",
532
+ "PageFigureNodeWithScore",
535
533
  "PageScreenshotMetadata",
536
534
  "PageScreenshotNodeWithScore",
537
535
  "PageSegmentationConfig",
@@ -5,6 +5,7 @@ import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .composite_retrieved_text_node_with_score import CompositeRetrievedTextNodeWithScore
8
+ from .page_figure_node_with_score import PageFigureNodeWithScore
8
9
  from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
9
10
 
10
11
  try:
@@ -21,7 +22,10 @@ class CompositeRetrievalResult(pydantic.BaseModel):
21
22
  description="The retrieved nodes from the composite retrieval."
22
23
  )
23
24
  image_nodes: typing.Optional[typing.List[PageScreenshotNodeWithScore]] = pydantic.Field(
24
- description="The image nodes retrieved by the pipeline for the given query."
25
+ description="The image nodes retrieved by the pipeline for the given query. Deprecated - will soon be replaced with 'page_screenshot_nodes'."
26
+ )
27
+ page_figure_nodes: typing.Optional[typing.List[PageFigureNodeWithScore]] = pydantic.Field(
28
+ description="The page figure nodes retrieved by the pipeline for the given query."
25
29
  )
26
30
 
27
31
  def json(self, **kwargs: typing.Any) -> str:
@@ -24,6 +24,9 @@ class ExtractConfig(pydantic.BaseModel):
24
24
 
25
25
  extraction_target: typing.Optional[ExtractTarget] = pydantic.Field(description="The extraction target specified.")
26
26
  extraction_mode: typing.Optional[ExtractMode] = pydantic.Field(description="The extraction mode specified.")
27
+ multimodal_fast_mode: typing.Optional[bool] = pydantic.Field(
28
+ description="Whether to use fast mode for multimodal extraction."
29
+ )
27
30
  system_prompt: typing.Optional[str]
28
31
  use_reasoning: typing.Optional[bool] = pydantic.Field(description="Whether to use reasoning for the extraction.")
29
32
  cite_sources: typing.Optional[bool] = pydantic.Field(description="Whether to cite sources for the extraction.")
@@ -7,27 +7,39 @@ T_Result = typing.TypeVar("T_Result")
7
7
 
8
8
 
9
9
  class ExtractModels(str, enum.Enum):
10
- GPT_4_O = "gpt-4o"
11
- GPT_4_O_MINI = "gpt-4o-mini"
12
10
  GPT_41 = "gpt-4.1"
13
11
  GPT_41_MINI = "gpt-4.1-mini"
12
+ GEMINI_20_FLASH = "gemini-2.0-flash"
14
13
  O_3_MINI = "o3-mini"
14
+ GEMINI_25_FLASH = "gemini-2.5-flash"
15
+ GEMINI_25_PRO = "gemini-2.5-pro"
16
+ GPT_4_O = "gpt-4o"
17
+ GPT_4_O_MINI = "gpt-4o-mini"
15
18
 
16
19
  def visit(
17
20
  self,
18
- gpt_4_o: typing.Callable[[], T_Result],
19
- gpt_4_o_mini: typing.Callable[[], T_Result],
20
21
  gpt_41: typing.Callable[[], T_Result],
21
22
  gpt_41_mini: typing.Callable[[], T_Result],
23
+ gemini_20_flash: typing.Callable[[], T_Result],
22
24
  o_3_mini: typing.Callable[[], T_Result],
25
+ gemini_25_flash: typing.Callable[[], T_Result],
26
+ gemini_25_pro: typing.Callable[[], T_Result],
27
+ gpt_4_o: typing.Callable[[], T_Result],
28
+ gpt_4_o_mini: typing.Callable[[], T_Result],
23
29
  ) -> T_Result:
24
- if self is ExtractModels.GPT_4_O:
25
- return gpt_4_o()
26
- if self is ExtractModels.GPT_4_O_MINI:
27
- return gpt_4_o_mini()
28
30
  if self is ExtractModels.GPT_41:
29
31
  return gpt_41()
30
32
  if self is ExtractModels.GPT_41_MINI:
31
33
  return gpt_41_mini()
34
+ if self is ExtractModels.GEMINI_20_FLASH:
35
+ return gemini_20_flash()
32
36
  if self is ExtractModels.O_3_MINI:
33
37
  return o_3_mini()
38
+ if self is ExtractModels.GEMINI_25_FLASH:
39
+ return gemini_25_flash()
40
+ if self is ExtractModels.GEMINI_25_PRO:
41
+ return gemini_25_pro()
42
+ if self is ExtractModels.GPT_4_O:
43
+ return gpt_4_o()
44
+ if self is ExtractModels.GPT_4_O_MINI:
45
+ return gpt_4_o_mini()
@@ -4,7 +4,6 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
- from .llm_config_result_llm_type import LlmConfigResultLlmType
8
7
 
9
8
  try:
10
9
  import pydantic
@@ -15,10 +14,15 @@ except ImportError:
15
14
  import pydantic # type: ignore
16
15
 
17
16
 
18
- class LlmConfigResult(pydantic.BaseModel):
19
- llm_type: LlmConfigResultLlmType
20
- valid: bool
21
- error_message: typing.Optional[str]
17
+ class FileIdPresignedUrl(pydantic.BaseModel):
18
+ """
19
+ Schema for a presigned URL with a file ID.
20
+ """
21
+
22
+ url: str = pydantic.Field(description="A presigned URL for IO operations against a private file")
23
+ expires_at: dt.datetime = pydantic.Field(description="The time at which the presigned URL expires")
24
+ form_fields: typing.Optional[typing.Dict[str, typing.Optional[str]]]
25
+ file_id: str = pydantic.Field(description="The ID of the file associated with the presigned URL")
22
26
 
23
27
  def json(self, **kwargs: typing.Any) -> str:
24
28
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -45,6 +45,7 @@ class LegacyParseJobConfig(pydantic.BaseModel):
45
45
  )
46
46
  invalidate_cache: bool = pydantic.Field(alias="invalidateCache", description="Whether to invalidate the cache.")
47
47
  output_pdf_of_document: typing.Optional[bool] = pydantic.Field(alias="outputPDFOfDocument")
48
+ outlined_table_extraction: typing.Optional[bool] = pydantic.Field(alias="outlinedTableExtraction")
48
49
  save_images: typing.Optional[bool] = pydantic.Field(alias="saveImages")
49
50
  gpt_4_o: typing.Optional[bool] = pydantic.Field(alias="gpt4o", description="Whether to use GPT4o.")
50
51
  open_aiapi_key: str = pydantic.Field(alias="openAIAPIKey", description="The OpenAI API key.")
@@ -39,7 +39,9 @@ class LlamaExtractSettings(pydantic.BaseModel):
39
39
  extraction_agent_config: typing.Optional[typing.Dict[str, StructParseConf]] = pydantic.Field(
40
40
  description="The configuration for the extraction agent."
41
41
  )
42
- use_multimodal_extraction: typing.Optional[bool]
42
+ use_pixel_extraction: typing.Optional[bool] = pydantic.Field(
43
+ description="Whether to use extraction over pixels for multimodal mode."
44
+ )
43
45
  llama_parse_params: typing.Optional[LlamaParseParameters] = pydantic.Field(
44
46
  description="LlamaParse related settings."
45
47
  )
@@ -31,6 +31,7 @@ class LlamaParseParameters(pydantic.BaseModel):
31
31
  disable_reconstruction: typing.Optional[bool]
32
32
  disable_image_extraction: typing.Optional[bool]
33
33
  invalidate_cache: typing.Optional[bool]
34
+ outlined_table_extraction: typing.Optional[bool]
34
35
  output_pdf_of_document: typing.Optional[bool]
35
36
  do_not_cache: typing.Optional[bool]
36
37
  fast_mode: typing.Optional[bool]
@@ -21,6 +21,7 @@ class PageFigureMetadata(pydantic.BaseModel):
21
21
  figure_size: int = pydantic.Field(description="The size of the figure in bytes")
22
22
  is_likely_noise: typing.Optional[bool] = pydantic.Field(description="Whether the figure is likely to be noise")
23
23
  confidence: float = pydantic.Field(description="The confidence of the figure")
24
+ metadata: typing.Optional[typing.Dict[str, typing.Any]]
24
25
 
25
26
  def json(self, **kwargs: typing.Any) -> str:
26
27
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -4,7 +4,7 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
- from .llm_config_result import LlmConfigResult
7
+ from .page_figure_metadata import PageFigureMetadata
8
8
 
9
9
  try:
10
10
  import pydantic
@@ -15,9 +15,14 @@ except ImportError:
15
15
  import pydantic # type: ignore
16
16
 
17
17
 
18
- class LlmConfigsResponse(pydantic.BaseModel):
19
- llm_configs: typing.List[LlmConfigResult]
20
- last_validated_at: str
18
+ class PageFigureNodeWithScore(pydantic.BaseModel):
19
+ """
20
+ Page figure metadata with score
21
+ """
22
+
23
+ node: PageFigureMetadata
24
+ score: float = pydantic.Field(description="The score of the figure node")
25
+ class_name: typing.Optional[str]
21
26
 
22
27
  def json(self, **kwargs: typing.Any) -> str:
23
28
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -33,6 +33,7 @@ class ParseJobConfig(pydantic.BaseModel):
33
33
  disable_reconstruction: typing.Optional[bool]
34
34
  disable_image_extraction: typing.Optional[bool]
35
35
  invalidate_cache: typing.Optional[bool]
36
+ outlined_table_extraction: typing.Optional[bool]
36
37
  output_pdf_of_document: typing.Optional[bool]
37
38
  do_not_cache: typing.Optional[bool]
38
39
  fast_mode: typing.Optional[bool]
@@ -37,6 +37,12 @@ class PresetRetrievalParams(pydantic.BaseModel):
37
37
  files_top_k: typing.Optional[int]
38
38
  retrieval_mode: typing.Optional[RetrievalMode] = pydantic.Field(description="The retrieval mode for the query.")
39
39
  retrieve_image_nodes: typing.Optional[bool] = pydantic.Field(description="Whether to retrieve image nodes.")
40
+ retrieve_page_screenshot_nodes: typing.Optional[bool] = pydantic.Field(
41
+ description="Whether to retrieve page screenshot nodes."
42
+ )
43
+ retrieve_page_figure_nodes: typing.Optional[bool] = pydantic.Field(
44
+ description="Whether to retrieve page figure nodes."
45
+ )
40
46
  class_name: typing.Optional[str]
41
47
 
42
48
  def json(self, **kwargs: typing.Any) -> str:
@@ -5,6 +5,7 @@ import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .metadata_filters import MetadataFilters
8
+ from .page_figure_node_with_score import PageFigureNodeWithScore
8
9
  from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
9
10
  from .text_node_with_score import TextNodeWithScore
10
11
 
@@ -27,7 +28,10 @@ class RetrieveResults(pydantic.BaseModel):
27
28
  description="The nodes retrieved by the pipeline for the given query."
28
29
  )
29
30
  image_nodes: typing.Optional[typing.List[PageScreenshotNodeWithScore]] = pydantic.Field(
30
- description="The image nodes retrieved by the pipeline for the given query."
31
+ description="The image nodes retrieved by the pipeline for the given query. Deprecated - will soon be replaced with 'page_screenshot_nodes'."
32
+ )
33
+ page_figure_nodes: typing.Optional[typing.List[PageFigureNodeWithScore]] = pydantic.Field(
34
+ description="The page figure nodes retrieved by the pipeline for the given query."
31
35
  )
32
36
  retrieval_latency: typing.Optional[typing.Dict[str, float]] = pydantic.Field(
33
37
  description="The end-to-end latency for retrieval and reranking."
@@ -9,9 +9,11 @@ T_Result = typing.TypeVar("T_Result")
9
9
  class SupportedLlmModelNames(str, enum.Enum):
10
10
  GPT_4_O = "GPT_4O"
11
11
  GPT_4_O_MINI = "GPT_4O_MINI"
12
+ GPT_4_1 = "GPT_4_1"
13
+ GPT_4_1_NANO = "GPT_4_1_NANO"
14
+ GPT_4_1_MINI = "GPT_4_1_MINI"
12
15
  AZURE_OPENAI_GPT_4_O = "AZURE_OPENAI_GPT_4O"
13
16
  AZURE_OPENAI_GPT_4_O_MINI = "AZURE_OPENAI_GPT_4O_MINI"
14
- AZURE_OPENAI_GPT_4 = "AZURE_OPENAI_GPT_4"
15
17
  CLAUDE_3_5_SONNET = "CLAUDE_3_5_SONNET"
16
18
  BEDROCK_CLAUDE_3_5_SONNET_V_1 = "BEDROCK_CLAUDE_3_5_SONNET_V1"
17
19
  BEDROCK_CLAUDE_3_5_SONNET_V_2 = "BEDROCK_CLAUDE_3_5_SONNET_V2"
@@ -21,9 +23,11 @@ class SupportedLlmModelNames(str, enum.Enum):
21
23
  self,
22
24
  gpt_4_o: typing.Callable[[], T_Result],
23
25
  gpt_4_o_mini: typing.Callable[[], T_Result],
26
+ gpt_4_1: typing.Callable[[], T_Result],
27
+ gpt_4_1_nano: typing.Callable[[], T_Result],
28
+ gpt_4_1_mini: typing.Callable[[], T_Result],
24
29
  azure_openai_gpt_4_o: typing.Callable[[], T_Result],
25
30
  azure_openai_gpt_4_o_mini: typing.Callable[[], T_Result],
26
- azure_openai_gpt_4: typing.Callable[[], T_Result],
27
31
  claude_3_5_sonnet: typing.Callable[[], T_Result],
28
32
  bedrock_claude_3_5_sonnet_v_1: typing.Callable[[], T_Result],
29
33
  bedrock_claude_3_5_sonnet_v_2: typing.Callable[[], T_Result],
@@ -33,12 +37,16 @@ class SupportedLlmModelNames(str, enum.Enum):
33
37
  return gpt_4_o()
34
38
  if self is SupportedLlmModelNames.GPT_4_O_MINI:
35
39
  return gpt_4_o_mini()
40
+ if self is SupportedLlmModelNames.GPT_4_1:
41
+ return gpt_4_1()
42
+ if self is SupportedLlmModelNames.GPT_4_1_NANO:
43
+ return gpt_4_1_nano()
44
+ if self is SupportedLlmModelNames.GPT_4_1_MINI:
45
+ return gpt_4_1_mini()
36
46
  if self is SupportedLlmModelNames.AZURE_OPENAI_GPT_4_O:
37
47
  return azure_openai_gpt_4_o()
38
48
  if self is SupportedLlmModelNames.AZURE_OPENAI_GPT_4_O_MINI:
39
49
  return azure_openai_gpt_4_o_mini()
40
- if self is SupportedLlmModelNames.AZURE_OPENAI_GPT_4:
41
- return azure_openai_gpt_4()
42
50
  if self is SupportedLlmModelNames.CLAUDE_3_5_SONNET:
43
51
  return claude_3_5_sonnet()
44
52
  if self is SupportedLlmModelNames.BEDROCK_CLAUDE_3_5_SONNET_V_1:
@@ -21,6 +21,7 @@ class UserOrganizationDelete(pydantic.BaseModel):
21
21
 
22
22
  user_id: typing.Optional[str]
23
23
  email: typing.Optional[str]
24
+ project_id_list: typing.Optional[typing.List[str]]
24
25
 
25
26
  def json(self, **kwargs: typing.Any) -> str:
26
27
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: llama-cloud
3
- Version: 0.1.23
3
+ Version: 0.1.25
4
4
  Summary:
5
5
  License: MIT
6
6
  Author: Logan Markewich
@@ -1,5 +1,5 @@
1
- llama_cloud/__init__.py,sha256=7ghywbsF_IGjBlM0qQ_ndijMnC4GL1KroZgXD8IWNpI,24345
2
- llama_cloud/client.py,sha256=JGV9tNaEgSsDy8vYv7RMPXTmXwbfc9gPvuEfQHYqrq4,5651
1
+ llama_cloud/__init__.py,sha256=8-yO1moSZzcHW8aAcqSu8z8Un9oMBnlV703ExtGQtYw,24279
2
+ llama_cloud/client.py,sha256=NuwtJJei6Wsa25GbYpxQnd_ZcD9_FxDJZp3Tu5OL914,5440
3
3
  llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
4
4
  llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
5
5
  llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
@@ -9,9 +9,7 @@ llama_cloud/core/remove_none_from_dict.py,sha256=8m91FC3YuVem0Gm9_sXhJ2tGvP33owJ
9
9
  llama_cloud/environment.py,sha256=feTjOebeFZMrBdnHat4RE5aHlpt-sJm4NhK4ntV1htI,167
10
10
  llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6g,170
11
11
  llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
12
- llama_cloud/resources/__init__.py,sha256=bDSUCwFqDzA8hFiWENlmoIsOcGu2rUZgWnCtVETKwVs,4145
13
- llama_cloud/resources/admin/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
14
- llama_cloud/resources/admin/client.py,sha256=tIfM5KMJXRL0AUAm_s_fx5OzgqMUIxksjhCGuviQXQk,3080
12
+ llama_cloud/resources/__init__.py,sha256=tVfSvZW9QzKBpC1XjqTMjZsyaMVHfetmwBpCPSjEGHk,4121
15
13
  llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
16
14
  llama_cloud/resources/beta/client.py,sha256=eRB3mGmNxbhVGTtUpp-j-2APkHUoCbUckIz9coYjCsM,14666
17
15
  llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
@@ -32,15 +30,15 @@ llama_cloud/resources/embedding_model_configs/types/embedding_model_config_creat
32
30
  llama_cloud/resources/evals/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
33
31
  llama_cloud/resources/evals/client.py,sha256=v2AyeQV0hVgC6xoP2gJNgneJMaeXALV1hIeirYGxlPw,3242
34
32
  llama_cloud/resources/files/__init__.py,sha256=3B0SNM8EE6PddD5LpxYllci9vflEXy1xjPzhEEd-OUk,293
35
- llama_cloud/resources/files/client.py,sha256=oPwDQAkf0zN1mxP_vT6Songp4scOq5k0jcfHo-zfCtY,50560
33
+ llama_cloud/resources/files/client.py,sha256=gEk7nhX2sZE3PYFCfvdeYtJfTK_vMlWLHtjsl73p73c,50646
36
34
  llama_cloud/resources/files/types/__init__.py,sha256=EPYENAwkjBWv1MLf8s7R5-RO-cxZ_8NPrqfR4ZoR7jY,418
37
35
  llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py,sha256=Wc8wFgujOO5pZvbbh2TMMzpa37GKZd14GYNJ9bdq7BE,214
38
36
  llama_cloud/resources/files/types/file_create_permission_info_value.py,sha256=KPCFuEaa8NiB85A5MfdXRAQ0poAUTl7Feg6BTfmdWas,209
39
37
  llama_cloud/resources/files/types/file_create_resource_info_value.py,sha256=R7Y-CJf7fnbvIqE3xOI5XOrmPwLbVJLC7zpxMu8Zopk,201
40
38
  llama_cloud/resources/jobs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
41
- llama_cloud/resources/jobs/client.py,sha256=tIrIVOZ6p60qTJ2dqhSrqHGGxLi0fzn9V7jJU9QS2Nw,6030
39
+ llama_cloud/resources/jobs/client.py,sha256=gv_N8e0lay7cjt6MCwx-Cj4FiCXKhbyCDaWbadaJpgY,6270
42
40
  llama_cloud/resources/llama_extract/__init__.py,sha256=jRUugj6XARMpKZi3e2RkfTdcOSuE-Zy0IfScRLlyYMs,819
43
- llama_cloud/resources/llama_extract/client.py,sha256=oL5_3ugPde9Ch8QnwvRvFeTMMl0y_ST6i1hoSt2S5E4,63503
41
+ llama_cloud/resources/llama_extract/client.py,sha256=XU9FyO4Q6R61VOYDRjwddhJWhONZncJ5R9-JE6iEHZY,66605
44
42
  llama_cloud/resources/llama_extract/types/__init__.py,sha256=ZRBD-jg1qdXyiJKTxgH7zaadoDzuof1TYpjK4P5z4zA,1216
45
43
  llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py,sha256=zB31hJQ8hKaIsPkfTWiX5hqsPVFMyyeWEDZ_Aq237jo,305
46
44
  llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py,sha256=xoyXH3f0Y5beMWBxmtXSz6QoB_df_-0QBsYdjBhZnGw,217
@@ -51,11 +49,11 @@ llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_o
51
49
  llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema.py,sha256=uMqpKJdCmUNtryS2bkQTNA1AgDlWdtsBOP31iMt3zNA,346
52
50
  llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema_zero_value.py,sha256=cUS7ez5r0Vx8T7SxwLYptZMmvpT5JoDVMyn54Q6VL-g,227
53
51
  llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
54
- llama_cloud/resources/organizations/client.py,sha256=56d5VcRg_3Lu_MMdfJCR8uoirBoxmj9vaQbHpwY2zWo,56243
52
+ llama_cloud/resources/organizations/client.py,sha256=CdrdNdB9R-bOsNqZ4Jbm1BzG1RafXMFjuCsrVYf2OrE,56567
55
53
  llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
56
- llama_cloud/resources/parsing/client.py,sha256=U5J2g78CWYgujfLXlLEIoqatFSPnaG1OYNanx1jm_GQ,78177
54
+ llama_cloud/resources/parsing/client.py,sha256=5xa47aiBNAjZzvKJNfmu5vM0ODuDzpWt-6SPOYTwhSg,78491
57
55
  llama_cloud/resources/pipelines/__init__.py,sha256=zyvVEOF_krvEZkCIj_kZoMKfhDqHo_R32a1mv9CriQc,1193
58
- llama_cloud/resources/pipelines/client.py,sha256=Vfth-8S_25P5jkaBAMoNA6Ve-5-iQhZLAwQxaUkDkJ0,131104
56
+ llama_cloud/resources/pipelines/client.py,sha256=BcBqzTPu1LUsdimXvuaaKjUu6w5xjbL-ZBfWsO183Vk,132360
59
57
  llama_cloud/resources/pipelines/types/__init__.py,sha256=C68NQ5QzA0dFXf9oePFFGmV1vn96jcAp-QAznSgoRYQ,1375
60
58
  llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
61
59
  llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
@@ -69,7 +67,7 @@ llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwH
69
67
  llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
70
68
  llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
71
69
  llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
72
- llama_cloud/types/__init__.py,sha256=-Xa6LnSne7UtW1wBpw0_MeNuXT58mZV1iohj-Vfus9A,28708
70
+ llama_cloud/types/__init__.py,sha256=q70Gh26R9ZvPLFZakF5ajLT4uXj6uoAw29ej2P7HkGM,28642
73
71
  llama_cloud/types/advanced_mode_transform_config.py,sha256=4xCXye0_cPmVS1F8aNTx81sIaEPjQH9kiCCAIoqUzlI,1502
74
72
  llama_cloud/types/advanced_mode_transform_config_chunking_config.py,sha256=wYbJnWLpeQDfhmDZz-wJfYzD1iGT5Jcxb9ga3mzUuvk,1983
75
73
  llama_cloud/types/advanced_mode_transform_config_segmentation_config.py,sha256=anNGq0F5-IlbIW3kpC8OilzLJnUq5tdIcWHnRnmlYsg,1303
@@ -116,7 +114,7 @@ llama_cloud/types/cloud_slack_data_source.py,sha256=tlsNj-hDj1gWmM0Q2A1BeyolfaPg
116
114
  llama_cloud/types/cohere_embedding.py,sha256=wkv_fVCA1WEroGawzPFExwmiJ75gPfzeeemty7NBlsM,1579
117
115
  llama_cloud/types/cohere_embedding_config.py,sha256=c0Kj1wuSsBX9TQ2AondKv5ZtX5PmkivsHj6P0M7tVB4,1142
118
116
  llama_cloud/types/composite_retrieval_mode.py,sha256=PtN0vQ90xyAJL4vyGRG4lMNOpnJ__2L1xiwosI9yfms,548
119
- llama_cloud/types/composite_retrieval_result.py,sha256=1GmLnT-PlpXdURfXn8vaWdEL9BjuWV-AyjqjPvJ4YGk,1479
117
+ llama_cloud/types/composite_retrieval_result.py,sha256=EulVseVvpK50kto4wQweLO7jJe6l6Ym1erKa4dOl4CU,1801
120
118
  llama_cloud/types/composite_retrieved_text_node.py,sha256=eTQ99cdZ2PASff5n4oVV1oaNiS9Ie3AtY_E55kBYpBs,1702
121
119
  llama_cloud/types/composite_retrieved_text_node_with_score.py,sha256=o-HvmyjqODc68zYuobtj10_62FMBAKRLfRoTHGDdmxw,1148
122
120
  llama_cloud/types/configurable_data_sink_names.py,sha256=0Yk9i8hcNXKCcSKpa5KwsCwy_EDeodqbny7qmF86_lM,1225
@@ -147,13 +145,13 @@ llama_cloud/types/embedding_model_config_update_embedding_config.py,sha256=mrXFx
147
145
  llama_cloud/types/eval_execution_params.py,sha256=ntVaJh5SMZMPL4QLUiihVjUlg2SKbrezvbMKGlrF66Q,1369
148
146
  llama_cloud/types/extract_agent.py,sha256=T98IOueut4M52Qm7hqcUOcWFFDhZ-ye0OFdXgfFGtS4,1763
149
147
  llama_cloud/types/extract_agent_data_schema_value.py,sha256=UaDQ2KjajLDccW7F4NKdfpefeTJrr1hl0c95WRETYkM,201
150
- llama_cloud/types/extract_config.py,sha256=tCVzcctlEaO4tx0ikAv0OyfiDDhiiynR2FYILgOCCe4,1936
148
+ llama_cloud/types/extract_config.py,sha256=j3792FdRUPpS6U3ynIDIUBaOTVSZfoNV3aBTnma_IjA,2082
151
149
  llama_cloud/types/extract_job.py,sha256=Yx4fDdCdylAji2LPTwqflVpz1o9slpj9tTLS93-1tzU,1431
152
150
  llama_cloud/types/extract_job_create.py,sha256=UK1mBIKyflo7e6m1MxMN95pLscj67jH_yvs8EvmBXqU,1545
153
151
  llama_cloud/types/extract_job_create_data_schema_override.py,sha256=vuiJ2lGJjbXEnvFKzVnKyvgwhMXPg1Pb5GZne2DrB60,330
154
152
  llama_cloud/types/extract_job_create_data_schema_override_zero_value.py,sha256=HHEYxOSQXXyBYOiUQg_qwfQtXFj-OtThMwbUDBIgZU0,223
155
153
  llama_cloud/types/extract_mode.py,sha256=DwTMzDq3HHJop_fxQelHEE_k8UcdDz-W_v_Oj2WWXLk,931
156
- llama_cloud/types/extract_models.py,sha256=qXsSXuPFiVBNsh1-4oTYAjJ_36AQJ7sOJyJ-5aNlYEE,977
154
+ llama_cloud/types/extract_models.py,sha256=GrNd7MPOfB7G4zzAtfTPfLw2EKJFoY_NJ9hbd-wZQfU,1519
157
155
  llama_cloud/types/extract_resultset.py,sha256=Alje0YQJUiA_aKi0hQs7TAnhDmZuQ_yL9b6HCNYBFQg,1627
158
156
  llama_cloud/types/extract_resultset_data.py,sha256=v9Ae4SxLsvYPE9crko4N16lBjsxuZpz1yrUOhnaM_VY,427
159
157
  llama_cloud/types/extract_resultset_data_item_value.py,sha256=JwqgDIGW0irr8QWaSTIrl24FhGxTUDOXIbxoSdIjuxs,209
@@ -172,6 +170,7 @@ llama_cloud/types/extract_target.py,sha256=Gt-FNqblzcjdfq1hxsqEjWWu-HNLXdKy4w98n
172
170
  llama_cloud/types/fail_page_mode.py,sha256=n4fgPpiEB5siPoEg0Sux4COg7ElNybjshxDoUihZwRU,786
173
171
  llama_cloud/types/file.py,sha256=rQXitPRKOYw91nK5qOZ0vpOmIx_MCpRb0g78d9dQs6w,1822
174
172
  llama_cloud/types/file_count_by_status_response.py,sha256=WuorbZvKjDs9Ql1hUiQu4gN5iCm8d6fr92KLyHpRvQU,1356
173
+ llama_cloud/types/file_id_presigned_url.py,sha256=Yr_MGFKbuBEHK4efRSK53fHcoo5bbAKnqQGGhMycUc0,1398
175
174
  llama_cloud/types/file_parse_public.py,sha256=sshZ0BcjHMGpuz4ylSurv0K_3ejfPrUGGyDxBHCtdMg,1378
176
175
  llama_cloud/types/file_permission_info_value.py,sha256=RyQlNbhvIKS87Ywu7XUaw5jDToZX64M9Wqzu1U_q2Us,197
177
176
  llama_cloud/types/file_resource_info_value.py,sha256=g6T6ELeLK9jgcvX6r-EuAl_4JkwnyqdS0RRoabMReSU,195
@@ -193,15 +192,12 @@ llama_cloud/types/job_record.py,sha256=7hdDPZU11EG8g6_9iq6vy-zqLEryeC7i8fZ-CkUB_
193
192
  llama_cloud/types/job_record_parameters.py,sha256=Oqxp5y0owPfjLc_NR7AYE8P3zM2PJo36N9olbyNl7AA,3425
194
193
  llama_cloud/types/job_record_with_usage_metrics.py,sha256=iNV2do5TB_0e3PoOz_DJyAaM6Cn9G8KG-dGPGgEs5SY,1198
195
194
  llama_cloud/types/l_lama_parse_transform_config.py,sha256=YQRJZvKh1Ee2FUyW_N0nqYJoW599qBgH3JCH9SH6YLo,1249
196
- llama_cloud/types/legacy_parse_job_config.py,sha256=kVBdiSLraI9rKQOPf0Ci9RtbNLkco0byBJC42uE_PCI,11698
197
- llama_cloud/types/llama_extract_settings.py,sha256=IQFxtKa4GtHKc9w-fLwsH0LSKDWzR9_vZ_cTFJ9cGBI,2288
195
+ llama_cloud/types/legacy_parse_job_config.py,sha256=4cC-p4d5bwkKnn_VZBuf7pCpwGFaD0QyJd5cPicaryY,11801
196
+ llama_cloud/types/llama_extract_settings.py,sha256=bHtF5AD0r896-248e7WKthcbbvrAUdptZrENP2Ed4LM,2388
198
197
  llama_cloud/types/llama_index_core_base_llms_types_chat_message.py,sha256=NelHo-T-ebVMhRKsqE_xV8AJW4c7o6lS0uEQnPsmTwg,1365
199
198
  llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py,sha256=-aL8fh-w2Xf4uQs_LHzb3q6LL_onLAcVzCR5yMI4qJw,1571
200
- llama_cloud/types/llama_parse_parameters.py,sha256=DNhVZm3YQ_3xZiz7WUrwH7E6jqW2fZ7YGFsdfsYalUk,5773
199
+ llama_cloud/types/llama_parse_parameters.py,sha256=Ap-F-8giC0pa45blPxMLNHJ-3zEgJc1zrNM1a9NWErs,5826
201
200
  llama_cloud/types/llama_parse_supported_file_extensions.py,sha256=B_0N3f8Aq59W9FbsH50mGBUiyWTIXQjHFl739uAyaQw,11207
202
- llama_cloud/types/llm_config_result.py,sha256=3NRAvq_jVJlkbLRDgBHTGMDli0av7d0GJME3aiXytKs,1106
203
- llama_cloud/types/llm_config_result_llm_type.py,sha256=yrijlC2f1egNDx-tCvDVp68pFmGaJZvVE_D1vS2wA34,1032
204
- llama_cloud/types/llm_configs_response.py,sha256=TQ9RLkOzdY-8k0l1NNUZNUl83dmKks0gOQvhj-4hRZs,1073
205
201
  llama_cloud/types/llm_model_data.py,sha256=6rrycqGwlK3LZ2S-WtgmeomithdLhDCgwBBZQ5KLaso,1300
206
202
  llama_cloud/types/llm_parameters.py,sha256=RTKYt09lm9a1MlnBfYuTP2x_Ww4byUNNc1TqIel5O1Y,1377
207
203
  llama_cloud/types/load_files_job_config.py,sha256=R5sFgFmV__0mqLUuD7dkFoBJHG2ZLw5px9zRapvYcpE,1069
@@ -221,7 +217,8 @@ llama_cloud/types/open_ai_embedding.py,sha256=RQijkvKyzbISy92LnBSEpjmIU8p7kMpdc4
221
217
  llama_cloud/types/open_ai_embedding_config.py,sha256=Mquc0JrtCo8lVYA2WW7q0ZikS3HRkiMtzDFu5XA-20o,1143
222
218
  llama_cloud/types/organization.py,sha256=p8mYRqSsGxw17AmdW8x8nP7P1UbdpYkwr51WTIjTVLw,1467
223
219
  llama_cloud/types/organization_create.py,sha256=hUXRwArIx_0D_lilpL7z-B0oJJ5yEX8Sbu2xqfH_9so,1086
224
- llama_cloud/types/page_figure_metadata.py,sha256=iIg6_f2SwJg6UcQo9X4MoSm_ygxnIBmFjS2LuUsI6qE,1528
220
+ llama_cloud/types/page_figure_metadata.py,sha256=0oasDkjnzoVQ4W-Ci0KoJHM0iHXTGvm3cbdVOgH9nHE,1588
221
+ llama_cloud/types/page_figure_node_with_score.py,sha256=VqNQx9RKmD_jY1kHPCvPjygshbfVLLSgtC5TX-Cy_cw,1208
225
222
  llama_cloud/types/page_screenshot_metadata.py,sha256=lobrq0AsOr8sDwMgA9ytop8lRmRFvJW2oiql3yLvbjM,1328
226
223
  llama_cloud/types/page_screenshot_node_with_score.py,sha256=EdqoXbmARCz1DV14E2saCPshIeII709uM4cLwxw_mkM,1232
227
224
  llama_cloud/types/page_segmentation_config.py,sha256=VH8uuxnubnJak1gSpS64OoMueHidhsDB-2eq2tVHbag,998
@@ -230,7 +227,7 @@ llama_cloud/types/paginated_jobs_history_with_metrics.py,sha256=Bxy6N0x0FARJhgwN
230
227
  llama_cloud/types/paginated_list_cloud_documents_response.py,sha256=MsjS0SWlT0syELDck4x2sxxR3_NC1e6QTdepgVmK9aY,1341
231
228
  llama_cloud/types/paginated_list_pipeline_files_response.py,sha256=2TKR2oHSQRyLMqWz1qQBSIvz-ZJb8U_94367lwOJ2S4,1317
232
229
  llama_cloud/types/paginated_report_response.py,sha256=o79QhQi9r0HZZrhvRlA6WGjxtyPuxN0xONhwXSwxtcs,1104
233
- llama_cloud/types/parse_job_config.py,sha256=KLBhRRGziH4eU2sZgab24c8-L9b8M9on1Dg0nVnObGc,6254
230
+ llama_cloud/types/parse_job_config.py,sha256=ajTNaLdfsuaimwL2PobOrTKvIe75SbSdIrDzJkFqp1U,6307
234
231
  llama_cloud/types/parse_plan_level.py,sha256=GBkDS19qfHseBa17EXfuTPNT4GNv5alyPrWEvWji3GY,528
235
232
  llama_cloud/types/parser_languages.py,sha256=Ps3IlaSt6tyxEI657N3-vZL96r2puk8wsf31cWnO-SI,10840
236
233
  llama_cloud/types/parsing_history_item.py,sha256=_MVzf43t84PbmjOzsMLZ_NBoyiisigLWz-fr0ZxU63g,1183
@@ -276,7 +273,7 @@ llama_cloud/types/plan_limits.py,sha256=WAbDbRl8gsQxvhmuVB0YT8mry-0uKg6c66uivypp
276
273
  llama_cloud/types/playground_session.py,sha256=F8u2KZL2YaOrsT-o1n4zbhyPxSsoduc3ZCzQB8AecFA,1858
277
274
  llama_cloud/types/pooling.py,sha256=5Fr6c8rx9SDWwWzEvD78suob2d79ktodUtLUAUHMbP8,651
278
275
  llama_cloud/types/preset_composite_retrieval_params.py,sha256=yEf1pk4Wz5J6SxgB8elklwuyVDCRSZqfWC6x3hJUS4Q,1366
279
- llama_cloud/types/preset_retrieval_params.py,sha256=tcWMNsT4IGVBqTYAj-yModGZbM4tk7mZIMGvpuBWH8Y,2079
276
+ llama_cloud/types/preset_retrieval_params.py,sha256=TcyljefpspJSveMR9L5DQHlqW4jZeexBsXus_LkHkJA,2365
280
277
  llama_cloud/types/preset_retrieval_params_search_filters_inference_schema_value.py,sha256=BOp-oJMIc3KVU89mmKIhVcwwsO0XBRnuErfsPqpUjSs,234
281
278
  llama_cloud/types/presigned_url.py,sha256=-DOQo7XKvUsl-9Gz7fX6VOHdQLzGH2XRau24ASvG92E,1275
282
279
  llama_cloud/types/progress_event.py,sha256=Bk73A8geTVaq0ze5pMnbkAmx7FSOHQIixYCpCas_dcY,1684
@@ -305,7 +302,7 @@ llama_cloud/types/report_state.py,sha256=gjexexoT8GaCamGKvfwivKrfRtvdhEtwSLkAt-j
305
302
  llama_cloud/types/report_state_event.py,sha256=_wf-Cl_skJdrag-7h11tz-HIy1jed_GIG3c-ksuAjT4,1270
306
303
  llama_cloud/types/report_update_event.py,sha256=uLRC79U3pvZ5-kY6pOseQyX1MNH-0m80GUtzpjd6mkI,1270
307
304
  llama_cloud/types/retrieval_mode.py,sha256=wV9q3OdHTuyDWbJCGdxq9Hw6U95WFlJcaMq6KWSTzyw,910
308
- llama_cloud/types/retrieve_results.py,sha256=ZuGIJrzxIegzsAopgA-X8690HrcYowhL4Xfhxm3m9H4,2053
305
+ llama_cloud/types/retrieve_results.py,sha256=rHArmu05K3NvIQepHX5nsVOfcMsZj3MaIcPkTC6mD_8,2375
309
306
  llama_cloud/types/retriever.py,sha256=ZItPsorL8x1XjtJT49ZodaMqU8h2GfwlB4U4cgnfZkM,1626
310
307
  llama_cloud/types/retriever_create.py,sha256=WyUR9DRzu3Q9tzKEeXCdQuzCY6WKi9ADJkZea9rqvxU,1286
311
308
  llama_cloud/types/retriever_pipeline.py,sha256=F1pZDxg8JdQXRHE6ciFezd7a-Wv5bHplPcGDED-J4b0,1330
@@ -317,7 +314,7 @@ llama_cloud/types/status_enum.py,sha256=cUBIlys89E8PUzmVqqawu7qTDF0aRqBwiijOmRDP
317
314
  llama_cloud/types/struct_mode.py,sha256=ROicwjXfFmgVU8_xSVxJlnFUzRNKG5VIEF1wYg9uOPU,1020
318
315
  llama_cloud/types/struct_parse_conf.py,sha256=WlL8y0IBvdzGsDtFUlEZLzoUODwmOWAJi0viS9unL18,2297
319
316
  llama_cloud/types/supported_llm_model.py,sha256=hubSopFICVNEegbJbtbpK6zRHwFPwUNtrw_NAw_3bfg,1380
320
- llama_cloud/types/supported_llm_model_names.py,sha256=ZBGVJDkoK4Kj0Z7DaQ7m-zQhymdf7_03_H4qJPrZ0FM,2213
317
+ llama_cloud/types/supported_llm_model_names.py,sha256=PXL0gA1lc0GJNzZHnjOscoxHpPW787A8Adh-2egAKo8,2512
321
318
  llama_cloud/types/text_block.py,sha256=X154sQkSyposXuRcEWNp_tWcDQ-AI6q_-MfJUN5exP8,958
322
319
  llama_cloud/types/text_node.py,sha256=Tq3QmuKC5cIHvC9wAtvhsXl1g2sACs2yJwQ0Uko8GSU,2846
323
320
  llama_cloud/types/text_node_relationships_value.py,sha256=qmXURTk1Xg7ZDzRSSV1uDEel0AXRLohND5ioezibHY0,217
@@ -330,14 +327,14 @@ llama_cloud/types/usage_response_active_alerts_item.py,sha256=5EgU7go_CPe2Bmio12
330
327
  llama_cloud/types/user_job_record.py,sha256=mJHdokJsemXJOwM2l7fsW3X0SlwSNcy7yHbcXZHh3I4,1098
331
328
  llama_cloud/types/user_organization.py,sha256=yKewpOrMcB-CbujGNTjkX6QiWYr5HVsRIFQ-WX8kp2I,1729
332
329
  llama_cloud/types/user_organization_create.py,sha256=Zj57s9xuYVnLW2p8i4j2QORL-G1y7Ab3avXE1baERQY,1189
333
- llama_cloud/types/user_organization_delete.py,sha256=IDYLKfFAXfcJfkEpA0ARbaA0JDcEBe7fTLv833DZXHs,1104
330
+ llama_cloud/types/user_organization_delete.py,sha256=bEfgQMdTd6oAMZXtvSm5BhZahG1wAVDBXZ8e7V9UN7w,1159
334
331
  llama_cloud/types/user_organization_role.py,sha256=vTM5pYG9NJpTQACn8vzSIt01Ul6jEHCVmyR3vV0isPg,1512
335
332
  llama_cloud/types/validation_error.py,sha256=yZDLtjUHDY5w82Ra6CW0H9sLAr18R0RY1UNgJKR72DQ,1084
336
333
  llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPXjdtN9EB7HrLVo6EP0,128
337
334
  llama_cloud/types/vertex_ai_embedding_config.py,sha256=DvQk2xMJFmo54MEXTzoM4KSADyhGm_ygmFyx6wIcQdw,1159
338
335
  llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2BaIMltDqGnIowU,1217
339
336
  llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
340
- llama_cloud-0.1.23.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
341
- llama_cloud-0.1.23.dist-info/METADATA,sha256=1aLs2Ad-tIIomTC_Nid1Tsedxj6ni0MPLxFa7u3xQEI,1143
342
- llama_cloud-0.1.23.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
343
- llama_cloud-0.1.23.dist-info/RECORD,,
337
+ llama_cloud-0.1.25.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
338
+ llama_cloud-0.1.25.dist-info/METADATA,sha256=EaOl_aKCf0kdbk6eSVh86xmUub8cGlQsk-XZEQIlEQ0,1143
339
+ llama_cloud-0.1.25.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
340
+ llama_cloud-0.1.25.dist-info/RECORD,,
@@ -1,2 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-