llama-cloud 0.1.23__py3-none-any.whl → 0.1.25__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +4 -8
- llama_cloud/client.py +0 -3
- llama_cloud/resources/__init__.py +0 -2
- llama_cloud/resources/files/client.py +5 -4
- llama_cloud/resources/jobs/client.py +8 -0
- llama_cloud/resources/llama_extract/client.py +92 -24
- llama_cloud/resources/organizations/client.py +14 -4
- llama_cloud/resources/parsing/client.py +8 -0
- llama_cloud/resources/pipelines/client.py +20 -0
- llama_cloud/types/__init__.py +4 -6
- llama_cloud/types/composite_retrieval_result.py +5 -1
- llama_cloud/types/extract_config.py +3 -0
- llama_cloud/types/extract_models.py +20 -8
- llama_cloud/types/{llm_config_result.py → file_id_presigned_url.py} +9 -5
- llama_cloud/types/legacy_parse_job_config.py +1 -0
- llama_cloud/types/llama_extract_settings.py +3 -1
- llama_cloud/types/llama_parse_parameters.py +1 -0
- llama_cloud/types/page_figure_metadata.py +1 -0
- llama_cloud/types/{llm_configs_response.py → page_figure_node_with_score.py} +9 -4
- llama_cloud/types/parse_job_config.py +1 -0
- llama_cloud/types/preset_retrieval_params.py +6 -0
- llama_cloud/types/retrieve_results.py +5 -1
- llama_cloud/types/supported_llm_model_names.py +12 -4
- llama_cloud/types/user_organization_delete.py +1 -0
- {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/METADATA +1 -1
- {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/RECORD +28 -31
- llama_cloud/resources/admin/__init__.py +0 -2
- llama_cloud/resources/admin/client.py +0 -78
- llama_cloud/types/llm_config_result_llm_type.py +0 -33
- {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/LICENSE +0 -0
- {llama_cloud-0.1.23.dist-info → llama_cloud-0.1.25.dist-info}/WHEEL +0 -0
|
@@ -1050,6 +1050,8 @@ class PipelinesClient:
|
|
|
1050
1050
|
files_top_k: typing.Optional[int] = OMIT,
|
|
1051
1051
|
retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
|
|
1052
1052
|
retrieve_image_nodes: typing.Optional[bool] = OMIT,
|
|
1053
|
+
retrieve_page_screenshot_nodes: typing.Optional[bool] = OMIT,
|
|
1054
|
+
retrieve_page_figure_nodes: typing.Optional[bool] = OMIT,
|
|
1053
1055
|
query: str,
|
|
1054
1056
|
class_name: typing.Optional[str] = OMIT,
|
|
1055
1057
|
) -> RetrieveResults:
|
|
@@ -1085,6 +1087,10 @@ class PipelinesClient:
|
|
|
1085
1087
|
|
|
1086
1088
|
- retrieve_image_nodes: typing.Optional[bool]. Whether to retrieve image nodes.
|
|
1087
1089
|
|
|
1090
|
+
- retrieve_page_screenshot_nodes: typing.Optional[bool]. Whether to retrieve page screenshot nodes.
|
|
1091
|
+
|
|
1092
|
+
- retrieve_page_figure_nodes: typing.Optional[bool]. Whether to retrieve page figure nodes.
|
|
1093
|
+
|
|
1088
1094
|
- query: str. The query to retrieve against.
|
|
1089
1095
|
|
|
1090
1096
|
- class_name: typing.Optional[str].
|
|
@@ -1128,6 +1134,10 @@ class PipelinesClient:
|
|
|
1128
1134
|
_request["retrieval_mode"] = retrieval_mode
|
|
1129
1135
|
if retrieve_image_nodes is not OMIT:
|
|
1130
1136
|
_request["retrieve_image_nodes"] = retrieve_image_nodes
|
|
1137
|
+
if retrieve_page_screenshot_nodes is not OMIT:
|
|
1138
|
+
_request["retrieve_page_screenshot_nodes"] = retrieve_page_screenshot_nodes
|
|
1139
|
+
if retrieve_page_figure_nodes is not OMIT:
|
|
1140
|
+
_request["retrieve_page_figure_nodes"] = retrieve_page_figure_nodes
|
|
1131
1141
|
if class_name is not OMIT:
|
|
1132
1142
|
_request["class_name"] = class_name
|
|
1133
1143
|
_response = self._client_wrapper.httpx_client.request(
|
|
@@ -2681,6 +2691,8 @@ class AsyncPipelinesClient:
|
|
|
2681
2691
|
files_top_k: typing.Optional[int] = OMIT,
|
|
2682
2692
|
retrieval_mode: typing.Optional[RetrievalMode] = OMIT,
|
|
2683
2693
|
retrieve_image_nodes: typing.Optional[bool] = OMIT,
|
|
2694
|
+
retrieve_page_screenshot_nodes: typing.Optional[bool] = OMIT,
|
|
2695
|
+
retrieve_page_figure_nodes: typing.Optional[bool] = OMIT,
|
|
2684
2696
|
query: str,
|
|
2685
2697
|
class_name: typing.Optional[str] = OMIT,
|
|
2686
2698
|
) -> RetrieveResults:
|
|
@@ -2716,6 +2728,10 @@ class AsyncPipelinesClient:
|
|
|
2716
2728
|
|
|
2717
2729
|
- retrieve_image_nodes: typing.Optional[bool]. Whether to retrieve image nodes.
|
|
2718
2730
|
|
|
2731
|
+
- retrieve_page_screenshot_nodes: typing.Optional[bool]. Whether to retrieve page screenshot nodes.
|
|
2732
|
+
|
|
2733
|
+
- retrieve_page_figure_nodes: typing.Optional[bool]. Whether to retrieve page figure nodes.
|
|
2734
|
+
|
|
2719
2735
|
- query: str. The query to retrieve against.
|
|
2720
2736
|
|
|
2721
2737
|
- class_name: typing.Optional[str].
|
|
@@ -2759,6 +2775,10 @@ class AsyncPipelinesClient:
|
|
|
2759
2775
|
_request["retrieval_mode"] = retrieval_mode
|
|
2760
2776
|
if retrieve_image_nodes is not OMIT:
|
|
2761
2777
|
_request["retrieve_image_nodes"] = retrieve_image_nodes
|
|
2778
|
+
if retrieve_page_screenshot_nodes is not OMIT:
|
|
2779
|
+
_request["retrieve_page_screenshot_nodes"] = retrieve_page_screenshot_nodes
|
|
2780
|
+
if retrieve_page_figure_nodes is not OMIT:
|
|
2781
|
+
_request["retrieve_page_figure_nodes"] = retrieve_page_figure_nodes
|
|
2762
2782
|
if class_name is not OMIT:
|
|
2763
2783
|
_request["class_name"] = class_name
|
|
2764
2784
|
_response = await self._client_wrapper.httpx_client.request(
|
llama_cloud/types/__init__.py
CHANGED
|
@@ -132,6 +132,7 @@ from .extract_target import ExtractTarget
|
|
|
132
132
|
from .fail_page_mode import FailPageMode
|
|
133
133
|
from .file import File
|
|
134
134
|
from .file_count_by_status_response import FileCountByStatusResponse
|
|
135
|
+
from .file_id_presigned_url import FileIdPresignedUrl
|
|
135
136
|
from .file_parse_public import FileParsePublic
|
|
136
137
|
from .file_permission_info_value import FilePermissionInfoValue
|
|
137
138
|
from .file_resource_info_value import FileResourceInfoValue
|
|
@@ -176,9 +177,6 @@ from .llama_index_core_base_llms_types_chat_message_blocks_item import (
|
|
|
176
177
|
)
|
|
177
178
|
from .llama_parse_parameters import LlamaParseParameters
|
|
178
179
|
from .llama_parse_supported_file_extensions import LlamaParseSupportedFileExtensions
|
|
179
|
-
from .llm_config_result import LlmConfigResult
|
|
180
|
-
from .llm_config_result_llm_type import LlmConfigResultLlmType
|
|
181
|
-
from .llm_configs_response import LlmConfigsResponse
|
|
182
180
|
from .llm_model_data import LlmModelData
|
|
183
181
|
from .llm_parameters import LlmParameters
|
|
184
182
|
from .load_files_job_config import LoadFilesJobConfig
|
|
@@ -199,6 +197,7 @@ from .open_ai_embedding_config import OpenAiEmbeddingConfig
|
|
|
199
197
|
from .organization import Organization
|
|
200
198
|
from .organization_create import OrganizationCreate
|
|
201
199
|
from .page_figure_metadata import PageFigureMetadata
|
|
200
|
+
from .page_figure_node_with_score import PageFigureNodeWithScore
|
|
202
201
|
from .page_screenshot_metadata import PageScreenshotMetadata
|
|
203
202
|
from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
|
|
204
203
|
from .page_segmentation_config import PageSegmentationConfig
|
|
@@ -469,6 +468,7 @@ __all__ = [
|
|
|
469
468
|
"FailPageMode",
|
|
470
469
|
"File",
|
|
471
470
|
"FileCountByStatusResponse",
|
|
471
|
+
"FileIdPresignedUrl",
|
|
472
472
|
"FileParsePublic",
|
|
473
473
|
"FilePermissionInfoValue",
|
|
474
474
|
"FileResourceInfoValue",
|
|
@@ -509,9 +509,6 @@ __all__ = [
|
|
|
509
509
|
"LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text",
|
|
510
510
|
"LlamaParseParameters",
|
|
511
511
|
"LlamaParseSupportedFileExtensions",
|
|
512
|
-
"LlmConfigResult",
|
|
513
|
-
"LlmConfigResultLlmType",
|
|
514
|
-
"LlmConfigsResponse",
|
|
515
512
|
"LlmModelData",
|
|
516
513
|
"LlmParameters",
|
|
517
514
|
"LoadFilesJobConfig",
|
|
@@ -532,6 +529,7 @@ __all__ = [
|
|
|
532
529
|
"Organization",
|
|
533
530
|
"OrganizationCreate",
|
|
534
531
|
"PageFigureMetadata",
|
|
532
|
+
"PageFigureNodeWithScore",
|
|
535
533
|
"PageScreenshotMetadata",
|
|
536
534
|
"PageScreenshotNodeWithScore",
|
|
537
535
|
"PageSegmentationConfig",
|
|
@@ -5,6 +5,7 @@ import typing
|
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .composite_retrieved_text_node_with_score import CompositeRetrievedTextNodeWithScore
|
|
8
|
+
from .page_figure_node_with_score import PageFigureNodeWithScore
|
|
8
9
|
from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
|
|
9
10
|
|
|
10
11
|
try:
|
|
@@ -21,7 +22,10 @@ class CompositeRetrievalResult(pydantic.BaseModel):
|
|
|
21
22
|
description="The retrieved nodes from the composite retrieval."
|
|
22
23
|
)
|
|
23
24
|
image_nodes: typing.Optional[typing.List[PageScreenshotNodeWithScore]] = pydantic.Field(
|
|
24
|
-
description="The image nodes retrieved by the pipeline for the given query."
|
|
25
|
+
description="The image nodes retrieved by the pipeline for the given query. Deprecated - will soon be replaced with 'page_screenshot_nodes'."
|
|
26
|
+
)
|
|
27
|
+
page_figure_nodes: typing.Optional[typing.List[PageFigureNodeWithScore]] = pydantic.Field(
|
|
28
|
+
description="The page figure nodes retrieved by the pipeline for the given query."
|
|
25
29
|
)
|
|
26
30
|
|
|
27
31
|
def json(self, **kwargs: typing.Any) -> str:
|
|
@@ -24,6 +24,9 @@ class ExtractConfig(pydantic.BaseModel):
|
|
|
24
24
|
|
|
25
25
|
extraction_target: typing.Optional[ExtractTarget] = pydantic.Field(description="The extraction target specified.")
|
|
26
26
|
extraction_mode: typing.Optional[ExtractMode] = pydantic.Field(description="The extraction mode specified.")
|
|
27
|
+
multimodal_fast_mode: typing.Optional[bool] = pydantic.Field(
|
|
28
|
+
description="Whether to use fast mode for multimodal extraction."
|
|
29
|
+
)
|
|
27
30
|
system_prompt: typing.Optional[str]
|
|
28
31
|
use_reasoning: typing.Optional[bool] = pydantic.Field(description="Whether to use reasoning for the extraction.")
|
|
29
32
|
cite_sources: typing.Optional[bool] = pydantic.Field(description="Whether to cite sources for the extraction.")
|
|
@@ -7,27 +7,39 @@ T_Result = typing.TypeVar("T_Result")
|
|
|
7
7
|
|
|
8
8
|
|
|
9
9
|
class ExtractModels(str, enum.Enum):
|
|
10
|
-
GPT_4_O = "gpt-4o"
|
|
11
|
-
GPT_4_O_MINI = "gpt-4o-mini"
|
|
12
10
|
GPT_41 = "gpt-4.1"
|
|
13
11
|
GPT_41_MINI = "gpt-4.1-mini"
|
|
12
|
+
GEMINI_20_FLASH = "gemini-2.0-flash"
|
|
14
13
|
O_3_MINI = "o3-mini"
|
|
14
|
+
GEMINI_25_FLASH = "gemini-2.5-flash"
|
|
15
|
+
GEMINI_25_PRO = "gemini-2.5-pro"
|
|
16
|
+
GPT_4_O = "gpt-4o"
|
|
17
|
+
GPT_4_O_MINI = "gpt-4o-mini"
|
|
15
18
|
|
|
16
19
|
def visit(
|
|
17
20
|
self,
|
|
18
|
-
gpt_4_o: typing.Callable[[], T_Result],
|
|
19
|
-
gpt_4_o_mini: typing.Callable[[], T_Result],
|
|
20
21
|
gpt_41: typing.Callable[[], T_Result],
|
|
21
22
|
gpt_41_mini: typing.Callable[[], T_Result],
|
|
23
|
+
gemini_20_flash: typing.Callable[[], T_Result],
|
|
22
24
|
o_3_mini: typing.Callable[[], T_Result],
|
|
25
|
+
gemini_25_flash: typing.Callable[[], T_Result],
|
|
26
|
+
gemini_25_pro: typing.Callable[[], T_Result],
|
|
27
|
+
gpt_4_o: typing.Callable[[], T_Result],
|
|
28
|
+
gpt_4_o_mini: typing.Callable[[], T_Result],
|
|
23
29
|
) -> T_Result:
|
|
24
|
-
if self is ExtractModels.GPT_4_O:
|
|
25
|
-
return gpt_4_o()
|
|
26
|
-
if self is ExtractModels.GPT_4_O_MINI:
|
|
27
|
-
return gpt_4_o_mini()
|
|
28
30
|
if self is ExtractModels.GPT_41:
|
|
29
31
|
return gpt_41()
|
|
30
32
|
if self is ExtractModels.GPT_41_MINI:
|
|
31
33
|
return gpt_41_mini()
|
|
34
|
+
if self is ExtractModels.GEMINI_20_FLASH:
|
|
35
|
+
return gemini_20_flash()
|
|
32
36
|
if self is ExtractModels.O_3_MINI:
|
|
33
37
|
return o_3_mini()
|
|
38
|
+
if self is ExtractModels.GEMINI_25_FLASH:
|
|
39
|
+
return gemini_25_flash()
|
|
40
|
+
if self is ExtractModels.GEMINI_25_PRO:
|
|
41
|
+
return gemini_25_pro()
|
|
42
|
+
if self is ExtractModels.GPT_4_O:
|
|
43
|
+
return gpt_4_o()
|
|
44
|
+
if self is ExtractModels.GPT_4_O_MINI:
|
|
45
|
+
return gpt_4_o_mini()
|
|
@@ -4,7 +4,6 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .llm_config_result_llm_type import LlmConfigResultLlmType
|
|
8
7
|
|
|
9
8
|
try:
|
|
10
9
|
import pydantic
|
|
@@ -15,10 +14,15 @@ except ImportError:
|
|
|
15
14
|
import pydantic # type: ignore
|
|
16
15
|
|
|
17
16
|
|
|
18
|
-
class
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
17
|
+
class FileIdPresignedUrl(pydantic.BaseModel):
|
|
18
|
+
"""
|
|
19
|
+
Schema for a presigned URL with a file ID.
|
|
20
|
+
"""
|
|
21
|
+
|
|
22
|
+
url: str = pydantic.Field(description="A presigned URL for IO operations against a private file")
|
|
23
|
+
expires_at: dt.datetime = pydantic.Field(description="The time at which the presigned URL expires")
|
|
24
|
+
form_fields: typing.Optional[typing.Dict[str, typing.Optional[str]]]
|
|
25
|
+
file_id: str = pydantic.Field(description="The ID of the file associated with the presigned URL")
|
|
22
26
|
|
|
23
27
|
def json(self, **kwargs: typing.Any) -> str:
|
|
24
28
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -45,6 +45,7 @@ class LegacyParseJobConfig(pydantic.BaseModel):
|
|
|
45
45
|
)
|
|
46
46
|
invalidate_cache: bool = pydantic.Field(alias="invalidateCache", description="Whether to invalidate the cache.")
|
|
47
47
|
output_pdf_of_document: typing.Optional[bool] = pydantic.Field(alias="outputPDFOfDocument")
|
|
48
|
+
outlined_table_extraction: typing.Optional[bool] = pydantic.Field(alias="outlinedTableExtraction")
|
|
48
49
|
save_images: typing.Optional[bool] = pydantic.Field(alias="saveImages")
|
|
49
50
|
gpt_4_o: typing.Optional[bool] = pydantic.Field(alias="gpt4o", description="Whether to use GPT4o.")
|
|
50
51
|
open_aiapi_key: str = pydantic.Field(alias="openAIAPIKey", description="The OpenAI API key.")
|
|
@@ -39,7 +39,9 @@ class LlamaExtractSettings(pydantic.BaseModel):
|
|
|
39
39
|
extraction_agent_config: typing.Optional[typing.Dict[str, StructParseConf]] = pydantic.Field(
|
|
40
40
|
description="The configuration for the extraction agent."
|
|
41
41
|
)
|
|
42
|
-
|
|
42
|
+
use_pixel_extraction: typing.Optional[bool] = pydantic.Field(
|
|
43
|
+
description="Whether to use extraction over pixels for multimodal mode."
|
|
44
|
+
)
|
|
43
45
|
llama_parse_params: typing.Optional[LlamaParseParameters] = pydantic.Field(
|
|
44
46
|
description="LlamaParse related settings."
|
|
45
47
|
)
|
|
@@ -31,6 +31,7 @@ class LlamaParseParameters(pydantic.BaseModel):
|
|
|
31
31
|
disable_reconstruction: typing.Optional[bool]
|
|
32
32
|
disable_image_extraction: typing.Optional[bool]
|
|
33
33
|
invalidate_cache: typing.Optional[bool]
|
|
34
|
+
outlined_table_extraction: typing.Optional[bool]
|
|
34
35
|
output_pdf_of_document: typing.Optional[bool]
|
|
35
36
|
do_not_cache: typing.Optional[bool]
|
|
36
37
|
fast_mode: typing.Optional[bool]
|
|
@@ -21,6 +21,7 @@ class PageFigureMetadata(pydantic.BaseModel):
|
|
|
21
21
|
figure_size: int = pydantic.Field(description="The size of the figure in bytes")
|
|
22
22
|
is_likely_noise: typing.Optional[bool] = pydantic.Field(description="Whether the figure is likely to be noise")
|
|
23
23
|
confidence: float = pydantic.Field(description="The confidence of the figure")
|
|
24
|
+
metadata: typing.Optional[typing.Dict[str, typing.Any]]
|
|
24
25
|
|
|
25
26
|
def json(self, **kwargs: typing.Any) -> str:
|
|
26
27
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -4,7 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .
|
|
7
|
+
from .page_figure_metadata import PageFigureMetadata
|
|
8
8
|
|
|
9
9
|
try:
|
|
10
10
|
import pydantic
|
|
@@ -15,9 +15,14 @@ except ImportError:
|
|
|
15
15
|
import pydantic # type: ignore
|
|
16
16
|
|
|
17
17
|
|
|
18
|
-
class
|
|
19
|
-
|
|
20
|
-
|
|
18
|
+
class PageFigureNodeWithScore(pydantic.BaseModel):
|
|
19
|
+
"""
|
|
20
|
+
Page figure metadata with score
|
|
21
|
+
"""
|
|
22
|
+
|
|
23
|
+
node: PageFigureMetadata
|
|
24
|
+
score: float = pydantic.Field(description="The score of the figure node")
|
|
25
|
+
class_name: typing.Optional[str]
|
|
21
26
|
|
|
22
27
|
def json(self, **kwargs: typing.Any) -> str:
|
|
23
28
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -33,6 +33,7 @@ class ParseJobConfig(pydantic.BaseModel):
|
|
|
33
33
|
disable_reconstruction: typing.Optional[bool]
|
|
34
34
|
disable_image_extraction: typing.Optional[bool]
|
|
35
35
|
invalidate_cache: typing.Optional[bool]
|
|
36
|
+
outlined_table_extraction: typing.Optional[bool]
|
|
36
37
|
output_pdf_of_document: typing.Optional[bool]
|
|
37
38
|
do_not_cache: typing.Optional[bool]
|
|
38
39
|
fast_mode: typing.Optional[bool]
|
|
@@ -37,6 +37,12 @@ class PresetRetrievalParams(pydantic.BaseModel):
|
|
|
37
37
|
files_top_k: typing.Optional[int]
|
|
38
38
|
retrieval_mode: typing.Optional[RetrievalMode] = pydantic.Field(description="The retrieval mode for the query.")
|
|
39
39
|
retrieve_image_nodes: typing.Optional[bool] = pydantic.Field(description="Whether to retrieve image nodes.")
|
|
40
|
+
retrieve_page_screenshot_nodes: typing.Optional[bool] = pydantic.Field(
|
|
41
|
+
description="Whether to retrieve page screenshot nodes."
|
|
42
|
+
)
|
|
43
|
+
retrieve_page_figure_nodes: typing.Optional[bool] = pydantic.Field(
|
|
44
|
+
description="Whether to retrieve page figure nodes."
|
|
45
|
+
)
|
|
40
46
|
class_name: typing.Optional[str]
|
|
41
47
|
|
|
42
48
|
def json(self, **kwargs: typing.Any) -> str:
|
|
@@ -5,6 +5,7 @@ import typing
|
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .metadata_filters import MetadataFilters
|
|
8
|
+
from .page_figure_node_with_score import PageFigureNodeWithScore
|
|
8
9
|
from .page_screenshot_node_with_score import PageScreenshotNodeWithScore
|
|
9
10
|
from .text_node_with_score import TextNodeWithScore
|
|
10
11
|
|
|
@@ -27,7 +28,10 @@ class RetrieveResults(pydantic.BaseModel):
|
|
|
27
28
|
description="The nodes retrieved by the pipeline for the given query."
|
|
28
29
|
)
|
|
29
30
|
image_nodes: typing.Optional[typing.List[PageScreenshotNodeWithScore]] = pydantic.Field(
|
|
30
|
-
description="The image nodes retrieved by the pipeline for the given query."
|
|
31
|
+
description="The image nodes retrieved by the pipeline for the given query. Deprecated - will soon be replaced with 'page_screenshot_nodes'."
|
|
32
|
+
)
|
|
33
|
+
page_figure_nodes: typing.Optional[typing.List[PageFigureNodeWithScore]] = pydantic.Field(
|
|
34
|
+
description="The page figure nodes retrieved by the pipeline for the given query."
|
|
31
35
|
)
|
|
32
36
|
retrieval_latency: typing.Optional[typing.Dict[str, float]] = pydantic.Field(
|
|
33
37
|
description="The end-to-end latency for retrieval and reranking."
|
|
@@ -9,9 +9,11 @@ T_Result = typing.TypeVar("T_Result")
|
|
|
9
9
|
class SupportedLlmModelNames(str, enum.Enum):
|
|
10
10
|
GPT_4_O = "GPT_4O"
|
|
11
11
|
GPT_4_O_MINI = "GPT_4O_MINI"
|
|
12
|
+
GPT_4_1 = "GPT_4_1"
|
|
13
|
+
GPT_4_1_NANO = "GPT_4_1_NANO"
|
|
14
|
+
GPT_4_1_MINI = "GPT_4_1_MINI"
|
|
12
15
|
AZURE_OPENAI_GPT_4_O = "AZURE_OPENAI_GPT_4O"
|
|
13
16
|
AZURE_OPENAI_GPT_4_O_MINI = "AZURE_OPENAI_GPT_4O_MINI"
|
|
14
|
-
AZURE_OPENAI_GPT_4 = "AZURE_OPENAI_GPT_4"
|
|
15
17
|
CLAUDE_3_5_SONNET = "CLAUDE_3_5_SONNET"
|
|
16
18
|
BEDROCK_CLAUDE_3_5_SONNET_V_1 = "BEDROCK_CLAUDE_3_5_SONNET_V1"
|
|
17
19
|
BEDROCK_CLAUDE_3_5_SONNET_V_2 = "BEDROCK_CLAUDE_3_5_SONNET_V2"
|
|
@@ -21,9 +23,11 @@ class SupportedLlmModelNames(str, enum.Enum):
|
|
|
21
23
|
self,
|
|
22
24
|
gpt_4_o: typing.Callable[[], T_Result],
|
|
23
25
|
gpt_4_o_mini: typing.Callable[[], T_Result],
|
|
26
|
+
gpt_4_1: typing.Callable[[], T_Result],
|
|
27
|
+
gpt_4_1_nano: typing.Callable[[], T_Result],
|
|
28
|
+
gpt_4_1_mini: typing.Callable[[], T_Result],
|
|
24
29
|
azure_openai_gpt_4_o: typing.Callable[[], T_Result],
|
|
25
30
|
azure_openai_gpt_4_o_mini: typing.Callable[[], T_Result],
|
|
26
|
-
azure_openai_gpt_4: typing.Callable[[], T_Result],
|
|
27
31
|
claude_3_5_sonnet: typing.Callable[[], T_Result],
|
|
28
32
|
bedrock_claude_3_5_sonnet_v_1: typing.Callable[[], T_Result],
|
|
29
33
|
bedrock_claude_3_5_sonnet_v_2: typing.Callable[[], T_Result],
|
|
@@ -33,12 +37,16 @@ class SupportedLlmModelNames(str, enum.Enum):
|
|
|
33
37
|
return gpt_4_o()
|
|
34
38
|
if self is SupportedLlmModelNames.GPT_4_O_MINI:
|
|
35
39
|
return gpt_4_o_mini()
|
|
40
|
+
if self is SupportedLlmModelNames.GPT_4_1:
|
|
41
|
+
return gpt_4_1()
|
|
42
|
+
if self is SupportedLlmModelNames.GPT_4_1_NANO:
|
|
43
|
+
return gpt_4_1_nano()
|
|
44
|
+
if self is SupportedLlmModelNames.GPT_4_1_MINI:
|
|
45
|
+
return gpt_4_1_mini()
|
|
36
46
|
if self is SupportedLlmModelNames.AZURE_OPENAI_GPT_4_O:
|
|
37
47
|
return azure_openai_gpt_4_o()
|
|
38
48
|
if self is SupportedLlmModelNames.AZURE_OPENAI_GPT_4_O_MINI:
|
|
39
49
|
return azure_openai_gpt_4_o_mini()
|
|
40
|
-
if self is SupportedLlmModelNames.AZURE_OPENAI_GPT_4:
|
|
41
|
-
return azure_openai_gpt_4()
|
|
42
50
|
if self is SupportedLlmModelNames.CLAUDE_3_5_SONNET:
|
|
43
51
|
return claude_3_5_sonnet()
|
|
44
52
|
if self is SupportedLlmModelNames.BEDROCK_CLAUDE_3_5_SONNET_V_1:
|
|
@@ -21,6 +21,7 @@ class UserOrganizationDelete(pydantic.BaseModel):
|
|
|
21
21
|
|
|
22
22
|
user_id: typing.Optional[str]
|
|
23
23
|
email: typing.Optional[str]
|
|
24
|
+
project_id_list: typing.Optional[typing.List[str]]
|
|
24
25
|
|
|
25
26
|
def json(self, **kwargs: typing.Any) -> str:
|
|
26
27
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
llama_cloud/__init__.py,sha256=
|
|
2
|
-
llama_cloud/client.py,sha256=
|
|
1
|
+
llama_cloud/__init__.py,sha256=8-yO1moSZzcHW8aAcqSu8z8Un9oMBnlV703ExtGQtYw,24279
|
|
2
|
+
llama_cloud/client.py,sha256=NuwtJJei6Wsa25GbYpxQnd_ZcD9_FxDJZp3Tu5OL914,5440
|
|
3
3
|
llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
|
|
4
4
|
llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
|
5
5
|
llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
|
|
@@ -9,9 +9,7 @@ llama_cloud/core/remove_none_from_dict.py,sha256=8m91FC3YuVem0Gm9_sXhJ2tGvP33owJ
|
|
|
9
9
|
llama_cloud/environment.py,sha256=feTjOebeFZMrBdnHat4RE5aHlpt-sJm4NhK4ntV1htI,167
|
|
10
10
|
llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6g,170
|
|
11
11
|
llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
|
|
12
|
-
llama_cloud/resources/__init__.py,sha256=
|
|
13
|
-
llama_cloud/resources/admin/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
14
|
-
llama_cloud/resources/admin/client.py,sha256=tIfM5KMJXRL0AUAm_s_fx5OzgqMUIxksjhCGuviQXQk,3080
|
|
12
|
+
llama_cloud/resources/__init__.py,sha256=tVfSvZW9QzKBpC1XjqTMjZsyaMVHfetmwBpCPSjEGHk,4121
|
|
15
13
|
llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
16
14
|
llama_cloud/resources/beta/client.py,sha256=eRB3mGmNxbhVGTtUpp-j-2APkHUoCbUckIz9coYjCsM,14666
|
|
17
15
|
llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
@@ -32,15 +30,15 @@ llama_cloud/resources/embedding_model_configs/types/embedding_model_config_creat
|
|
|
32
30
|
llama_cloud/resources/evals/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
33
31
|
llama_cloud/resources/evals/client.py,sha256=v2AyeQV0hVgC6xoP2gJNgneJMaeXALV1hIeirYGxlPw,3242
|
|
34
32
|
llama_cloud/resources/files/__init__.py,sha256=3B0SNM8EE6PddD5LpxYllci9vflEXy1xjPzhEEd-OUk,293
|
|
35
|
-
llama_cloud/resources/files/client.py,sha256=
|
|
33
|
+
llama_cloud/resources/files/client.py,sha256=gEk7nhX2sZE3PYFCfvdeYtJfTK_vMlWLHtjsl73p73c,50646
|
|
36
34
|
llama_cloud/resources/files/types/__init__.py,sha256=EPYENAwkjBWv1MLf8s7R5-RO-cxZ_8NPrqfR4ZoR7jY,418
|
|
37
35
|
llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py,sha256=Wc8wFgujOO5pZvbbh2TMMzpa37GKZd14GYNJ9bdq7BE,214
|
|
38
36
|
llama_cloud/resources/files/types/file_create_permission_info_value.py,sha256=KPCFuEaa8NiB85A5MfdXRAQ0poAUTl7Feg6BTfmdWas,209
|
|
39
37
|
llama_cloud/resources/files/types/file_create_resource_info_value.py,sha256=R7Y-CJf7fnbvIqE3xOI5XOrmPwLbVJLC7zpxMu8Zopk,201
|
|
40
38
|
llama_cloud/resources/jobs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
41
|
-
llama_cloud/resources/jobs/client.py,sha256=
|
|
39
|
+
llama_cloud/resources/jobs/client.py,sha256=gv_N8e0lay7cjt6MCwx-Cj4FiCXKhbyCDaWbadaJpgY,6270
|
|
42
40
|
llama_cloud/resources/llama_extract/__init__.py,sha256=jRUugj6XARMpKZi3e2RkfTdcOSuE-Zy0IfScRLlyYMs,819
|
|
43
|
-
llama_cloud/resources/llama_extract/client.py,sha256=
|
|
41
|
+
llama_cloud/resources/llama_extract/client.py,sha256=XU9FyO4Q6R61VOYDRjwddhJWhONZncJ5R9-JE6iEHZY,66605
|
|
44
42
|
llama_cloud/resources/llama_extract/types/__init__.py,sha256=ZRBD-jg1qdXyiJKTxgH7zaadoDzuof1TYpjK4P5z4zA,1216
|
|
45
43
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py,sha256=zB31hJQ8hKaIsPkfTWiX5hqsPVFMyyeWEDZ_Aq237jo,305
|
|
46
44
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py,sha256=xoyXH3f0Y5beMWBxmtXSz6QoB_df_-0QBsYdjBhZnGw,217
|
|
@@ -51,11 +49,11 @@ llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_o
|
|
|
51
49
|
llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema.py,sha256=uMqpKJdCmUNtryS2bkQTNA1AgDlWdtsBOP31iMt3zNA,346
|
|
52
50
|
llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema_zero_value.py,sha256=cUS7ez5r0Vx8T7SxwLYptZMmvpT5JoDVMyn54Q6VL-g,227
|
|
53
51
|
llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
54
|
-
llama_cloud/resources/organizations/client.py,sha256=
|
|
52
|
+
llama_cloud/resources/organizations/client.py,sha256=CdrdNdB9R-bOsNqZ4Jbm1BzG1RafXMFjuCsrVYf2OrE,56567
|
|
55
53
|
llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
56
|
-
llama_cloud/resources/parsing/client.py,sha256=
|
|
54
|
+
llama_cloud/resources/parsing/client.py,sha256=5xa47aiBNAjZzvKJNfmu5vM0ODuDzpWt-6SPOYTwhSg,78491
|
|
57
55
|
llama_cloud/resources/pipelines/__init__.py,sha256=zyvVEOF_krvEZkCIj_kZoMKfhDqHo_R32a1mv9CriQc,1193
|
|
58
|
-
llama_cloud/resources/pipelines/client.py,sha256=
|
|
56
|
+
llama_cloud/resources/pipelines/client.py,sha256=BcBqzTPu1LUsdimXvuaaKjUu6w5xjbL-ZBfWsO183Vk,132360
|
|
59
57
|
llama_cloud/resources/pipelines/types/__init__.py,sha256=C68NQ5QzA0dFXf9oePFFGmV1vn96jcAp-QAznSgoRYQ,1375
|
|
60
58
|
llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
|
|
61
59
|
llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
|
|
@@ -69,7 +67,7 @@ llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwH
|
|
|
69
67
|
llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
|
|
70
68
|
llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
71
69
|
llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
|
|
72
|
-
llama_cloud/types/__init__.py,sha256
|
|
70
|
+
llama_cloud/types/__init__.py,sha256=q70Gh26R9ZvPLFZakF5ajLT4uXj6uoAw29ej2P7HkGM,28642
|
|
73
71
|
llama_cloud/types/advanced_mode_transform_config.py,sha256=4xCXye0_cPmVS1F8aNTx81sIaEPjQH9kiCCAIoqUzlI,1502
|
|
74
72
|
llama_cloud/types/advanced_mode_transform_config_chunking_config.py,sha256=wYbJnWLpeQDfhmDZz-wJfYzD1iGT5Jcxb9ga3mzUuvk,1983
|
|
75
73
|
llama_cloud/types/advanced_mode_transform_config_segmentation_config.py,sha256=anNGq0F5-IlbIW3kpC8OilzLJnUq5tdIcWHnRnmlYsg,1303
|
|
@@ -116,7 +114,7 @@ llama_cloud/types/cloud_slack_data_source.py,sha256=tlsNj-hDj1gWmM0Q2A1BeyolfaPg
|
|
|
116
114
|
llama_cloud/types/cohere_embedding.py,sha256=wkv_fVCA1WEroGawzPFExwmiJ75gPfzeeemty7NBlsM,1579
|
|
117
115
|
llama_cloud/types/cohere_embedding_config.py,sha256=c0Kj1wuSsBX9TQ2AondKv5ZtX5PmkivsHj6P0M7tVB4,1142
|
|
118
116
|
llama_cloud/types/composite_retrieval_mode.py,sha256=PtN0vQ90xyAJL4vyGRG4lMNOpnJ__2L1xiwosI9yfms,548
|
|
119
|
-
llama_cloud/types/composite_retrieval_result.py,sha256=
|
|
117
|
+
llama_cloud/types/composite_retrieval_result.py,sha256=EulVseVvpK50kto4wQweLO7jJe6l6Ym1erKa4dOl4CU,1801
|
|
120
118
|
llama_cloud/types/composite_retrieved_text_node.py,sha256=eTQ99cdZ2PASff5n4oVV1oaNiS9Ie3AtY_E55kBYpBs,1702
|
|
121
119
|
llama_cloud/types/composite_retrieved_text_node_with_score.py,sha256=o-HvmyjqODc68zYuobtj10_62FMBAKRLfRoTHGDdmxw,1148
|
|
122
120
|
llama_cloud/types/configurable_data_sink_names.py,sha256=0Yk9i8hcNXKCcSKpa5KwsCwy_EDeodqbny7qmF86_lM,1225
|
|
@@ -147,13 +145,13 @@ llama_cloud/types/embedding_model_config_update_embedding_config.py,sha256=mrXFx
|
|
|
147
145
|
llama_cloud/types/eval_execution_params.py,sha256=ntVaJh5SMZMPL4QLUiihVjUlg2SKbrezvbMKGlrF66Q,1369
|
|
148
146
|
llama_cloud/types/extract_agent.py,sha256=T98IOueut4M52Qm7hqcUOcWFFDhZ-ye0OFdXgfFGtS4,1763
|
|
149
147
|
llama_cloud/types/extract_agent_data_schema_value.py,sha256=UaDQ2KjajLDccW7F4NKdfpefeTJrr1hl0c95WRETYkM,201
|
|
150
|
-
llama_cloud/types/extract_config.py,sha256=
|
|
148
|
+
llama_cloud/types/extract_config.py,sha256=j3792FdRUPpS6U3ynIDIUBaOTVSZfoNV3aBTnma_IjA,2082
|
|
151
149
|
llama_cloud/types/extract_job.py,sha256=Yx4fDdCdylAji2LPTwqflVpz1o9slpj9tTLS93-1tzU,1431
|
|
152
150
|
llama_cloud/types/extract_job_create.py,sha256=UK1mBIKyflo7e6m1MxMN95pLscj67jH_yvs8EvmBXqU,1545
|
|
153
151
|
llama_cloud/types/extract_job_create_data_schema_override.py,sha256=vuiJ2lGJjbXEnvFKzVnKyvgwhMXPg1Pb5GZne2DrB60,330
|
|
154
152
|
llama_cloud/types/extract_job_create_data_schema_override_zero_value.py,sha256=HHEYxOSQXXyBYOiUQg_qwfQtXFj-OtThMwbUDBIgZU0,223
|
|
155
153
|
llama_cloud/types/extract_mode.py,sha256=DwTMzDq3HHJop_fxQelHEE_k8UcdDz-W_v_Oj2WWXLk,931
|
|
156
|
-
llama_cloud/types/extract_models.py,sha256=
|
|
154
|
+
llama_cloud/types/extract_models.py,sha256=GrNd7MPOfB7G4zzAtfTPfLw2EKJFoY_NJ9hbd-wZQfU,1519
|
|
157
155
|
llama_cloud/types/extract_resultset.py,sha256=Alje0YQJUiA_aKi0hQs7TAnhDmZuQ_yL9b6HCNYBFQg,1627
|
|
158
156
|
llama_cloud/types/extract_resultset_data.py,sha256=v9Ae4SxLsvYPE9crko4N16lBjsxuZpz1yrUOhnaM_VY,427
|
|
159
157
|
llama_cloud/types/extract_resultset_data_item_value.py,sha256=JwqgDIGW0irr8QWaSTIrl24FhGxTUDOXIbxoSdIjuxs,209
|
|
@@ -172,6 +170,7 @@ llama_cloud/types/extract_target.py,sha256=Gt-FNqblzcjdfq1hxsqEjWWu-HNLXdKy4w98n
|
|
|
172
170
|
llama_cloud/types/fail_page_mode.py,sha256=n4fgPpiEB5siPoEg0Sux4COg7ElNybjshxDoUihZwRU,786
|
|
173
171
|
llama_cloud/types/file.py,sha256=rQXitPRKOYw91nK5qOZ0vpOmIx_MCpRb0g78d9dQs6w,1822
|
|
174
172
|
llama_cloud/types/file_count_by_status_response.py,sha256=WuorbZvKjDs9Ql1hUiQu4gN5iCm8d6fr92KLyHpRvQU,1356
|
|
173
|
+
llama_cloud/types/file_id_presigned_url.py,sha256=Yr_MGFKbuBEHK4efRSK53fHcoo5bbAKnqQGGhMycUc0,1398
|
|
175
174
|
llama_cloud/types/file_parse_public.py,sha256=sshZ0BcjHMGpuz4ylSurv0K_3ejfPrUGGyDxBHCtdMg,1378
|
|
176
175
|
llama_cloud/types/file_permission_info_value.py,sha256=RyQlNbhvIKS87Ywu7XUaw5jDToZX64M9Wqzu1U_q2Us,197
|
|
177
176
|
llama_cloud/types/file_resource_info_value.py,sha256=g6T6ELeLK9jgcvX6r-EuAl_4JkwnyqdS0RRoabMReSU,195
|
|
@@ -193,15 +192,12 @@ llama_cloud/types/job_record.py,sha256=7hdDPZU11EG8g6_9iq6vy-zqLEryeC7i8fZ-CkUB_
|
|
|
193
192
|
llama_cloud/types/job_record_parameters.py,sha256=Oqxp5y0owPfjLc_NR7AYE8P3zM2PJo36N9olbyNl7AA,3425
|
|
194
193
|
llama_cloud/types/job_record_with_usage_metrics.py,sha256=iNV2do5TB_0e3PoOz_DJyAaM6Cn9G8KG-dGPGgEs5SY,1198
|
|
195
194
|
llama_cloud/types/l_lama_parse_transform_config.py,sha256=YQRJZvKh1Ee2FUyW_N0nqYJoW599qBgH3JCH9SH6YLo,1249
|
|
196
|
-
llama_cloud/types/legacy_parse_job_config.py,sha256=
|
|
197
|
-
llama_cloud/types/llama_extract_settings.py,sha256=
|
|
195
|
+
llama_cloud/types/legacy_parse_job_config.py,sha256=4cC-p4d5bwkKnn_VZBuf7pCpwGFaD0QyJd5cPicaryY,11801
|
|
196
|
+
llama_cloud/types/llama_extract_settings.py,sha256=bHtF5AD0r896-248e7WKthcbbvrAUdptZrENP2Ed4LM,2388
|
|
198
197
|
llama_cloud/types/llama_index_core_base_llms_types_chat_message.py,sha256=NelHo-T-ebVMhRKsqE_xV8AJW4c7o6lS0uEQnPsmTwg,1365
|
|
199
198
|
llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py,sha256=-aL8fh-w2Xf4uQs_LHzb3q6LL_onLAcVzCR5yMI4qJw,1571
|
|
200
|
-
llama_cloud/types/llama_parse_parameters.py,sha256=
|
|
199
|
+
llama_cloud/types/llama_parse_parameters.py,sha256=Ap-F-8giC0pa45blPxMLNHJ-3zEgJc1zrNM1a9NWErs,5826
|
|
201
200
|
llama_cloud/types/llama_parse_supported_file_extensions.py,sha256=B_0N3f8Aq59W9FbsH50mGBUiyWTIXQjHFl739uAyaQw,11207
|
|
202
|
-
llama_cloud/types/llm_config_result.py,sha256=3NRAvq_jVJlkbLRDgBHTGMDli0av7d0GJME3aiXytKs,1106
|
|
203
|
-
llama_cloud/types/llm_config_result_llm_type.py,sha256=yrijlC2f1egNDx-tCvDVp68pFmGaJZvVE_D1vS2wA34,1032
|
|
204
|
-
llama_cloud/types/llm_configs_response.py,sha256=TQ9RLkOzdY-8k0l1NNUZNUl83dmKks0gOQvhj-4hRZs,1073
|
|
205
201
|
llama_cloud/types/llm_model_data.py,sha256=6rrycqGwlK3LZ2S-WtgmeomithdLhDCgwBBZQ5KLaso,1300
|
|
206
202
|
llama_cloud/types/llm_parameters.py,sha256=RTKYt09lm9a1MlnBfYuTP2x_Ww4byUNNc1TqIel5O1Y,1377
|
|
207
203
|
llama_cloud/types/load_files_job_config.py,sha256=R5sFgFmV__0mqLUuD7dkFoBJHG2ZLw5px9zRapvYcpE,1069
|
|
@@ -221,7 +217,8 @@ llama_cloud/types/open_ai_embedding.py,sha256=RQijkvKyzbISy92LnBSEpjmIU8p7kMpdc4
|
|
|
221
217
|
llama_cloud/types/open_ai_embedding_config.py,sha256=Mquc0JrtCo8lVYA2WW7q0ZikS3HRkiMtzDFu5XA-20o,1143
|
|
222
218
|
llama_cloud/types/organization.py,sha256=p8mYRqSsGxw17AmdW8x8nP7P1UbdpYkwr51WTIjTVLw,1467
|
|
223
219
|
llama_cloud/types/organization_create.py,sha256=hUXRwArIx_0D_lilpL7z-B0oJJ5yEX8Sbu2xqfH_9so,1086
|
|
224
|
-
llama_cloud/types/page_figure_metadata.py,sha256=
|
|
220
|
+
llama_cloud/types/page_figure_metadata.py,sha256=0oasDkjnzoVQ4W-Ci0KoJHM0iHXTGvm3cbdVOgH9nHE,1588
|
|
221
|
+
llama_cloud/types/page_figure_node_with_score.py,sha256=VqNQx9RKmD_jY1kHPCvPjygshbfVLLSgtC5TX-Cy_cw,1208
|
|
225
222
|
llama_cloud/types/page_screenshot_metadata.py,sha256=lobrq0AsOr8sDwMgA9ytop8lRmRFvJW2oiql3yLvbjM,1328
|
|
226
223
|
llama_cloud/types/page_screenshot_node_with_score.py,sha256=EdqoXbmARCz1DV14E2saCPshIeII709uM4cLwxw_mkM,1232
|
|
227
224
|
llama_cloud/types/page_segmentation_config.py,sha256=VH8uuxnubnJak1gSpS64OoMueHidhsDB-2eq2tVHbag,998
|
|
@@ -230,7 +227,7 @@ llama_cloud/types/paginated_jobs_history_with_metrics.py,sha256=Bxy6N0x0FARJhgwN
|
|
|
230
227
|
llama_cloud/types/paginated_list_cloud_documents_response.py,sha256=MsjS0SWlT0syELDck4x2sxxR3_NC1e6QTdepgVmK9aY,1341
|
|
231
228
|
llama_cloud/types/paginated_list_pipeline_files_response.py,sha256=2TKR2oHSQRyLMqWz1qQBSIvz-ZJb8U_94367lwOJ2S4,1317
|
|
232
229
|
llama_cloud/types/paginated_report_response.py,sha256=o79QhQi9r0HZZrhvRlA6WGjxtyPuxN0xONhwXSwxtcs,1104
|
|
233
|
-
llama_cloud/types/parse_job_config.py,sha256=
|
|
230
|
+
llama_cloud/types/parse_job_config.py,sha256=ajTNaLdfsuaimwL2PobOrTKvIe75SbSdIrDzJkFqp1U,6307
|
|
234
231
|
llama_cloud/types/parse_plan_level.py,sha256=GBkDS19qfHseBa17EXfuTPNT4GNv5alyPrWEvWji3GY,528
|
|
235
232
|
llama_cloud/types/parser_languages.py,sha256=Ps3IlaSt6tyxEI657N3-vZL96r2puk8wsf31cWnO-SI,10840
|
|
236
233
|
llama_cloud/types/parsing_history_item.py,sha256=_MVzf43t84PbmjOzsMLZ_NBoyiisigLWz-fr0ZxU63g,1183
|
|
@@ -276,7 +273,7 @@ llama_cloud/types/plan_limits.py,sha256=WAbDbRl8gsQxvhmuVB0YT8mry-0uKg6c66uivypp
|
|
|
276
273
|
llama_cloud/types/playground_session.py,sha256=F8u2KZL2YaOrsT-o1n4zbhyPxSsoduc3ZCzQB8AecFA,1858
|
|
277
274
|
llama_cloud/types/pooling.py,sha256=5Fr6c8rx9SDWwWzEvD78suob2d79ktodUtLUAUHMbP8,651
|
|
278
275
|
llama_cloud/types/preset_composite_retrieval_params.py,sha256=yEf1pk4Wz5J6SxgB8elklwuyVDCRSZqfWC6x3hJUS4Q,1366
|
|
279
|
-
llama_cloud/types/preset_retrieval_params.py,sha256=
|
|
276
|
+
llama_cloud/types/preset_retrieval_params.py,sha256=TcyljefpspJSveMR9L5DQHlqW4jZeexBsXus_LkHkJA,2365
|
|
280
277
|
llama_cloud/types/preset_retrieval_params_search_filters_inference_schema_value.py,sha256=BOp-oJMIc3KVU89mmKIhVcwwsO0XBRnuErfsPqpUjSs,234
|
|
281
278
|
llama_cloud/types/presigned_url.py,sha256=-DOQo7XKvUsl-9Gz7fX6VOHdQLzGH2XRau24ASvG92E,1275
|
|
282
279
|
llama_cloud/types/progress_event.py,sha256=Bk73A8geTVaq0ze5pMnbkAmx7FSOHQIixYCpCas_dcY,1684
|
|
@@ -305,7 +302,7 @@ llama_cloud/types/report_state.py,sha256=gjexexoT8GaCamGKvfwivKrfRtvdhEtwSLkAt-j
|
|
|
305
302
|
llama_cloud/types/report_state_event.py,sha256=_wf-Cl_skJdrag-7h11tz-HIy1jed_GIG3c-ksuAjT4,1270
|
|
306
303
|
llama_cloud/types/report_update_event.py,sha256=uLRC79U3pvZ5-kY6pOseQyX1MNH-0m80GUtzpjd6mkI,1270
|
|
307
304
|
llama_cloud/types/retrieval_mode.py,sha256=wV9q3OdHTuyDWbJCGdxq9Hw6U95WFlJcaMq6KWSTzyw,910
|
|
308
|
-
llama_cloud/types/retrieve_results.py,sha256=
|
|
305
|
+
llama_cloud/types/retrieve_results.py,sha256=rHArmu05K3NvIQepHX5nsVOfcMsZj3MaIcPkTC6mD_8,2375
|
|
309
306
|
llama_cloud/types/retriever.py,sha256=ZItPsorL8x1XjtJT49ZodaMqU8h2GfwlB4U4cgnfZkM,1626
|
|
310
307
|
llama_cloud/types/retriever_create.py,sha256=WyUR9DRzu3Q9tzKEeXCdQuzCY6WKi9ADJkZea9rqvxU,1286
|
|
311
308
|
llama_cloud/types/retriever_pipeline.py,sha256=F1pZDxg8JdQXRHE6ciFezd7a-Wv5bHplPcGDED-J4b0,1330
|
|
@@ -317,7 +314,7 @@ llama_cloud/types/status_enum.py,sha256=cUBIlys89E8PUzmVqqawu7qTDF0aRqBwiijOmRDP
|
|
|
317
314
|
llama_cloud/types/struct_mode.py,sha256=ROicwjXfFmgVU8_xSVxJlnFUzRNKG5VIEF1wYg9uOPU,1020
|
|
318
315
|
llama_cloud/types/struct_parse_conf.py,sha256=WlL8y0IBvdzGsDtFUlEZLzoUODwmOWAJi0viS9unL18,2297
|
|
319
316
|
llama_cloud/types/supported_llm_model.py,sha256=hubSopFICVNEegbJbtbpK6zRHwFPwUNtrw_NAw_3bfg,1380
|
|
320
|
-
llama_cloud/types/supported_llm_model_names.py,sha256=
|
|
317
|
+
llama_cloud/types/supported_llm_model_names.py,sha256=PXL0gA1lc0GJNzZHnjOscoxHpPW787A8Adh-2egAKo8,2512
|
|
321
318
|
llama_cloud/types/text_block.py,sha256=X154sQkSyposXuRcEWNp_tWcDQ-AI6q_-MfJUN5exP8,958
|
|
322
319
|
llama_cloud/types/text_node.py,sha256=Tq3QmuKC5cIHvC9wAtvhsXl1g2sACs2yJwQ0Uko8GSU,2846
|
|
323
320
|
llama_cloud/types/text_node_relationships_value.py,sha256=qmXURTk1Xg7ZDzRSSV1uDEel0AXRLohND5ioezibHY0,217
|
|
@@ -330,14 +327,14 @@ llama_cloud/types/usage_response_active_alerts_item.py,sha256=5EgU7go_CPe2Bmio12
|
|
|
330
327
|
llama_cloud/types/user_job_record.py,sha256=mJHdokJsemXJOwM2l7fsW3X0SlwSNcy7yHbcXZHh3I4,1098
|
|
331
328
|
llama_cloud/types/user_organization.py,sha256=yKewpOrMcB-CbujGNTjkX6QiWYr5HVsRIFQ-WX8kp2I,1729
|
|
332
329
|
llama_cloud/types/user_organization_create.py,sha256=Zj57s9xuYVnLW2p8i4j2QORL-G1y7Ab3avXE1baERQY,1189
|
|
333
|
-
llama_cloud/types/user_organization_delete.py,sha256=
|
|
330
|
+
llama_cloud/types/user_organization_delete.py,sha256=bEfgQMdTd6oAMZXtvSm5BhZahG1wAVDBXZ8e7V9UN7w,1159
|
|
334
331
|
llama_cloud/types/user_organization_role.py,sha256=vTM5pYG9NJpTQACn8vzSIt01Ul6jEHCVmyR3vV0isPg,1512
|
|
335
332
|
llama_cloud/types/validation_error.py,sha256=yZDLtjUHDY5w82Ra6CW0H9sLAr18R0RY1UNgJKR72DQ,1084
|
|
336
333
|
llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPXjdtN9EB7HrLVo6EP0,128
|
|
337
334
|
llama_cloud/types/vertex_ai_embedding_config.py,sha256=DvQk2xMJFmo54MEXTzoM4KSADyhGm_ygmFyx6wIcQdw,1159
|
|
338
335
|
llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2BaIMltDqGnIowU,1217
|
|
339
336
|
llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
|
|
340
|
-
llama_cloud-0.1.
|
|
341
|
-
llama_cloud-0.1.
|
|
342
|
-
llama_cloud-0.1.
|
|
343
|
-
llama_cloud-0.1.
|
|
337
|
+
llama_cloud-0.1.25.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
|
|
338
|
+
llama_cloud-0.1.25.dist-info/METADATA,sha256=EaOl_aKCf0kdbk6eSVh86xmUub8cGlQsk-XZEQIlEQ0,1143
|
|
339
|
+
llama_cloud-0.1.25.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
|
|
340
|
+
llama_cloud-0.1.25.dist-info/RECORD,,
|