llama-cloud 0.1.20__py3-none-any.whl → 0.1.22__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +18 -136
- llama_cloud/client.py +3 -3
- llama_cloud/resources/__init__.py +4 -43
- llama_cloud/resources/admin/client.py +78 -0
- llama_cloud/resources/data_sinks/__init__.py +2 -18
- llama_cloud/resources/data_sinks/client.py +94 -2
- llama_cloud/resources/data_sinks/types/__init__.py +2 -18
- llama_cloud/resources/data_sinks/types/data_sink_update_component.py +7 -65
- llama_cloud/resources/data_sources/__init__.py +2 -30
- llama_cloud/resources/data_sources/types/__init__.py +1 -28
- llama_cloud/resources/data_sources/types/data_source_update_component.py +23 -2
- llama_cloud/resources/jobs/client.py +10 -2
- llama_cloud/resources/llama_extract/client.py +50 -6
- llama_cloud/resources/organizations/client.py +12 -2
- llama_cloud/resources/parsing/client.py +30 -0
- llama_cloud/resources/pipelines/client.py +22 -0
- llama_cloud/resources/retrievers/client.py +14 -0
- llama_cloud/types/__init__.py +18 -112
- llama_cloud/types/cloud_jira_data_source.py +4 -0
- llama_cloud/types/cloud_s_3_data_source.py +1 -0
- llama_cloud/types/data_sink_component.py +7 -65
- llama_cloud/types/data_sink_create_component.py +7 -65
- llama_cloud/types/data_source_component.py +23 -2
- llama_cloud/types/data_source_create_component.py +23 -2
- llama_cloud/types/{data_sink_definition.py → document_block.py} +6 -15
- llama_cloud/types/document_chunk_mode.py +17 -0
- llama_cloud/types/extract_config.py +7 -0
- llama_cloud/types/extract_mode.py +4 -0
- llama_cloud/types/extract_models.py +33 -0
- llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +11 -0
- llama_cloud/types/{data_source_definition.py → llm_config_result.py} +6 -15
- llama_cloud/types/llm_config_result_llm_type.py +33 -0
- llama_cloud/types/llm_configs_response.py +33 -0
- llama_cloud/types/pipeline.py +2 -0
- llama_cloud/types/pipeline_create.py +1 -3
- llama_cloud/types/pipeline_data_source_component.py +23 -2
- llama_cloud/types/pipeline_status.py +17 -0
- llama_cloud/types/prompt_conf.py +1 -0
- llama_cloud/types/struct_parse_conf.py +2 -1
- llama_cloud/types/supported_llm_model.py +1 -2
- llama_cloud/types/supported_llm_model_names.py +4 -4
- llama_cloud/types/user_organization_role.py +1 -0
- {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/METADATA +5 -3
- {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/RECORD +47 -49
- {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/WHEEL +1 -1
- llama_cloud/resources/component_definitions/client.py +0 -189
- llama_cloud/resources/data_sources/types/data_source_update_component_one.py +0 -122
- llama_cloud/types/configurable_transformation_definition.py +0 -48
- llama_cloud/types/configurable_transformation_names.py +0 -41
- llama_cloud/types/data_source_component_one.py +0 -122
- llama_cloud/types/data_source_create_component_one.py +0 -122
- llama_cloud/types/pipeline_data_source_component_one.py +0 -122
- llama_cloud/types/transformation_category_names.py +0 -17
- /llama_cloud/resources/{component_definitions → admin}/__init__.py +0 -0
- {llama_cloud-0.1.20.dist-info → llama_cloud-0.1.22.dist-info}/LICENSE +0 -0
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class ExtractModels(str, enum.Enum):
|
|
10
|
+
GPT_4_O = "gpt-4o"
|
|
11
|
+
GPT_4_O_MINI = "gpt-4o-mini"
|
|
12
|
+
GPT_41 = "gpt-4.1"
|
|
13
|
+
GPT_41_MINI = "gpt-4.1-mini"
|
|
14
|
+
O_3_MINI = "o3-mini"
|
|
15
|
+
|
|
16
|
+
def visit(
|
|
17
|
+
self,
|
|
18
|
+
gpt_4_o: typing.Callable[[], T_Result],
|
|
19
|
+
gpt_4_o_mini: typing.Callable[[], T_Result],
|
|
20
|
+
gpt_41: typing.Callable[[], T_Result],
|
|
21
|
+
gpt_41_mini: typing.Callable[[], T_Result],
|
|
22
|
+
o_3_mini: typing.Callable[[], T_Result],
|
|
23
|
+
) -> T_Result:
|
|
24
|
+
if self is ExtractModels.GPT_4_O:
|
|
25
|
+
return gpt_4_o()
|
|
26
|
+
if self is ExtractModels.GPT_4_O_MINI:
|
|
27
|
+
return gpt_4_o_mini()
|
|
28
|
+
if self is ExtractModels.GPT_41:
|
|
29
|
+
return gpt_41()
|
|
30
|
+
if self is ExtractModels.GPT_41_MINI:
|
|
31
|
+
return gpt_41_mini()
|
|
32
|
+
if self is ExtractModels.O_3_MINI:
|
|
33
|
+
return o_3_mini()
|
|
@@ -7,6 +7,7 @@ import typing
|
|
|
7
7
|
import typing_extensions
|
|
8
8
|
|
|
9
9
|
from .audio_block import AudioBlock
|
|
10
|
+
from .document_block import DocumentBlock
|
|
10
11
|
from .image_block import ImageBlock
|
|
11
12
|
from .text_block import TextBlock
|
|
12
13
|
|
|
@@ -20,6 +21,15 @@ class LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Audio(AudioBlock):
|
|
|
20
21
|
allow_population_by_field_name = True
|
|
21
22
|
|
|
22
23
|
|
|
24
|
+
class LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Document(DocumentBlock):
|
|
25
|
+
block_type: typing_extensions.Literal["document"]
|
|
26
|
+
|
|
27
|
+
class Config:
|
|
28
|
+
frozen = True
|
|
29
|
+
smart_union = True
|
|
30
|
+
allow_population_by_field_name = True
|
|
31
|
+
|
|
32
|
+
|
|
23
33
|
class LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image(ImageBlock):
|
|
24
34
|
block_type: typing_extensions.Literal["image"]
|
|
25
35
|
|
|
@@ -40,6 +50,7 @@ class LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text(TextBlock):
|
|
|
40
50
|
|
|
41
51
|
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem = typing.Union[
|
|
42
52
|
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Audio,
|
|
53
|
+
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Document,
|
|
43
54
|
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Image,
|
|
44
55
|
LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem_Text,
|
|
45
56
|
]
|
|
@@ -4,7 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .
|
|
7
|
+
from .llm_config_result_llm_type import LlmConfigResultLlmType
|
|
8
8
|
|
|
9
9
|
try:
|
|
10
10
|
import pydantic
|
|
@@ -15,20 +15,11 @@ except ImportError:
|
|
|
15
15
|
import pydantic # type: ignore
|
|
16
16
|
|
|
17
17
|
|
|
18
|
-
class
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
label: str = pydantic.Field(
|
|
24
|
-
description="The label field will be used to display the name of the component in the UI"
|
|
25
|
-
)
|
|
26
|
-
json_schema: typing.Dict[str, typing.Any] = pydantic.Field(
|
|
27
|
-
description="The json_schema field can be used by clients to determine how to construct the component"
|
|
28
|
-
)
|
|
29
|
-
source_type: ConfigurableDataSourceNames = pydantic.Field(
|
|
30
|
-
description="The name field will act as the unique identifier of DataSourceDefinition objects"
|
|
31
|
-
)
|
|
18
|
+
class LlmConfigResult(pydantic.BaseModel):
|
|
19
|
+
llm_type: LlmConfigResultLlmType
|
|
20
|
+
is_enabled: bool
|
|
21
|
+
valid: bool
|
|
22
|
+
error_message: typing.Optional[str]
|
|
32
23
|
|
|
33
24
|
def json(self, **kwargs: typing.Any) -> str:
|
|
34
25
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class LlmConfigResultLlmType(str, enum.Enum):
|
|
10
|
+
OPENAI = "openai"
|
|
11
|
+
ANTHROPIC = "anthropic"
|
|
12
|
+
GEMINI = "gemini"
|
|
13
|
+
AWS_BEDROCK = "aws_bedrock"
|
|
14
|
+
AZURE_OPENAI = "azure_openai"
|
|
15
|
+
|
|
16
|
+
def visit(
|
|
17
|
+
self,
|
|
18
|
+
openai: typing.Callable[[], T_Result],
|
|
19
|
+
anthropic: typing.Callable[[], T_Result],
|
|
20
|
+
gemini: typing.Callable[[], T_Result],
|
|
21
|
+
aws_bedrock: typing.Callable[[], T_Result],
|
|
22
|
+
azure_openai: typing.Callable[[], T_Result],
|
|
23
|
+
) -> T_Result:
|
|
24
|
+
if self is LlmConfigResultLlmType.OPENAI:
|
|
25
|
+
return openai()
|
|
26
|
+
if self is LlmConfigResultLlmType.ANTHROPIC:
|
|
27
|
+
return anthropic()
|
|
28
|
+
if self is LlmConfigResultLlmType.GEMINI:
|
|
29
|
+
return gemini()
|
|
30
|
+
if self is LlmConfigResultLlmType.AWS_BEDROCK:
|
|
31
|
+
return aws_bedrock()
|
|
32
|
+
if self is LlmConfigResultLlmType.AZURE_OPENAI:
|
|
33
|
+
return azure_openai()
|
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import datetime as dt
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
|
7
|
+
from .llm_config_result import LlmConfigResult
|
|
8
|
+
|
|
9
|
+
try:
|
|
10
|
+
import pydantic
|
|
11
|
+
if pydantic.__version__.startswith("1."):
|
|
12
|
+
raise ImportError
|
|
13
|
+
import pydantic.v1 as pydantic # type: ignore
|
|
14
|
+
except ImportError:
|
|
15
|
+
import pydantic # type: ignore
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
class LlmConfigsResponse(pydantic.BaseModel):
|
|
19
|
+
llm_configs: typing.List[LlmConfigResult]
|
|
20
|
+
last_validated_at: str
|
|
21
|
+
|
|
22
|
+
def json(self, **kwargs: typing.Any) -> str:
|
|
23
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
24
|
+
return super().json(**kwargs_with_defaults)
|
|
25
|
+
|
|
26
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
|
27
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
28
|
+
return super().dict(**kwargs_with_defaults)
|
|
29
|
+
|
|
30
|
+
class Config:
|
|
31
|
+
frozen = True
|
|
32
|
+
smart_union = True
|
|
33
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
llama_cloud/types/pipeline.py
CHANGED
|
@@ -10,6 +10,7 @@ from .llama_parse_parameters import LlamaParseParameters
|
|
|
10
10
|
from .pipeline_configuration_hashes import PipelineConfigurationHashes
|
|
11
11
|
from .pipeline_embedding_config import PipelineEmbeddingConfig
|
|
12
12
|
from .pipeline_metadata_config import PipelineMetadataConfig
|
|
13
|
+
from .pipeline_status import PipelineStatus
|
|
13
14
|
from .pipeline_transform_config import PipelineTransformConfig
|
|
14
15
|
from .pipeline_type import PipelineType
|
|
15
16
|
from .preset_retrieval_params import PresetRetrievalParams
|
|
@@ -51,6 +52,7 @@ class Pipeline(pydantic.BaseModel):
|
|
|
51
52
|
)
|
|
52
53
|
llama_parse_parameters: typing.Optional[LlamaParseParameters]
|
|
53
54
|
data_sink: typing.Optional[DataSink]
|
|
55
|
+
status: typing.Optional[PipelineStatus]
|
|
54
56
|
metadata_config: typing.Optional[PipelineMetadataConfig]
|
|
55
57
|
|
|
56
58
|
def json(self, **kwargs: typing.Any) -> str:
|
|
@@ -40,9 +40,7 @@ class PipelineCreate(pydantic.BaseModel):
|
|
|
40
40
|
eval_parameters: typing.Optional[EvalExecutionParams] = pydantic.Field(
|
|
41
41
|
description="Eval parameters for the pipeline."
|
|
42
42
|
)
|
|
43
|
-
llama_parse_parameters: typing.Optional[LlamaParseParameters]
|
|
44
|
-
description="Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline."
|
|
45
|
-
)
|
|
43
|
+
llama_parse_parameters: typing.Optional[LlamaParseParameters]
|
|
46
44
|
status: typing.Optional[str]
|
|
47
45
|
metadata_config: typing.Optional[PipelineMetadataConfig]
|
|
48
46
|
name: str
|
|
@@ -2,6 +2,27 @@
|
|
|
2
2
|
|
|
3
3
|
import typing
|
|
4
4
|
|
|
5
|
-
from .
|
|
5
|
+
from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
|
|
6
|
+
from .cloud_box_data_source import CloudBoxDataSource
|
|
7
|
+
from .cloud_confluence_data_source import CloudConfluenceDataSource
|
|
8
|
+
from .cloud_google_drive_data_source import CloudGoogleDriveDataSource
|
|
9
|
+
from .cloud_jira_data_source import CloudJiraDataSource
|
|
10
|
+
from .cloud_notion_page_data_source import CloudNotionPageDataSource
|
|
11
|
+
from .cloud_one_drive_data_source import CloudOneDriveDataSource
|
|
12
|
+
from .cloud_s_3_data_source import CloudS3DataSource
|
|
13
|
+
from .cloud_sharepoint_data_source import CloudSharepointDataSource
|
|
14
|
+
from .cloud_slack_data_source import CloudSlackDataSource
|
|
6
15
|
|
|
7
|
-
PipelineDataSourceComponent = typing.Union[
|
|
16
|
+
PipelineDataSourceComponent = typing.Union[
|
|
17
|
+
typing.Dict[str, typing.Any],
|
|
18
|
+
CloudS3DataSource,
|
|
19
|
+
CloudAzStorageBlobDataSource,
|
|
20
|
+
CloudGoogleDriveDataSource,
|
|
21
|
+
CloudOneDriveDataSource,
|
|
22
|
+
CloudSharepointDataSource,
|
|
23
|
+
CloudSlackDataSource,
|
|
24
|
+
CloudNotionPageDataSource,
|
|
25
|
+
CloudConfluenceDataSource,
|
|
26
|
+
CloudJiraDataSource,
|
|
27
|
+
CloudBoxDataSource,
|
|
28
|
+
]
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class PipelineStatus(str, enum.Enum):
|
|
10
|
+
CREATED = "CREATED"
|
|
11
|
+
DELETING = "DELETING"
|
|
12
|
+
|
|
13
|
+
def visit(self, created: typing.Callable[[], T_Result], deleting: typing.Callable[[], T_Result]) -> T_Result:
|
|
14
|
+
if self is PipelineStatus.CREATED:
|
|
15
|
+
return created()
|
|
16
|
+
if self is PipelineStatus.DELETING:
|
|
17
|
+
return deleting()
|
llama_cloud/types/prompt_conf.py
CHANGED
|
@@ -22,6 +22,7 @@ class PromptConf(pydantic.BaseModel):
|
|
|
22
22
|
cite_sources_prompt: typing.Optional[typing.Dict[str, str]] = pydantic.Field(
|
|
23
23
|
description="The prompt to use for citing sources."
|
|
24
24
|
)
|
|
25
|
+
scratchpad_prompt: typing.Optional[str] = pydantic.Field(description="The prompt to use for scratchpad.")
|
|
25
26
|
|
|
26
27
|
def json(self, **kwargs: typing.Any) -> str:
|
|
27
28
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -4,6 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
+
from .extract_models import ExtractModels
|
|
7
8
|
from .prompt_conf import PromptConf
|
|
8
9
|
from .schema_relax_mode import SchemaRelaxMode
|
|
9
10
|
from .struct_mode import StructMode
|
|
@@ -22,7 +23,7 @@ class StructParseConf(pydantic.BaseModel):
|
|
|
22
23
|
Configuration for the structured parsing agent.
|
|
23
24
|
"""
|
|
24
25
|
|
|
25
|
-
model: typing.Optional[
|
|
26
|
+
model: typing.Optional[ExtractModels] = pydantic.Field(description="The model to use for the structured parsing.")
|
|
26
27
|
temperature: typing.Optional[float] = pydantic.Field(
|
|
27
28
|
description="The temperature to use for the structured parsing."
|
|
28
29
|
)
|
|
@@ -5,7 +5,6 @@ import typing
|
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .llm_model_data import LlmModelData
|
|
8
|
-
from .supported_llm_model_names import SupportedLlmModelNames
|
|
9
8
|
|
|
10
9
|
try:
|
|
11
10
|
import pydantic
|
|
@@ -21,7 +20,7 @@ class SupportedLlmModel(pydantic.BaseModel):
|
|
|
21
20
|
Response Schema for a supported eval LLM model.
|
|
22
21
|
"""
|
|
23
22
|
|
|
24
|
-
name:
|
|
23
|
+
name: str = pydantic.Field(description="The name of the supported LLM model.")
|
|
25
24
|
enabled: typing.Optional[bool] = pydantic.Field(
|
|
26
25
|
description="Whether the LLM model is enabled for use in LlamaCloud."
|
|
27
26
|
)
|
|
@@ -15,7 +15,7 @@ class SupportedLlmModelNames(str, enum.Enum):
|
|
|
15
15
|
AZURE_OPENAI_GPT_4 = "AZURE_OPENAI_GPT_4"
|
|
16
16
|
CLAUDE_3_5_SONNET = "CLAUDE_3_5_SONNET"
|
|
17
17
|
BEDROCK_CLAUDE_3_5_SONNET = "BEDROCK_CLAUDE_3_5_SONNET"
|
|
18
|
-
|
|
18
|
+
VERTEX_AI_CLAUDE_3_5_SONNET_V_2 = "VERTEX_AI_CLAUDE_3_5_SONNET_V2"
|
|
19
19
|
|
|
20
20
|
def visit(
|
|
21
21
|
self,
|
|
@@ -27,7 +27,7 @@ class SupportedLlmModelNames(str, enum.Enum):
|
|
|
27
27
|
azure_openai_gpt_4: typing.Callable[[], T_Result],
|
|
28
28
|
claude_3_5_sonnet: typing.Callable[[], T_Result],
|
|
29
29
|
bedrock_claude_3_5_sonnet: typing.Callable[[], T_Result],
|
|
30
|
-
|
|
30
|
+
vertex_ai_claude_3_5_sonnet_v_2: typing.Callable[[], T_Result],
|
|
31
31
|
) -> T_Result:
|
|
32
32
|
if self is SupportedLlmModelNames.GPT_4_O:
|
|
33
33
|
return gpt_4_o()
|
|
@@ -45,5 +45,5 @@ class SupportedLlmModelNames(str, enum.Enum):
|
|
|
45
45
|
return claude_3_5_sonnet()
|
|
46
46
|
if self is SupportedLlmModelNames.BEDROCK_CLAUDE_3_5_SONNET:
|
|
47
47
|
return bedrock_claude_3_5_sonnet()
|
|
48
|
-
if self is SupportedLlmModelNames.
|
|
49
|
-
return
|
|
48
|
+
if self is SupportedLlmModelNames.VERTEX_AI_CLAUDE_3_5_SONNET_V_2:
|
|
49
|
+
return vertex_ai_claude_3_5_sonnet_v_2()
|
|
@@ -25,6 +25,7 @@ class UserOrganizationRole(pydantic.BaseModel):
|
|
|
25
25
|
updated_at: typing.Optional[dt.datetime]
|
|
26
26
|
user_id: str = pydantic.Field(description="The user's ID.")
|
|
27
27
|
organization_id: str = pydantic.Field(description="The organization's ID.")
|
|
28
|
+
project_ids: typing.Optional[typing.List[str]]
|
|
28
29
|
role_id: str = pydantic.Field(description="The role's ID.")
|
|
29
30
|
role: Role = pydantic.Field(description="The role.")
|
|
30
31
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
|
-
Metadata-Version: 2.
|
|
1
|
+
Metadata-Version: 2.3
|
|
2
2
|
Name: llama-cloud
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.22
|
|
4
4
|
Summary:
|
|
5
5
|
License: MIT
|
|
6
6
|
Author: Logan Markewich
|
|
@@ -13,6 +13,7 @@ Classifier: Programming Language :: Python :: 3.9
|
|
|
13
13
|
Classifier: Programming Language :: Python :: 3.10
|
|
14
14
|
Classifier: Programming Language :: Python :: 3.11
|
|
15
15
|
Classifier: Programming Language :: Python :: 3.12
|
|
16
|
+
Classifier: Programming Language :: Python :: 3.13
|
|
16
17
|
Requires-Dist: certifi (>=2024.7.4)
|
|
17
18
|
Requires-Dist: httpx (>=0.20.0)
|
|
18
19
|
Requires-Dist: pydantic (>=1.10)
|
|
@@ -26,6 +27,7 @@ To publish:
|
|
|
26
27
|
- update the version in `pyproject.toml`
|
|
27
28
|
- run `poetry publish --build`
|
|
28
29
|
|
|
29
|
-
|
|
30
|
+
Setup credentials:
|
|
30
31
|
- run `poetry config pypi-token.pypi <my-token>`
|
|
32
|
+
- Get token form PyPi once logged in with credentials in [1Password](https://start.1password.com/open/i?a=32SA66TZ3JCRXOCMASLSDCT5TI&v=lhv7hvb5o46cwo257c3hviqkle&i=yvslwei7jtf6tgqamzcdantqi4&h=llamaindex.1password.com)
|
|
31
33
|
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
llama_cloud/__init__.py,sha256=
|
|
2
|
-
llama_cloud/client.py,sha256=
|
|
1
|
+
llama_cloud/__init__.py,sha256=dwxeZ2TI6QaKfB791DPGvzcGkg2VBV3hAqAhs_4aMcY,24113
|
|
2
|
+
llama_cloud/client.py,sha256=JGV9tNaEgSsDy8vYv7RMPXTmXwbfc9gPvuEfQHYqrq4,5651
|
|
3
3
|
llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
|
|
4
4
|
llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
|
5
5
|
llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
|
|
@@ -9,22 +9,21 @@ llama_cloud/core/remove_none_from_dict.py,sha256=8m91FC3YuVem0Gm9_sXhJ2tGvP33owJ
|
|
|
9
9
|
llama_cloud/environment.py,sha256=feTjOebeFZMrBdnHat4RE5aHlpt-sJm4NhK4ntV1htI,167
|
|
10
10
|
llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6g,170
|
|
11
11
|
llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
|
|
12
|
-
llama_cloud/resources/__init__.py,sha256=
|
|
12
|
+
llama_cloud/resources/__init__.py,sha256=lggYCM9fT5FLxGeehXyAkl6WA_7VxyYwLKdWOOSAgQ4,4035
|
|
13
|
+
llama_cloud/resources/admin/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
14
|
+
llama_cloud/resources/admin/client.py,sha256=tIfM5KMJXRL0AUAm_s_fx5OzgqMUIxksjhCGuviQXQk,3080
|
|
13
15
|
llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
14
16
|
llama_cloud/resources/beta/client.py,sha256=eRB3mGmNxbhVGTtUpp-j-2APkHUoCbUckIz9coYjCsM,14666
|
|
15
17
|
llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
16
18
|
llama_cloud/resources/chat_apps/client.py,sha256=orSI8rpQbUwVEToolEeiEi5Qe--suXFvfu6D9JDii5I,23595
|
|
17
|
-
llama_cloud/resources/
|
|
18
|
-
llama_cloud/resources/
|
|
19
|
-
llama_cloud/resources/data_sinks/__init__.py,sha256=
|
|
20
|
-
llama_cloud/resources/data_sinks/
|
|
21
|
-
llama_cloud/resources/
|
|
22
|
-
llama_cloud/resources/data_sinks/types/data_sink_update_component.py,sha256=jSMVBQWeMw2BJQA-rviqC6a4PexBulmFg8IdzkqKVvQ,2284
|
|
23
|
-
llama_cloud/resources/data_sources/__init__.py,sha256=o3JSGGw1IGSLg_E219fFVlVCCzBuTqlnAl-yoQS3nME,1235
|
|
19
|
+
llama_cloud/resources/data_sinks/__init__.py,sha256=ZHUjn3HbKhq_7QS1q74r2m5RGKF5lxcvF2P6pGvpcis,147
|
|
20
|
+
llama_cloud/resources/data_sinks/client.py,sha256=GpD6FhbGqkg2oUToyMG6J8hPxG_iG7W5ZJRo0qg3yzk,20639
|
|
21
|
+
llama_cloud/resources/data_sinks/types/__init__.py,sha256=M1aTcufJwiEZo9B0KmYj9PfkSd6I1ooFt9tpIRGwgg8,168
|
|
22
|
+
llama_cloud/resources/data_sinks/types/data_sink_update_component.py,sha256=EWbsPt3k_w_vySf01iiFanyN7UVNzSOM3weHzx-Y_rk,809
|
|
23
|
+
llama_cloud/resources/data_sources/__init__.py,sha256=McURkcNBGHXH1hmRDRmZI1dRzJrekCTHZsgv03r2oZI,227
|
|
24
24
|
llama_cloud/resources/data_sources/client.py,sha256=SZFm8bW5nkaXringdSnmxHqvVjKM7cNNOtqVXjgTKhc,21855
|
|
25
|
-
llama_cloud/resources/data_sources/types/__init__.py,sha256=
|
|
26
|
-
llama_cloud/resources/data_sources/types/data_source_update_component.py,sha256=
|
|
27
|
-
llama_cloud/resources/data_sources/types/data_source_update_component_one.py,sha256=BeXgQB9aRR88S_pRlU5Ru5HBVorumM1oTq-zer0uCyE,3742
|
|
25
|
+
llama_cloud/resources/data_sources/types/__init__.py,sha256=Cd5xEECTzXqQSfJALfJPSjudlSLeb3RENeJVi8vwPbM,303
|
|
26
|
+
llama_cloud/resources/data_sources/types/data_source_update_component.py,sha256=u9sYcs3A4ZDzKjWCH3W9xIXCcLkZkVZxwoFOhEluqJU,1173
|
|
28
27
|
llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py,sha256=3aFC-p8MSxjhOu2nFtqk0pixj6RqNqcFnbOYngUdZUk,215
|
|
29
28
|
llama_cloud/resources/embedding_model_configs/__init__.py,sha256=cXDtKKq-gj7yjFjdQ5GrGyPs-T5tRV_0JjUMGlAbdUs,1115
|
|
30
29
|
llama_cloud/resources/embedding_model_configs/client.py,sha256=2JDvZJtSger9QJ8luPct-2zvwjaJAR8VcKsTZ1wgYTE,17769
|
|
@@ -39,9 +38,9 @@ llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py,sh
|
|
|
39
38
|
llama_cloud/resources/files/types/file_create_permission_info_value.py,sha256=KPCFuEaa8NiB85A5MfdXRAQ0poAUTl7Feg6BTfmdWas,209
|
|
40
39
|
llama_cloud/resources/files/types/file_create_resource_info_value.py,sha256=R7Y-CJf7fnbvIqE3xOI5XOrmPwLbVJLC7zpxMu8Zopk,201
|
|
41
40
|
llama_cloud/resources/jobs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
42
|
-
llama_cloud/resources/jobs/client.py,sha256=
|
|
41
|
+
llama_cloud/resources/jobs/client.py,sha256=tIrIVOZ6p60qTJ2dqhSrqHGGxLi0fzn9V7jJU9QS2Nw,6030
|
|
43
42
|
llama_cloud/resources/llama_extract/__init__.py,sha256=jRUugj6XARMpKZi3e2RkfTdcOSuE-Zy0IfScRLlyYMs,819
|
|
44
|
-
llama_cloud/resources/llama_extract/client.py,sha256=
|
|
43
|
+
llama_cloud/resources/llama_extract/client.py,sha256=oL5_3ugPde9Ch8QnwvRvFeTMMl0y_ST6i1hoSt2S5E4,63503
|
|
45
44
|
llama_cloud/resources/llama_extract/types/__init__.py,sha256=ZRBD-jg1qdXyiJKTxgH7zaadoDzuof1TYpjK4P5z4zA,1216
|
|
46
45
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py,sha256=zB31hJQ8hKaIsPkfTWiX5hqsPVFMyyeWEDZ_Aq237jo,305
|
|
47
46
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py,sha256=xoyXH3f0Y5beMWBxmtXSz6QoB_df_-0QBsYdjBhZnGw,217
|
|
@@ -52,11 +51,11 @@ llama_cloud/resources/llama_extract/types/extract_job_create_batch_data_schema_o
|
|
|
52
51
|
llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema.py,sha256=uMqpKJdCmUNtryS2bkQTNA1AgDlWdtsBOP31iMt3zNA,346
|
|
53
52
|
llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_schema_zero_value.py,sha256=cUS7ez5r0Vx8T7SxwLYptZMmvpT5JoDVMyn54Q6VL-g,227
|
|
54
53
|
llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
55
|
-
llama_cloud/resources/organizations/client.py,sha256=
|
|
54
|
+
llama_cloud/resources/organizations/client.py,sha256=56d5VcRg_3Lu_MMdfJCR8uoirBoxmj9vaQbHpwY2zWo,56243
|
|
56
55
|
llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
57
|
-
llama_cloud/resources/parsing/client.py,sha256=
|
|
56
|
+
llama_cloud/resources/parsing/client.py,sha256=U5J2g78CWYgujfLXlLEIoqatFSPnaG1OYNanx1jm_GQ,78177
|
|
58
57
|
llama_cloud/resources/pipelines/__init__.py,sha256=Mx7p3jDZRLMltsfywSufam_4AnHvmAfsxtMHVI72e-8,1083
|
|
59
|
-
llama_cloud/resources/pipelines/client.py,sha256=
|
|
58
|
+
llama_cloud/resources/pipelines/client.py,sha256=pfHugOoBmmfawJIGw4tegLxCNQS5Pxki0aBvYn3dn0Q,129708
|
|
60
59
|
llama_cloud/resources/pipelines/types/__init__.py,sha256=jjaMc0V3K1HZLMYZ6WT4ydMtBCVy-oF5koqTCovbDws,1202
|
|
61
60
|
llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
|
|
62
61
|
llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
|
|
@@ -68,8 +67,8 @@ llama_cloud/resources/reports/client.py,sha256=kHjtXVVc1Xi3T1GyBvSW5K4mTdr6xQwZA
|
|
|
68
67
|
llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwHww2YU90lOonBPTmZIk,291
|
|
69
68
|
llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
|
|
70
69
|
llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
71
|
-
llama_cloud/resources/retrievers/client.py,sha256=
|
|
72
|
-
llama_cloud/types/__init__.py,sha256=
|
|
70
|
+
llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
|
|
71
|
+
llama_cloud/types/__init__.py,sha256=_aT6mWFd4wz3Bw63_HQoGr0PDdbpf0bDhDpMEypGSGU,28507
|
|
73
72
|
llama_cloud/types/advanced_mode_transform_config.py,sha256=4xCXye0_cPmVS1F8aNTx81sIaEPjQH9kiCCAIoqUzlI,1502
|
|
74
73
|
llama_cloud/types/advanced_mode_transform_config_chunking_config.py,sha256=wYbJnWLpeQDfhmDZz-wJfYzD1iGT5Jcxb9ga3mzUuvk,1983
|
|
75
74
|
llama_cloud/types/advanced_mode_transform_config_segmentation_config.py,sha256=anNGq0F5-IlbIW3kpC8OilzLJnUq5tdIcWHnRnmlYsg,1303
|
|
@@ -102,7 +101,7 @@ llama_cloud/types/cloud_confluence_data_source.py,sha256=ok8BOv51SC4Ia9kX3DC8LuZ
|
|
|
102
101
|
llama_cloud/types/cloud_document.py,sha256=Rg_H8lcz2TzxEAIdU-m5mGpkM7s0j1Cn4JHkXYddmGs,1255
|
|
103
102
|
llama_cloud/types/cloud_document_create.py,sha256=fQ1gZAtLCpr-a-sPbMez_5fK9JMU3uyp2tNvIzWNG3U,1278
|
|
104
103
|
llama_cloud/types/cloud_google_drive_data_source.py,sha256=jf5k7SY8scR-8_X27ShYSh1vOiFteqIH6cNcG7xZGLE,1304
|
|
105
|
-
llama_cloud/types/cloud_jira_data_source.py,sha256=
|
|
104
|
+
llama_cloud/types/cloud_jira_data_source.py,sha256=9R20k8Ne0Bl9X5dgSxpM_IGOFmC70Llz0pJ93rAKRvw,1458
|
|
106
105
|
llama_cloud/types/cloud_milvus_vector_store.py,sha256=CHFTJSYPZKYPUU-jpB1MG8OwRvnPiT07o7cYCvQMZLA,1235
|
|
107
106
|
llama_cloud/types/cloud_mongo_db_atlas_vector_search.py,sha256=CQ9euGBd3a72dvpTapRBhakme-fQbY2OaSoe0GDSHDo,1771
|
|
108
107
|
llama_cloud/types/cloud_notion_page_data_source.py,sha256=DxYullFctkpd0A75lfTmPzf-9EjBlusMTtNs3RbmIag,1230
|
|
@@ -110,7 +109,7 @@ llama_cloud/types/cloud_one_drive_data_source.py,sha256=ryDLKD7FVvXGo5maj92CSe52
|
|
|
110
109
|
llama_cloud/types/cloud_pinecone_vector_store.py,sha256=d1jEezwE6ndNG-2izgoO_m9tG3N1ZFvmeCXI2r3miFc,1724
|
|
111
110
|
llama_cloud/types/cloud_postgres_vector_store.py,sha256=xWACT9JPqCfoBTGu68IVO9F52W2bTugFOoVQo49oi3M,1391
|
|
112
111
|
llama_cloud/types/cloud_qdrant_vector_store.py,sha256=F-gjNArzwLWmqgPcC-ZxRqSrhTFZbv5kqmIXmnLPB5o,1718
|
|
113
|
-
llama_cloud/types/cloud_s_3_data_source.py,sha256=
|
|
112
|
+
llama_cloud/types/cloud_s_3_data_source.py,sha256=tTT0us3oNatduTpuLPiOqBg-YPaIKX1HVujJwzlmmBA,1416
|
|
114
113
|
llama_cloud/types/cloud_sharepoint_data_source.py,sha256=iJtlgb4hsj8CP2IJ7TxdK1GOb3MdyKr7_jsOlY3kFiE,1609
|
|
115
114
|
llama_cloud/types/cloud_slack_data_source.py,sha256=tlsNj-hDj1gWmM0Q2A1BeyolfaPg_wfvSlJGTETknAo,1374
|
|
116
115
|
llama_cloud/types/cohere_embedding.py,sha256=wkv_fVCA1WEroGawzPFExwmiJ75gPfzeeemty7NBlsM,1579
|
|
@@ -121,25 +120,21 @@ llama_cloud/types/composite_retrieved_text_node.py,sha256=eTQ99cdZ2PASff5n4oVV1o
|
|
|
121
120
|
llama_cloud/types/composite_retrieved_text_node_with_score.py,sha256=o-HvmyjqODc68zYuobtj10_62FMBAKRLfRoTHGDdmxw,1148
|
|
122
121
|
llama_cloud/types/configurable_data_sink_names.py,sha256=0Yk9i8hcNXKCcSKpa5KwsCwy_EDeodqbny7qmF86_lM,1225
|
|
123
122
|
llama_cloud/types/configurable_data_source_names.py,sha256=mNW71sSgcVhU3kePAOUgRxeqK1Vo7F_J1xIzmYKPRq0,1971
|
|
124
|
-
llama_cloud/types/configurable_transformation_definition.py,sha256=LDOhI5IDxlLDWM_p_xwCFM7qq1y-aGA8UxN7dnplDlU,1886
|
|
125
|
-
llama_cloud/types/configurable_transformation_names.py,sha256=N_YhY8IuQxsqBteCibaQwEaY0zd6Ncb6jW69d9mjrdU,1898
|
|
126
123
|
llama_cloud/types/credit_type.py,sha256=nwSRKDWgHk_msdWitctqtyeZwj5EFd6VLto6NF2yCd4,971
|
|
127
124
|
llama_cloud/types/data_sink.py,sha256=PeexYHHoD8WkVp9WsFtfC-AIWszcgeJUprG1bwC8WsQ,1498
|
|
128
|
-
llama_cloud/types/data_sink_component.py,sha256=
|
|
125
|
+
llama_cloud/types/data_sink_component.py,sha256=uvuxLY3MPDpv_bkT0y-tHSZVPRSHCkDBDHVff-036Dg,749
|
|
129
126
|
llama_cloud/types/data_sink_create.py,sha256=dAaFPCwZ5oX0Fbf7ij62dzSaYnrhj3EHmnLnYnw2KgI,1360
|
|
130
|
-
llama_cloud/types/data_sink_create_component.py,sha256=
|
|
131
|
-
llama_cloud/types/data_sink_definition.py,sha256=5ve_pq02s8szc34-wWobMe6BAPj_c7e9n9FFsfDqEQ0,1561
|
|
127
|
+
llama_cloud/types/data_sink_create_component.py,sha256=8QfNKSTJV_sQ0nJxlpfh0fBkMTSnQD1DTJR8ZMYaesI,755
|
|
132
128
|
llama_cloud/types/data_source.py,sha256=4_lTRToLO4u9LYK66VygCPycrZuyct_aiovlxG5H2sE,1768
|
|
133
|
-
llama_cloud/types/data_source_component.py,sha256=
|
|
134
|
-
llama_cloud/types/data_source_component_one.py,sha256=TRYXPzpIt79ZcwIJn0RKnozeS7IgDyxkLiALo7jpSFM,3526
|
|
129
|
+
llama_cloud/types/data_source_component.py,sha256=yfXHoeHaqUMum7fIs3tZB0pOFMhDbAq7oCJtnob0gWY,1077
|
|
135
130
|
llama_cloud/types/data_source_create.py,sha256=s0bAX_GUwiRdrL-PXS9ROrvq3xpmqbqzdMa6thqL2P4,1581
|
|
136
|
-
llama_cloud/types/data_source_create_component.py,sha256
|
|
137
|
-
llama_cloud/types/data_source_create_component_one.py,sha256=HE2afkjNfnI3xwWoVbS3GXZEAEQa2tt4Wl0fs2RqfAs,3652
|
|
131
|
+
llama_cloud/types/data_source_create_component.py,sha256=-P4FGv9Xg951n-77_bb-2_CF-33ZXcUkw52LPQNunBY,1083
|
|
138
132
|
llama_cloud/types/data_source_create_custom_metadata_value.py,sha256=ejSsQNbszYQaUWFh9r9kQpHf88qbhuRv1SI9J_MOSC0,215
|
|
139
133
|
llama_cloud/types/data_source_custom_metadata_value.py,sha256=pTZn5yjZYmuOhsLABFJOKZblZUkRqo1CqLAuP5tKji4,209
|
|
140
|
-
llama_cloud/types/data_source_definition.py,sha256=HlSlTxzYcQJOSo_2OSroAE8vAr-otDvTNBSEkA54vL8,1575
|
|
141
134
|
llama_cloud/types/data_source_update_dispatcher_config.py,sha256=Sh6HhXfEV2Z6PYhkYQucs2MxyKVpL3UPV-I4cbf--bA,1242
|
|
142
135
|
llama_cloud/types/delete_params.py,sha256=1snPrd3WO9C1bKf0WdMslE2HQMF0yYLI3U7N53cmurM,1285
|
|
136
|
+
llama_cloud/types/document_block.py,sha256=OYKd5M3LgJ0Cz0K0YNuVRoHz9HcUdVuf2Vcqku8fck4,1116
|
|
137
|
+
llama_cloud/types/document_chunk_mode.py,sha256=6qH43Q0lIob2DMU95GsmSEOs4kQxOIyUFXj_kRDnyV4,470
|
|
143
138
|
llama_cloud/types/document_ingestion_job_params.py,sha256=33xTAl-K-m1j_Ufkj7w2GaYg9EUH5Hwsjn869X-fWMk,1524
|
|
144
139
|
llama_cloud/types/edit_suggestion.py,sha256=uzXSZYJiU3FaNN-TvEd3EXdaXvjQIe7Mf4kntKkyB2I,1202
|
|
145
140
|
llama_cloud/types/edit_suggestion_blocks_item.py,sha256=ojTk4lh0IHmrWP5wLPTIlsc2jAUDoHvdjJ5sm2uMut0,236
|
|
@@ -151,12 +146,13 @@ llama_cloud/types/embedding_model_config_update_embedding_config.py,sha256=mrXFx
|
|
|
151
146
|
llama_cloud/types/eval_execution_params.py,sha256=ntVaJh5SMZMPL4QLUiihVjUlg2SKbrezvbMKGlrF66Q,1369
|
|
152
147
|
llama_cloud/types/extract_agent.py,sha256=T98IOueut4M52Qm7hqcUOcWFFDhZ-ye0OFdXgfFGtS4,1763
|
|
153
148
|
llama_cloud/types/extract_agent_data_schema_value.py,sha256=UaDQ2KjajLDccW7F4NKdfpefeTJrr1hl0c95WRETYkM,201
|
|
154
|
-
llama_cloud/types/extract_config.py,sha256=
|
|
149
|
+
llama_cloud/types/extract_config.py,sha256=tCVzcctlEaO4tx0ikAv0OyfiDDhiiynR2FYILgOCCe4,1936
|
|
155
150
|
llama_cloud/types/extract_job.py,sha256=Yx4fDdCdylAji2LPTwqflVpz1o9slpj9tTLS93-1tzU,1431
|
|
156
151
|
llama_cloud/types/extract_job_create.py,sha256=UK1mBIKyflo7e6m1MxMN95pLscj67jH_yvs8EvmBXqU,1545
|
|
157
152
|
llama_cloud/types/extract_job_create_data_schema_override.py,sha256=vuiJ2lGJjbXEnvFKzVnKyvgwhMXPg1Pb5GZne2DrB60,330
|
|
158
153
|
llama_cloud/types/extract_job_create_data_schema_override_zero_value.py,sha256=HHEYxOSQXXyBYOiUQg_qwfQtXFj-OtThMwbUDBIgZU0,223
|
|
159
|
-
llama_cloud/types/extract_mode.py,sha256=
|
|
154
|
+
llama_cloud/types/extract_mode.py,sha256=DwTMzDq3HHJop_fxQelHEE_k8UcdDz-W_v_Oj2WWXLk,931
|
|
155
|
+
llama_cloud/types/extract_models.py,sha256=qXsSXuPFiVBNsh1-4oTYAjJ_36AQJ7sOJyJ-5aNlYEE,977
|
|
160
156
|
llama_cloud/types/extract_resultset.py,sha256=Alje0YQJUiA_aKi0hQs7TAnhDmZuQ_yL9b6HCNYBFQg,1627
|
|
161
157
|
llama_cloud/types/extract_resultset_data.py,sha256=v9Ae4SxLsvYPE9crko4N16lBjsxuZpz1yrUOhnaM_VY,427
|
|
162
158
|
llama_cloud/types/extract_resultset_data_item_value.py,sha256=JwqgDIGW0irr8QWaSTIrl24FhGxTUDOXIbxoSdIjuxs,209
|
|
@@ -199,9 +195,12 @@ llama_cloud/types/l_lama_parse_transform_config.py,sha256=YQRJZvKh1Ee2FUyW_N0nqY
|
|
|
199
195
|
llama_cloud/types/legacy_parse_job_config.py,sha256=kVBdiSLraI9rKQOPf0Ci9RtbNLkco0byBJC42uE_PCI,11698
|
|
200
196
|
llama_cloud/types/llama_extract_settings.py,sha256=IQFxtKa4GtHKc9w-fLwsH0LSKDWzR9_vZ_cTFJ9cGBI,2288
|
|
201
197
|
llama_cloud/types/llama_index_core_base_llms_types_chat_message.py,sha256=NelHo-T-ebVMhRKsqE_xV8AJW4c7o6lS0uEQnPsmTwg,1365
|
|
202
|
-
llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py,sha256
|
|
198
|
+
llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py,sha256=-aL8fh-w2Xf4uQs_LHzb3q6LL_onLAcVzCR5yMI4qJw,1571
|
|
203
199
|
llama_cloud/types/llama_parse_parameters.py,sha256=DNhVZm3YQ_3xZiz7WUrwH7E6jqW2fZ7YGFsdfsYalUk,5773
|
|
204
200
|
llama_cloud/types/llama_parse_supported_file_extensions.py,sha256=B_0N3f8Aq59W9FbsH50mGBUiyWTIXQjHFl739uAyaQw,11207
|
|
201
|
+
llama_cloud/types/llm_config_result.py,sha256=rruY77YClxo6X7gUFyknmMHoBFI-sWYMsdlSUvMn83I,1127
|
|
202
|
+
llama_cloud/types/llm_config_result_llm_type.py,sha256=yrijlC2f1egNDx-tCvDVp68pFmGaJZvVE_D1vS2wA34,1032
|
|
203
|
+
llama_cloud/types/llm_configs_response.py,sha256=TQ9RLkOzdY-8k0l1NNUZNUl83dmKks0gOQvhj-4hRZs,1073
|
|
205
204
|
llama_cloud/types/llm_model_data.py,sha256=6rrycqGwlK3LZ2S-WtgmeomithdLhDCgwBBZQ5KLaso,1300
|
|
206
205
|
llama_cloud/types/llm_parameters.py,sha256=RTKYt09lm9a1MlnBfYuTP2x_Ww4byUNNc1TqIel5O1Y,1377
|
|
207
206
|
llama_cloud/types/load_files_job_config.py,sha256=R5sFgFmV__0mqLUuD7dkFoBJHG2ZLw5px9zRapvYcpE,1069
|
|
@@ -245,14 +244,13 @@ llama_cloud/types/permission.py,sha256=LjhZdo0oLvk7ZVIF1d6Qja--AKH5Ri0naUhuJvZS6
|
|
|
245
244
|
llama_cloud/types/pg_vector_distance_method.py,sha256=U81o0ARjPR-HuFcVspHiJUrjIDJo3jLhB46vkITDu7M,1203
|
|
246
245
|
llama_cloud/types/pg_vector_hnsw_settings.py,sha256=-RE59xUgHwNEyAwRYmOQ8SHeAqkSYBfCAROw7QomxUU,1758
|
|
247
246
|
llama_cloud/types/pg_vector_vector_type.py,sha256=VwOohN566zw42UMlnuKTJopYJypsSnzWjCFmKRoU-bo,952
|
|
248
|
-
llama_cloud/types/pipeline.py,sha256=
|
|
247
|
+
llama_cloud/types/pipeline.py,sha256=4m1NIqTtG2DItvW69SWW3NjZPBL848VEW69Qbt2B7uo,2728
|
|
249
248
|
llama_cloud/types/pipeline_configuration_hashes.py,sha256=7_MbOcPWV6iyMflJeXoo9vLzD04E5WM7YxYp4ls0jQs,1169
|
|
250
|
-
llama_cloud/types/pipeline_create.py,sha256=
|
|
249
|
+
llama_cloud/types/pipeline_create.py,sha256=PKchM5cxkidXVFv2qON0uVh5lv8aqsy5OrZvT5UzqTU,2496
|
|
251
250
|
llama_cloud/types/pipeline_create_embedding_config.py,sha256=PQqmVBFUyZXYKKBmVQF2zPsGp1L6rje6g3RtXEcdfc8,2811
|
|
252
251
|
llama_cloud/types/pipeline_create_transform_config.py,sha256=HP6tzLsw_pomK1Ye2PYCS_XDZK_TMgg22mz17_zYKFg,303
|
|
253
252
|
llama_cloud/types/pipeline_data_source.py,sha256=g8coq6ohp09TtqzvB3_A8Nzery3J5knIfxGWzUtozmg,2381
|
|
254
|
-
llama_cloud/types/pipeline_data_source_component.py,sha256=
|
|
255
|
-
llama_cloud/types/pipeline_data_source_component_one.py,sha256=7GnJ1tPFIPjpL949T56pFYHhbvdRPw5VTMHmuOmVV8c,3694
|
|
253
|
+
llama_cloud/types/pipeline_data_source_component.py,sha256=c_R2aBl7XXsfJ_ZuK_-PXzzL2nDI4jrbJ0BStlzp87Y,1085
|
|
256
254
|
llama_cloud/types/pipeline_data_source_create.py,sha256=wMsymqB-YGyf3jdQr-N5ODVG6v0w68EMxGBNdQXeJe0,1178
|
|
257
255
|
llama_cloud/types/pipeline_data_source_custom_metadata_value.py,sha256=8n3r60sxMx4_udW0yzJZxzyWeK6L3cc2-jLGZFW4EDs,217
|
|
258
256
|
llama_cloud/types/pipeline_data_source_status.py,sha256=BD4xoftwp9lWC8EjJTnf3boIG_AyzjLPuP4qJxGhmcc,1039
|
|
@@ -270,6 +268,7 @@ llama_cloud/types/pipeline_file_update_dispatcher_config.py,sha256=PiJ1brbKGyq07
|
|
|
270
268
|
llama_cloud/types/pipeline_file_updater_config.py,sha256=KMHBYpH3fYDQaDVvxVgckosiWz0Dl3v5dC53Cgnmtb8,1761
|
|
271
269
|
llama_cloud/types/pipeline_managed_ingestion_job_params.py,sha256=ahliOe6YnLI-upIq1v5HZd9p8xH6pPdkh2M_n_zM9TA,1180
|
|
272
270
|
llama_cloud/types/pipeline_metadata_config.py,sha256=yMnPu6FnhagjuJ_rQ756WbIvVG5dzyXT1fmCYUAmCS0,1291
|
|
271
|
+
llama_cloud/types/pipeline_status.py,sha256=aC340nhfuPSrFVZOH_DhgYHWe985J3WNHrwvUtjXTRA,481
|
|
273
272
|
llama_cloud/types/pipeline_transform_config.py,sha256=zMr-ePLKGjbaScxbAHaSwYBL7rrNibVlnn0cbgElDfU,824
|
|
274
273
|
llama_cloud/types/pipeline_type.py,sha256=tTqrhxHP5xd7W2dQGD0e5FOv886nwJssyaVlXpWrtRo,551
|
|
275
274
|
llama_cloud/types/plan_limits.py,sha256=WAbDbRl8gsQxvhmuVB0YT8mry-0uKg6c66uivyppdQU,2056
|
|
@@ -282,7 +281,7 @@ llama_cloud/types/progress_event.py,sha256=Bk73A8geTVaq0ze5pMnbkAmx7FSOHQIixYCpC
|
|
|
282
281
|
llama_cloud/types/progress_event_status.py,sha256=yb4RAXwOKU6Bi7iyYy-3lwhF6_mLz0ZFyGjxIdaByoE,893
|
|
283
282
|
llama_cloud/types/project.py,sha256=4NNh_ZAjEkoWl5st6b1jsPVf_SYKtUTB6rS1701G4IQ,1441
|
|
284
283
|
llama_cloud/types/project_create.py,sha256=GxGmsXGJM-cHrvPFLktEkj9JtNsSdFae7-HPZFB4er0,1014
|
|
285
|
-
llama_cloud/types/prompt_conf.py,sha256=
|
|
284
|
+
llama_cloud/types/prompt_conf.py,sha256=hh8I3jxk3K6e5QZoBCLqszohMYtk73PERYoL36lLmTk,1660
|
|
286
285
|
llama_cloud/types/re_rank_config.py,sha256=mxRWwrC5BLg3DP1yEyRwW2lIpv5BuXZfTy8f4RbcOp0,1262
|
|
287
286
|
llama_cloud/types/re_ranker_type.py,sha256=qYItMEHrf80ePBp7gNGBSL67mkTIsqco92WJaJiYweo,1123
|
|
288
287
|
llama_cloud/types/recurring_credit_grant.py,sha256=19qI3p5k1mQ1Qoo-gCQU02Aa42XpEsmwxPF1F88F-Yg,1517
|
|
@@ -314,15 +313,14 @@ llama_cloud/types/semantic_chunking_config.py,sha256=dFDniTVWpRc7UcmVFvljUoyL5Zt
|
|
|
314
313
|
llama_cloud/types/sentence_chunking_config.py,sha256=NA9xidK5ICxJPkEMQZWNcsV0Hw9Co_bzRWeYe4uSh9I,1116
|
|
315
314
|
llama_cloud/types/status_enum.py,sha256=cUBIlys89E8PUzmVqqawu7qTDF0aRqBwiijOmRDPvx0,1018
|
|
316
315
|
llama_cloud/types/struct_mode.py,sha256=ROicwjXfFmgVU8_xSVxJlnFUzRNKG5VIEF1wYg9uOPU,1020
|
|
317
|
-
llama_cloud/types/struct_parse_conf.py,sha256=
|
|
318
|
-
llama_cloud/types/supported_llm_model.py,sha256=
|
|
319
|
-
llama_cloud/types/supported_llm_model_names.py,sha256=
|
|
316
|
+
llama_cloud/types/struct_parse_conf.py,sha256=WlL8y0IBvdzGsDtFUlEZLzoUODwmOWAJi0viS9unL18,2297
|
|
317
|
+
llama_cloud/types/supported_llm_model.py,sha256=hubSopFICVNEegbJbtbpK6zRHwFPwUNtrw_NAw_3bfg,1380
|
|
318
|
+
llama_cloud/types/supported_llm_model_names.py,sha256=NplDWGpsGoikd3By5sYhqL5IRFWBUnkz-8A3yf5pPcE,2180
|
|
320
319
|
llama_cloud/types/text_block.py,sha256=X154sQkSyposXuRcEWNp_tWcDQ-AI6q_-MfJUN5exP8,958
|
|
321
320
|
llama_cloud/types/text_node.py,sha256=Tq3QmuKC5cIHvC9wAtvhsXl1g2sACs2yJwQ0Uko8GSU,2846
|
|
322
321
|
llama_cloud/types/text_node_relationships_value.py,sha256=qmXURTk1Xg7ZDzRSSV1uDEel0AXRLohND5ioezibHY0,217
|
|
323
322
|
llama_cloud/types/text_node_with_score.py,sha256=k-KYWO_mgJBvO6xUfOD5W6v1Ku9E586_HsvDoQbLfuQ,1229
|
|
324
323
|
llama_cloud/types/token_chunking_config.py,sha256=XNvnTsNd--YOMQ_Ad8hoqhYgQftqkBHKVn6i7nJnMqs,1067
|
|
325
|
-
llama_cloud/types/transformation_category_names.py,sha256=Wb7NBB0f-tEtfEZQis-iKy71SUKmmHFcXf6XLn6g0XU,545
|
|
326
324
|
llama_cloud/types/usage_and_plan.py,sha256=bclc7TE7CTBu7RLiTHG426dziyj--I8m5NVu86I2AV4,1065
|
|
327
325
|
llama_cloud/types/usage_metric_response.py,sha256=ukvtNZLeLacv-5F0-GQ5wTBZOPUPEjAeurgYPc4s7nA,1047
|
|
328
326
|
llama_cloud/types/usage_response.py,sha256=o0u15PGNQmOOie4kJFfc4Rw0jKGLckBJdH0NCAfT8_k,1499
|
|
@@ -331,13 +329,13 @@ llama_cloud/types/user_job_record.py,sha256=mJHdokJsemXJOwM2l7fsW3X0SlwSNcy7yHbc
|
|
|
331
329
|
llama_cloud/types/user_organization.py,sha256=Ydel7grMnKiPMWJmSWhCFCm3v_n286Gk36ANtDLNLd4,1770
|
|
332
330
|
llama_cloud/types/user_organization_create.py,sha256=Zj57s9xuYVnLW2p8i4j2QORL-G1y7Ab3avXE1baERQY,1189
|
|
333
331
|
llama_cloud/types/user_organization_delete.py,sha256=IDYLKfFAXfcJfkEpA0ARbaA0JDcEBe7fTLv833DZXHs,1104
|
|
334
|
-
llama_cloud/types/user_organization_role.py,sha256=
|
|
332
|
+
llama_cloud/types/user_organization_role.py,sha256=vTM5pYG9NJpTQACn8vzSIt01Ul6jEHCVmyR3vV0isPg,1512
|
|
335
333
|
llama_cloud/types/validation_error.py,sha256=yZDLtjUHDY5w82Ra6CW0H9sLAr18R0RY1UNgJKR72DQ,1084
|
|
336
334
|
llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPXjdtN9EB7HrLVo6EP0,128
|
|
337
335
|
llama_cloud/types/vertex_ai_embedding_config.py,sha256=DvQk2xMJFmo54MEXTzoM4KSADyhGm_ygmFyx6wIcQdw,1159
|
|
338
336
|
llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2BaIMltDqGnIowU,1217
|
|
339
337
|
llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
|
|
340
|
-
llama_cloud-0.1.
|
|
341
|
-
llama_cloud-0.1.
|
|
342
|
-
llama_cloud-0.1.
|
|
343
|
-
llama_cloud-0.1.
|
|
338
|
+
llama_cloud-0.1.22.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
|
|
339
|
+
llama_cloud-0.1.22.dist-info/METADATA,sha256=8oiNq-d89044LUmFV1wybkJIdF5qcK4poEPPjh6UrtQ,1194
|
|
340
|
+
llama_cloud-0.1.22.dist-info/WHEEL,sha256=b4K_helf-jlQoXBBETfwnf4B04YC67LOev0jo4fX5m8,88
|
|
341
|
+
llama_cloud-0.1.22.dist-info/RECORD,,
|