llama-cloud 0.1.39__py3-none-any.whl → 0.1.41__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_cloud/__init__.py +24 -72
- llama_cloud/client.py +2 -5
- llama_cloud/resources/__init__.py +0 -4
- llama_cloud/resources/admin/client.py +5 -5
- llama_cloud/resources/beta/client.py +1351 -335
- llama_cloud/resources/llama_extract/client.py +56 -0
- llama_cloud/resources/parsing/client.py +8 -0
- llama_cloud/resources/pipelines/client.py +37 -0
- llama_cloud/types/__init__.py +24 -72
- llama_cloud/types/{src_app_schema_chat_chat_message.py → chat_message.py} +1 -1
- llama_cloud/types/extract_config.py +8 -2
- llama_cloud/types/extract_job_create.py +2 -0
- llama_cloud/types/extract_job_create_priority.py +29 -0
- llama_cloud/types/extract_models.py +28 -28
- llama_cloud/types/job_names.py +0 -4
- llama_cloud/types/{document_block.py → llama_extract_feature_availability.py} +5 -6
- llama_cloud/types/llama_extract_mode_availability.py +4 -3
- llama_cloud/types/llama_extract_settings.py +1 -1
- llama_cloud/types/llama_parse_parameters.py +1 -0
- llama_cloud/types/{progress_event.py → parse_configuration.py} +12 -12
- llama_cloud/types/{llama_index_core_base_llms_types_chat_message.py → parse_configuration_create.py} +9 -7
- llama_cloud/types/{edit_suggestion.py → parse_configuration_filter.py} +8 -6
- llama_cloud/types/{report_update_event.py → parse_configuration_query_response.py} +6 -6
- llama_cloud/types/parse_job_config.py +1 -0
- llama_cloud/types/pipeline.py +4 -0
- llama_cloud/types/pipeline_create.py +2 -0
- llama_cloud/types/playground_session.py +2 -2
- llama_cloud/types/public_model_name.py +97 -0
- llama_cloud/types/{report_create_response.py → schema_generation_availability.py} +4 -2
- llama_cloud/types/schema_generation_availability_status.py +17 -0
- llama_cloud/types/{report_event_item.py → sparse_model_config.py} +10 -8
- llama_cloud/types/sparse_model_type.py +33 -0
- llama_cloud/types/webhook_configuration.py +1 -0
- llama_cloud-0.1.41.dist-info/METADATA +106 -0
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/RECORD +37 -56
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/WHEEL +1 -1
- llama_cloud/resources/reports/__init__.py +0 -5
- llama_cloud/resources/reports/client.py +0 -1230
- llama_cloud/resources/reports/types/__init__.py +0 -7
- llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py +0 -25
- llama_cloud/types/audio_block.py +0 -34
- llama_cloud/types/edit_suggestion_blocks_item.py +0 -8
- llama_cloud/types/image_block.py +0 -35
- llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py +0 -56
- llama_cloud/types/paginated_report_response.py +0 -35
- llama_cloud/types/progress_event_status.py +0 -33
- llama_cloud/types/report.py +0 -33
- llama_cloud/types/report_block.py +0 -35
- llama_cloud/types/report_block_dependency.py +0 -29
- llama_cloud/types/report_event_item_event_data.py +0 -45
- llama_cloud/types/report_event_type.py +0 -37
- llama_cloud/types/report_metadata.py +0 -43
- llama_cloud/types/report_plan.py +0 -36
- llama_cloud/types/report_plan_block.py +0 -36
- llama_cloud/types/report_query.py +0 -33
- llama_cloud/types/report_response.py +0 -41
- llama_cloud/types/report_state.py +0 -37
- llama_cloud/types/report_state_event.py +0 -38
- llama_cloud/types/text_block.py +0 -31
- llama_cloud-0.1.39.dist-info/METADATA +0 -32
- {llama_cloud-0.1.39.dist-info → llama_cloud-0.1.41.dist-info}/LICENSE +0 -0
llama_cloud/types/pipeline.py
CHANGED
|
@@ -5,6 +5,7 @@ import typing
|
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .data_sink import DataSink
|
|
8
|
+
from .embedding_model_config import EmbeddingModelConfig
|
|
8
9
|
from .eval_execution_params import EvalExecutionParams
|
|
9
10
|
from .llama_parse_parameters import LlamaParseParameters
|
|
10
11
|
from .pipeline_configuration_hashes import PipelineConfigurationHashes
|
|
@@ -14,6 +15,7 @@ from .pipeline_status import PipelineStatus
|
|
|
14
15
|
from .pipeline_transform_config import PipelineTransformConfig
|
|
15
16
|
from .pipeline_type import PipelineType
|
|
16
17
|
from .preset_retrieval_params import PresetRetrievalParams
|
|
18
|
+
from .sparse_model_config import SparseModelConfig
|
|
17
19
|
|
|
18
20
|
try:
|
|
19
21
|
import pydantic
|
|
@@ -35,11 +37,13 @@ class Pipeline(pydantic.BaseModel):
|
|
|
35
37
|
name: str
|
|
36
38
|
project_id: str
|
|
37
39
|
embedding_model_config_id: typing.Optional[str]
|
|
40
|
+
embedding_model_config: typing.Optional[EmbeddingModelConfig]
|
|
38
41
|
pipeline_type: typing.Optional[PipelineType] = pydantic.Field(
|
|
39
42
|
description="Type of pipeline. Either PLAYGROUND or MANAGED."
|
|
40
43
|
)
|
|
41
44
|
managed_pipeline_id: typing.Optional[str]
|
|
42
45
|
embedding_config: PipelineEmbeddingConfig
|
|
46
|
+
sparse_model_config: typing.Optional[SparseModelConfig]
|
|
43
47
|
config_hash: typing.Optional[PipelineConfigurationHashes]
|
|
44
48
|
transform_config: typing.Optional[PipelineTransformConfig] = pydantic.Field(
|
|
45
49
|
description="Configuration for the transformation."
|
|
@@ -12,6 +12,7 @@ from .pipeline_create_transform_config import PipelineCreateTransformConfig
|
|
|
12
12
|
from .pipeline_metadata_config import PipelineMetadataConfig
|
|
13
13
|
from .pipeline_type import PipelineType
|
|
14
14
|
from .preset_retrieval_params import PresetRetrievalParams
|
|
15
|
+
from .sparse_model_config import SparseModelConfig
|
|
15
16
|
|
|
16
17
|
try:
|
|
17
18
|
import pydantic
|
|
@@ -31,6 +32,7 @@ class PipelineCreate(pydantic.BaseModel):
|
|
|
31
32
|
transform_config: typing.Optional[PipelineCreateTransformConfig] = pydantic.Field(
|
|
32
33
|
description="Configuration for the transformation."
|
|
33
34
|
)
|
|
35
|
+
sparse_model_config: typing.Optional[SparseModelConfig]
|
|
34
36
|
data_sink_id: typing.Optional[str]
|
|
35
37
|
embedding_model_config_id: typing.Optional[str]
|
|
36
38
|
data_sink: typing.Optional[DataSinkCreate]
|
|
@@ -4,9 +4,9 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
+
from .chat_message import ChatMessage
|
|
7
8
|
from .llm_parameters import LlmParameters
|
|
8
9
|
from .preset_retrieval_params import PresetRetrievalParams
|
|
9
|
-
from .src_app_schema_chat_chat_message import SrcAppSchemaChatChatMessage
|
|
10
10
|
|
|
11
11
|
try:
|
|
12
12
|
import pydantic
|
|
@@ -33,7 +33,7 @@ class PlaygroundSession(pydantic.BaseModel):
|
|
|
33
33
|
retrieval_params: typing.Optional[PresetRetrievalParams] = pydantic.Field(
|
|
34
34
|
description="Preset retrieval parameters last used in this session."
|
|
35
35
|
)
|
|
36
|
-
chat_messages: typing.Optional[typing.List[
|
|
36
|
+
chat_messages: typing.Optional[typing.List[ChatMessage]] = pydantic.Field(
|
|
37
37
|
description="Chat message history for this session."
|
|
38
38
|
)
|
|
39
39
|
|
|
@@ -0,0 +1,97 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class PublicModelName(str, enum.Enum):
|
|
10
|
+
OPENAI_GPT_4_O = "openai-gpt-4o"
|
|
11
|
+
OPENAI_GPT_4_O_MINI = "openai-gpt-4o-mini"
|
|
12
|
+
OPENAI_GPT_41 = "openai-gpt-4-1"
|
|
13
|
+
OPENAI_GPT_41_MINI = "openai-gpt-4-1-mini"
|
|
14
|
+
OPENAI_GPT_41_NANO = "openai-gpt-4-1-nano"
|
|
15
|
+
OPENAI_GPT_5 = "openai-gpt-5"
|
|
16
|
+
OPENAI_GPT_5_MINI = "openai-gpt-5-mini"
|
|
17
|
+
OPENAI_GPT_5_NANO = "openai-gpt-5-nano"
|
|
18
|
+
OPENAI_TEXT_EMBEDDING_3_SMALL = "openai-text-embedding-3-small"
|
|
19
|
+
OPENAI_TEXT_EMBEDDING_3_LARGE = "openai-text-embedding-3-large"
|
|
20
|
+
OPENAI_WHISPER_1 = "openai-whisper-1"
|
|
21
|
+
ANTHROPIC_SONNET_35 = "anthropic-sonnet-3.5"
|
|
22
|
+
ANTHROPIC_SONNET_35_V_2 = "anthropic-sonnet-3.5-v2"
|
|
23
|
+
ANTHROPIC_SONNET_37 = "anthropic-sonnet-3.7"
|
|
24
|
+
ANTHROPIC_SONNET_40 = "anthropic-sonnet-4.0"
|
|
25
|
+
GEMINI_25_FLASH = "gemini-2.5-flash"
|
|
26
|
+
GEMINI_25_PRO = "gemini-2.5-pro"
|
|
27
|
+
GEMINI_20_FLASH = "gemini-2.0-flash"
|
|
28
|
+
GEMINI_20_FLASH_LITE = "gemini-2.0-flash-lite"
|
|
29
|
+
GEMINI_15_FLASH = "gemini-1.5-flash"
|
|
30
|
+
GEMINI_15_PRO = "gemini-1.5-pro"
|
|
31
|
+
|
|
32
|
+
def visit(
|
|
33
|
+
self,
|
|
34
|
+
openai_gpt_4_o: typing.Callable[[], T_Result],
|
|
35
|
+
openai_gpt_4_o_mini: typing.Callable[[], T_Result],
|
|
36
|
+
openai_gpt_41: typing.Callable[[], T_Result],
|
|
37
|
+
openai_gpt_41_mini: typing.Callable[[], T_Result],
|
|
38
|
+
openai_gpt_41_nano: typing.Callable[[], T_Result],
|
|
39
|
+
openai_gpt_5: typing.Callable[[], T_Result],
|
|
40
|
+
openai_gpt_5_mini: typing.Callable[[], T_Result],
|
|
41
|
+
openai_gpt_5_nano: typing.Callable[[], T_Result],
|
|
42
|
+
openai_text_embedding_3_small: typing.Callable[[], T_Result],
|
|
43
|
+
openai_text_embedding_3_large: typing.Callable[[], T_Result],
|
|
44
|
+
openai_whisper_1: typing.Callable[[], T_Result],
|
|
45
|
+
anthropic_sonnet_35: typing.Callable[[], T_Result],
|
|
46
|
+
anthropic_sonnet_35_v_2: typing.Callable[[], T_Result],
|
|
47
|
+
anthropic_sonnet_37: typing.Callable[[], T_Result],
|
|
48
|
+
anthropic_sonnet_40: typing.Callable[[], T_Result],
|
|
49
|
+
gemini_25_flash: typing.Callable[[], T_Result],
|
|
50
|
+
gemini_25_pro: typing.Callable[[], T_Result],
|
|
51
|
+
gemini_20_flash: typing.Callable[[], T_Result],
|
|
52
|
+
gemini_20_flash_lite: typing.Callable[[], T_Result],
|
|
53
|
+
gemini_15_flash: typing.Callable[[], T_Result],
|
|
54
|
+
gemini_15_pro: typing.Callable[[], T_Result],
|
|
55
|
+
) -> T_Result:
|
|
56
|
+
if self is PublicModelName.OPENAI_GPT_4_O:
|
|
57
|
+
return openai_gpt_4_o()
|
|
58
|
+
if self is PublicModelName.OPENAI_GPT_4_O_MINI:
|
|
59
|
+
return openai_gpt_4_o_mini()
|
|
60
|
+
if self is PublicModelName.OPENAI_GPT_41:
|
|
61
|
+
return openai_gpt_41()
|
|
62
|
+
if self is PublicModelName.OPENAI_GPT_41_MINI:
|
|
63
|
+
return openai_gpt_41_mini()
|
|
64
|
+
if self is PublicModelName.OPENAI_GPT_41_NANO:
|
|
65
|
+
return openai_gpt_41_nano()
|
|
66
|
+
if self is PublicModelName.OPENAI_GPT_5:
|
|
67
|
+
return openai_gpt_5()
|
|
68
|
+
if self is PublicModelName.OPENAI_GPT_5_MINI:
|
|
69
|
+
return openai_gpt_5_mini()
|
|
70
|
+
if self is PublicModelName.OPENAI_GPT_5_NANO:
|
|
71
|
+
return openai_gpt_5_nano()
|
|
72
|
+
if self is PublicModelName.OPENAI_TEXT_EMBEDDING_3_SMALL:
|
|
73
|
+
return openai_text_embedding_3_small()
|
|
74
|
+
if self is PublicModelName.OPENAI_TEXT_EMBEDDING_3_LARGE:
|
|
75
|
+
return openai_text_embedding_3_large()
|
|
76
|
+
if self is PublicModelName.OPENAI_WHISPER_1:
|
|
77
|
+
return openai_whisper_1()
|
|
78
|
+
if self is PublicModelName.ANTHROPIC_SONNET_35:
|
|
79
|
+
return anthropic_sonnet_35()
|
|
80
|
+
if self is PublicModelName.ANTHROPIC_SONNET_35_V_2:
|
|
81
|
+
return anthropic_sonnet_35_v_2()
|
|
82
|
+
if self is PublicModelName.ANTHROPIC_SONNET_37:
|
|
83
|
+
return anthropic_sonnet_37()
|
|
84
|
+
if self is PublicModelName.ANTHROPIC_SONNET_40:
|
|
85
|
+
return anthropic_sonnet_40()
|
|
86
|
+
if self is PublicModelName.GEMINI_25_FLASH:
|
|
87
|
+
return gemini_25_flash()
|
|
88
|
+
if self is PublicModelName.GEMINI_25_PRO:
|
|
89
|
+
return gemini_25_pro()
|
|
90
|
+
if self is PublicModelName.GEMINI_20_FLASH:
|
|
91
|
+
return gemini_20_flash()
|
|
92
|
+
if self is PublicModelName.GEMINI_20_FLASH_LITE:
|
|
93
|
+
return gemini_20_flash_lite()
|
|
94
|
+
if self is PublicModelName.GEMINI_15_FLASH:
|
|
95
|
+
return gemini_15_flash()
|
|
96
|
+
if self is PublicModelName.GEMINI_15_PRO:
|
|
97
|
+
return gemini_15_pro()
|
|
@@ -4,6 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
+
from .schema_generation_availability_status import SchemaGenerationAvailabilityStatus
|
|
7
8
|
|
|
8
9
|
try:
|
|
9
10
|
import pydantic
|
|
@@ -14,8 +15,9 @@ except ImportError:
|
|
|
14
15
|
import pydantic # type: ignore
|
|
15
16
|
|
|
16
17
|
|
|
17
|
-
class
|
|
18
|
-
|
|
18
|
+
class SchemaGenerationAvailability(pydantic.BaseModel):
|
|
19
|
+
model: str
|
|
20
|
+
status: SchemaGenerationAvailabilityStatus
|
|
19
21
|
|
|
20
22
|
def json(self, **kwargs: typing.Any) -> str:
|
|
21
23
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class SchemaGenerationAvailabilityStatus(str, enum.Enum):
|
|
10
|
+
AVAILABLE = "available"
|
|
11
|
+
UNAVAILABLE = "unavailable"
|
|
12
|
+
|
|
13
|
+
def visit(self, available: typing.Callable[[], T_Result], unavailable: typing.Callable[[], T_Result]) -> T_Result:
|
|
14
|
+
if self is SchemaGenerationAvailabilityStatus.AVAILABLE:
|
|
15
|
+
return available()
|
|
16
|
+
if self is SchemaGenerationAvailabilityStatus.UNAVAILABLE:
|
|
17
|
+
return unavailable()
|
|
@@ -4,7 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .
|
|
7
|
+
from .sparse_model_type import SparseModelType
|
|
8
8
|
|
|
9
9
|
try:
|
|
10
10
|
import pydantic
|
|
@@ -15,16 +15,18 @@ except ImportError:
|
|
|
15
15
|
import pydantic # type: ignore
|
|
16
16
|
|
|
17
17
|
|
|
18
|
-
class
|
|
18
|
+
class SparseModelConfig(pydantic.BaseModel):
|
|
19
19
|
"""
|
|
20
|
-
|
|
20
|
+
Configuration for sparse embedding models used in hybrid search.
|
|
21
|
+
|
|
22
|
+
This allows users to choose between Splade and BM25 models for
|
|
23
|
+
sparse retrieval in managed data sinks.
|
|
21
24
|
"""
|
|
22
25
|
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
timestamp: dt.datetime = pydantic.Field(description="The timestamp for the event")
|
|
26
|
+
model_type: typing.Optional[SparseModelType] = pydantic.Field(
|
|
27
|
+
description="The sparse model type to use. 'auto' selects based on deployment mode (BYOC uses term frequency, Cloud uses Splade), 'splade' uses HuggingFace Splade model, 'bm25' uses Qdrant's FastEmbed BM25 model."
|
|
28
|
+
)
|
|
29
|
+
class_name: typing.Optional[str]
|
|
28
30
|
|
|
29
31
|
def json(self, **kwargs: typing.Any) -> str:
|
|
30
32
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class SparseModelType(str, enum.Enum):
|
|
10
|
+
"""
|
|
11
|
+
Enum for sparse model types supported in LlamaCloud.
|
|
12
|
+
|
|
13
|
+
SPLADE: Uses HuggingFace Splade model for sparse embeddings
|
|
14
|
+
BM25: Uses Qdrant's FastEmbed BM25 model for sparse embeddings
|
|
15
|
+
AUTO: Automatically selects based on deployment mode (BYOC uses term frequency, Cloud uses Splade)
|
|
16
|
+
"""
|
|
17
|
+
|
|
18
|
+
SPLADE = "splade"
|
|
19
|
+
BM_25 = "bm25"
|
|
20
|
+
AUTO = "auto"
|
|
21
|
+
|
|
22
|
+
def visit(
|
|
23
|
+
self,
|
|
24
|
+
splade: typing.Callable[[], T_Result],
|
|
25
|
+
bm_25: typing.Callable[[], T_Result],
|
|
26
|
+
auto: typing.Callable[[], T_Result],
|
|
27
|
+
) -> T_Result:
|
|
28
|
+
if self is SparseModelType.SPLADE:
|
|
29
|
+
return splade()
|
|
30
|
+
if self is SparseModelType.BM_25:
|
|
31
|
+
return bm_25()
|
|
32
|
+
if self is SparseModelType.AUTO:
|
|
33
|
+
return auto()
|
|
@@ -23,6 +23,7 @@ class WebhookConfiguration(pydantic.BaseModel):
|
|
|
23
23
|
webhook_url: typing.Optional[str]
|
|
24
24
|
webhook_headers: typing.Optional[typing.Dict[str, typing.Optional[str]]]
|
|
25
25
|
webhook_events: typing.Optional[typing.List[WebhookConfigurationWebhookEventsItem]]
|
|
26
|
+
webhook_output_format: typing.Optional[str]
|
|
26
27
|
|
|
27
28
|
def json(self, **kwargs: typing.Any) -> str:
|
|
28
29
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -0,0 +1,106 @@
|
|
|
1
|
+
Metadata-Version: 2.3
|
|
2
|
+
Name: llama-cloud
|
|
3
|
+
Version: 0.1.41
|
|
4
|
+
Summary:
|
|
5
|
+
License: MIT
|
|
6
|
+
Author: Logan Markewich
|
|
7
|
+
Author-email: logan@runllama.ai
|
|
8
|
+
Requires-Python: >=3.8,<4
|
|
9
|
+
Classifier: License :: OSI Approved :: MIT License
|
|
10
|
+
Classifier: Programming Language :: Python :: 3
|
|
11
|
+
Classifier: Programming Language :: Python :: 3.8
|
|
12
|
+
Classifier: Programming Language :: Python :: 3.9
|
|
13
|
+
Classifier: Programming Language :: Python :: 3.10
|
|
14
|
+
Classifier: Programming Language :: Python :: 3.11
|
|
15
|
+
Classifier: Programming Language :: Python :: 3.12
|
|
16
|
+
Classifier: Programming Language :: Python :: 3.13
|
|
17
|
+
Requires-Dist: certifi (>=2024.7.4)
|
|
18
|
+
Requires-Dist: httpx (>=0.20.0)
|
|
19
|
+
Requires-Dist: pydantic (>=1.10)
|
|
20
|
+
Description-Content-Type: text/markdown
|
|
21
|
+
|
|
22
|
+
# LlamaIndex Python Client
|
|
23
|
+
|
|
24
|
+
This client is auto-generated using [Fern](https://buildwithfern.com/docs/intro)
|
|
25
|
+
|
|
26
|
+
## Setup
|
|
27
|
+
|
|
28
|
+
Before testing or releasing, ensure you have the development environment set up:
|
|
29
|
+
|
|
30
|
+
```bash
|
|
31
|
+
# Install Poetry (if not already installed)
|
|
32
|
+
curl -sSL https://install.python-poetry.org | python3 -
|
|
33
|
+
|
|
34
|
+
# Install project dependencies
|
|
35
|
+
poetry install
|
|
36
|
+
|
|
37
|
+
# Verify setup
|
|
38
|
+
poetry --version
|
|
39
|
+
python --version
|
|
40
|
+
```
|
|
41
|
+
|
|
42
|
+
## Quick Start
|
|
43
|
+
|
|
44
|
+
```bash
|
|
45
|
+
# Set up development environment (first time only)
|
|
46
|
+
make setup
|
|
47
|
+
|
|
48
|
+
# See all available commands
|
|
49
|
+
make help
|
|
50
|
+
|
|
51
|
+
# Test before releasing
|
|
52
|
+
make test
|
|
53
|
+
|
|
54
|
+
# Complete release workflow
|
|
55
|
+
make release
|
|
56
|
+
```
|
|
57
|
+
|
|
58
|
+
## Commands
|
|
59
|
+
|
|
60
|
+
| Command | Description |
|
|
61
|
+
|---------|-------------|
|
|
62
|
+
| `make setup` | Set up development environment (first time only) |
|
|
63
|
+
| `make help` | Show all available commands |
|
|
64
|
+
| `make build` | Build the package |
|
|
65
|
+
| `make test` | Build, install locally, and run comprehensive tests |
|
|
66
|
+
| `make publish` | Test and publish to PyPI (with confirmation) |
|
|
67
|
+
| `make release` | Complete release workflow (clean → build → test → publish) |
|
|
68
|
+
| `make clean` | Clean build artifacts |
|
|
69
|
+
| `make setup-credentials` | Set up PyPI credentials (one-time setup) |
|
|
70
|
+
|
|
71
|
+
## Release Process
|
|
72
|
+
|
|
73
|
+
**First time setup:**
|
|
74
|
+
```bash
|
|
75
|
+
make setup # Set up development environment
|
|
76
|
+
```
|
|
77
|
+
|
|
78
|
+
**For each release:**
|
|
79
|
+
1. **Update version** in `pyproject.toml`
|
|
80
|
+
2. **Run release workflow:**
|
|
81
|
+
```bash
|
|
82
|
+
make release
|
|
83
|
+
```
|
|
84
|
+
|
|
85
|
+
That's it! The Makefile handles building, testing, and publishing with safety checks.
|
|
86
|
+
|
|
87
|
+
## Manual Steps (if needed)
|
|
88
|
+
|
|
89
|
+
<details>
|
|
90
|
+
<summary>Expand for manual commands</summary>
|
|
91
|
+
|
|
92
|
+
**Build and test manually:**
|
|
93
|
+
```bash
|
|
94
|
+
poetry build
|
|
95
|
+
pip install --force-reinstall dist/llama_cloud-*-py3-none-any.whl
|
|
96
|
+
python test_local_build.py
|
|
97
|
+
```
|
|
98
|
+
|
|
99
|
+
**Setup credentials manually:**
|
|
100
|
+
```bash
|
|
101
|
+
poetry config pypi-token.pypi <your-token>
|
|
102
|
+
```
|
|
103
|
+
Get token from [1Password](https://start.1password.com/open/i?a=32SA66TZ3JCRXOCMASLSDCT5TI&v=lhv7hvb5o46cwo257c3hviqkle&i=yvslwei7jtf6tgqamzcdantqi4&h=llamaindex.1password.com)
|
|
104
|
+
|
|
105
|
+
</details>
|
|
106
|
+
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
llama_cloud/__init__.py,sha256=
|
|
2
|
-
llama_cloud/client.py,sha256=
|
|
1
|
+
llama_cloud/__init__.py,sha256=D3PI7WP06XxMUsb_pO5oEM26Sa42KdKOjbCZZPL7kxk,26329
|
|
2
|
+
llama_cloud/client.py,sha256=GDYFdv8HLjksP7v9Srg2s0R1k_nouz2toh27EG3y110,6385
|
|
3
3
|
llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
|
|
4
4
|
llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
|
5
5
|
llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
|
|
@@ -9,15 +9,15 @@ llama_cloud/core/remove_none_from_dict.py,sha256=8m91FC3YuVem0Gm9_sXhJ2tGvP33owJ
|
|
|
9
9
|
llama_cloud/environment.py,sha256=feTjOebeFZMrBdnHat4RE5aHlpt-sJm4NhK4ntV1htI,167
|
|
10
10
|
llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6g,170
|
|
11
11
|
llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
|
|
12
|
-
llama_cloud/resources/__init__.py,sha256=
|
|
12
|
+
llama_cloud/resources/__init__.py,sha256=j5itg2tAAkpAKMu5oLAl9I9a79D9BMUYYHTvc1nHZfM,4147
|
|
13
13
|
llama_cloud/resources/admin/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
14
|
-
llama_cloud/resources/admin/client.py,sha256=
|
|
14
|
+
llama_cloud/resources/admin/client.py,sha256=iJClMzp6OQ_TOnAwgcPSb0BkEuuFeIq0r15lDmWUD0s,8502
|
|
15
15
|
llama_cloud/resources/agent_deployments/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
16
16
|
llama_cloud/resources/agent_deployments/client.py,sha256=3EOzOjmRs4KISgJ566enq3FCuN3YtskjO0OHqQGtkQ0,6122
|
|
17
17
|
llama_cloud/resources/alpha/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
18
18
|
llama_cloud/resources/alpha/client.py,sha256=d5cRIUykNpnVryuxWBPUpmo-2L1vMIDeZIF3DvTIx7E,4322
|
|
19
19
|
llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
20
|
-
llama_cloud/resources/beta/client.py,sha256=
|
|
20
|
+
llama_cloud/resources/beta/client.py,sha256=VznYZ3GeFHKcPXp_bgsitf4GEgqsqo2bvlseLc1d9gA,100996
|
|
21
21
|
llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
22
22
|
llama_cloud/resources/chat_apps/client.py,sha256=orSI8rpQbUwVEToolEeiEi5Qe--suXFvfu6D9JDii5I,23595
|
|
23
23
|
llama_cloud/resources/classifier/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
@@ -44,7 +44,7 @@ llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py,sh
|
|
|
44
44
|
llama_cloud/resources/jobs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
45
45
|
llama_cloud/resources/jobs/client.py,sha256=b2R_Oj2OCtcv-IIJNz9aq42hDgrOk_huqTSJhTB9VaA,6202
|
|
46
46
|
llama_cloud/resources/llama_extract/__init__.py,sha256=V6VZ8hQXwAuvOOZyk43nnbINoDQqEr03AjKQPhYKluk,997
|
|
47
|
-
llama_cloud/resources/llama_extract/client.py,sha256=
|
|
47
|
+
llama_cloud/resources/llama_extract/client.py,sha256=gk57ynYMMNlR0n_57w2MoGXLt_IfAQ5tJIvqFh0GYcM,85612
|
|
48
48
|
llama_cloud/resources/llama_extract/types/__init__.py,sha256=2Iu4w5LXZY2Govr1RzahIfY0b84y658SQjMDtj7rH_0,1497
|
|
49
49
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py,sha256=zB31hJQ8hKaIsPkfTWiX5hqsPVFMyyeWEDZ_Aq237jo,305
|
|
50
50
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py,sha256=xoyXH3f0Y5beMWBxmtXSz6QoB_df_-0QBsYdjBhZnGw,217
|
|
@@ -59,9 +59,9 @@ llama_cloud/resources/llama_extract/types/extract_stateless_request_data_schema_
|
|
|
59
59
|
llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
60
60
|
llama_cloud/resources/organizations/client.py,sha256=RoN-nkN7VeRZnrrElXhaPrgQFzGMHgNY41_XpbCXP0g,56623
|
|
61
61
|
llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
62
|
-
llama_cloud/resources/parsing/client.py,sha256=
|
|
62
|
+
llama_cloud/resources/parsing/client.py,sha256=aOoOGjpPqAYFNrkfNoqPjG1sXKQwGFYPGTzjpVuOLsY,89205
|
|
63
63
|
llama_cloud/resources/pipelines/__init__.py,sha256=zyvVEOF_krvEZkCIj_kZoMKfhDqHo_R32a1mv9CriQc,1193
|
|
64
|
-
llama_cloud/resources/pipelines/client.py,sha256=
|
|
64
|
+
llama_cloud/resources/pipelines/client.py,sha256=hecRBRZYiEbwKoM1-HEdilN_SSikkoqUu_mTDkg_Lq0,136403
|
|
65
65
|
llama_cloud/resources/pipelines/types/__init__.py,sha256=C68NQ5QzA0dFXf9oePFFGmV1vn96jcAp-QAznSgoRYQ,1375
|
|
66
66
|
llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
|
|
67
67
|
llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
|
|
@@ -69,15 +69,11 @@ llama_cloud/resources/pipelines/types/pipeline_update_transform_config.py,sha256
|
|
|
69
69
|
llama_cloud/resources/pipelines/types/retrieval_params_search_filters_inference_schema_value.py,sha256=hZWXYlTib0af85ECcerC4xD-bUQe8rG3Q6G1jFTMQcI,228
|
|
70
70
|
llama_cloud/resources/projects/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
71
71
|
llama_cloud/resources/projects/client.py,sha256=PF36iWtSa5amUt3q56YwLypOZjclIXSubCRv9NttpLs,25404
|
|
72
|
-
llama_cloud/resources/reports/__init__.py,sha256=cruYbQ1bIuJbRpkfaQY7ajUEslffjd7KzvzMzbtPH94,217
|
|
73
|
-
llama_cloud/resources/reports/client.py,sha256=kHjtXVVc1Xi3T1GyBvSW5K4mTdr6xQwZA3vw-liRKBg,46736
|
|
74
|
-
llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwHww2YU90lOonBPTmZIk,291
|
|
75
|
-
llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
|
|
76
72
|
llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
77
73
|
llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
|
|
78
74
|
llama_cloud/resources/users/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
79
75
|
llama_cloud/resources/users/client.py,sha256=A2s8e2syQHkkSwPz-Lrt_Zxp1K-8nqJqj5EafE6NWYs,5545
|
|
80
|
-
llama_cloud/types/__init__.py,sha256=
|
|
76
|
+
llama_cloud/types/__init__.py,sha256=RRUY02ATn2w51dyjZUFJ4zc8aDb-XmSm40I_cnqGEck,32169
|
|
81
77
|
llama_cloud/types/advanced_mode_transform_config.py,sha256=4xCXye0_cPmVS1F8aNTx81sIaEPjQH9kiCCAIoqUzlI,1502
|
|
82
78
|
llama_cloud/types/advanced_mode_transform_config_chunking_config.py,sha256=wYbJnWLpeQDfhmDZz-wJfYzD1iGT5Jcxb9ga3mzUuvk,1983
|
|
83
79
|
llama_cloud/types/advanced_mode_transform_config_segmentation_config.py,sha256=anNGq0F5-IlbIW3kpC8OilzLJnUq5tdIcWHnRnmlYsg,1303
|
|
@@ -85,7 +81,6 @@ llama_cloud/types/agent_data.py,sha256=Onaoc1QeIn3Il-8r1vgEzqvef92gHclCO7AC4kucE
|
|
|
85
81
|
llama_cloud/types/agent_deployment_list.py,sha256=7PWm2GHumo8CfqKU8fDRTJVDV4QQh8My1dhvBPO2zaA,1120
|
|
86
82
|
llama_cloud/types/agent_deployment_summary.py,sha256=9IKjSwu5uNpGfVyilr2W5d0aMApOdWNbtTk5w8GegEY,1642
|
|
87
83
|
llama_cloud/types/aggregate_group.py,sha256=LybxFl_1snA9VgG6f7sogwO7kYAwH_I88pkYc0oMOH0,1164
|
|
88
|
-
llama_cloud/types/audio_block.py,sha256=9JIGjZ8GU3C7ICv6XdNVN6_gWXyF18TJPaDuM9OUoMU,1071
|
|
89
84
|
llama_cloud/types/auto_transform_config.py,sha256=HVeHZM75DMRznScqLTfrMwcZwIdyWPuaEYbPewnHqwc,1168
|
|
90
85
|
llama_cloud/types/azure_open_ai_embedding.py,sha256=MeDqZoPYFN7Nv_imY9cfqDU9SPlEyAY4HcQZ4PF5X3g,2264
|
|
91
86
|
llama_cloud/types/azure_open_ai_embedding_config.py,sha256=o1zZhzcGElH3SeixFErrm7P_WFHQ6LvrLem_nKJWunw,1170
|
|
@@ -105,6 +100,7 @@ llama_cloud/types/character_chunking_config.py,sha256=2ooAnrlVVbKj4nDi_lR66x5E6n
|
|
|
105
100
|
llama_cloud/types/chat_app.py,sha256=fLuzYkXLq51C_Y23hoLwfmG-OiT7jlyHt2JGe6-f1IA,1795
|
|
106
101
|
llama_cloud/types/chat_app_response.py,sha256=WSKr1KI9_pGTSstr3I53kZ8qb3y87Q4ulh8fR0C7sSU,1784
|
|
107
102
|
llama_cloud/types/chat_data.py,sha256=ZYqVtjXF6qPGajU4IWZu3InpU54TXJwBFiqxBepylP0,1197
|
|
103
|
+
llama_cloud/types/chat_message.py,sha256=94GmO2Kp6VXhb7ZXhBPz8ExeZX1b_2LZLowuWtj6Tuw,1580
|
|
108
104
|
llama_cloud/types/chunk_mode.py,sha256=J4vqAQfQG6PWsIv1Fe_99nVsAfDbv_P81_KVsJ9AkU4,790
|
|
109
105
|
llama_cloud/types/classification_result.py,sha256=1faExxbtJLoYjy0h0Gl38Shk2idySEOenJBjQlcRpXs,1309
|
|
110
106
|
llama_cloud/types/classifier_rule.py,sha256=-64iBABkQ_IXN8rA77xA6L4xSsj8epTVT9Z1C7ypGx0,1533
|
|
@@ -154,11 +150,8 @@ llama_cloud/types/data_source_reader_version_metadata.py,sha256=hh7Hunen9GHlvtLb
|
|
|
154
150
|
llama_cloud/types/data_source_reader_version_metadata_reader_version.py,sha256=qZtQtHEnpWE48CjBPdljoYSzuk2rdrw5CCpWbLtM6Ps,735
|
|
155
151
|
llama_cloud/types/data_source_update_dispatcher_config.py,sha256=Sh6HhXfEV2Z6PYhkYQucs2MxyKVpL3UPV-I4cbf--bA,1242
|
|
156
152
|
llama_cloud/types/delete_params.py,sha256=1snPrd3WO9C1bKf0WdMslE2HQMF0yYLI3U7N53cmurM,1285
|
|
157
|
-
llama_cloud/types/document_block.py,sha256=OYKd5M3LgJ0Cz0K0YNuVRoHz9HcUdVuf2Vcqku8fck4,1116
|
|
158
153
|
llama_cloud/types/document_chunk_mode.py,sha256=6qH43Q0lIob2DMU95GsmSEOs4kQxOIyUFXj_kRDnyV4,470
|
|
159
154
|
llama_cloud/types/document_ingestion_job_params.py,sha256=33xTAl-K-m1j_Ufkj7w2GaYg9EUH5Hwsjn869X-fWMk,1524
|
|
160
|
-
llama_cloud/types/edit_suggestion.py,sha256=uzXSZYJiU3FaNN-TvEd3EXdaXvjQIe7Mf4kntKkyB2I,1202
|
|
161
|
-
llama_cloud/types/edit_suggestion_blocks_item.py,sha256=ojTk4lh0IHmrWP5wLPTIlsc2jAUDoHvdjJ5sm2uMut0,236
|
|
162
155
|
llama_cloud/types/element_segmentation_config.py,sha256=QOBk8YFrgK0I2m3caqV5bpYaGXbk0fMSjZ4hUPZXZDI,959
|
|
163
156
|
llama_cloud/types/embedding_model_config.py,sha256=6-o0vsAX89eHQdCAG5sI317Aivr4Tvs6ycg9TqNgybo,1525
|
|
164
157
|
llama_cloud/types/embedding_model_config_embedding_config.py,sha256=9rmfeiJYhBPmSJCXp-qxkOAd9WPwL5Hks7jIKd8XCPM,2901
|
|
@@ -167,14 +160,15 @@ llama_cloud/types/embedding_model_config_update_embedding_config.py,sha256=mrXFx
|
|
|
167
160
|
llama_cloud/types/eval_execution_params.py,sha256=ntVaJh5SMZMPL4QLUiihVjUlg2SKbrezvbMKGlrF66Q,1369
|
|
168
161
|
llama_cloud/types/extract_agent.py,sha256=Vj6tg8aEjUPADsUlkhHSCotrfWt8uoktaV45J81KeLc,1869
|
|
169
162
|
llama_cloud/types/extract_agent_data_schema_value.py,sha256=UaDQ2KjajLDccW7F4NKdfpefeTJrr1hl0c95WRETYkM,201
|
|
170
|
-
llama_cloud/types/extract_config.py,sha256=
|
|
163
|
+
llama_cloud/types/extract_config.py,sha256=FgkMtDgzeFZPNPHUR0HB00EauAz9Zh4ef6uC3yPyRO8,2780
|
|
171
164
|
llama_cloud/types/extract_config_priority.py,sha256=btl5lxl25Ve6_lTbQzQyjOKle8XoY0r16lk3364c3uw,795
|
|
172
165
|
llama_cloud/types/extract_job.py,sha256=Yx4fDdCdylAji2LPTwqflVpz1o9slpj9tTLS93-1tzU,1431
|
|
173
|
-
llama_cloud/types/extract_job_create.py,sha256=
|
|
166
|
+
llama_cloud/types/extract_job_create.py,sha256=5CcKnYprImF0wEqUJDqi6flAIJ0rzOWxmrCvtl_b8WM,1802
|
|
174
167
|
llama_cloud/types/extract_job_create_data_schema_override.py,sha256=vuiJ2lGJjbXEnvFKzVnKyvgwhMXPg1Pb5GZne2DrB60,330
|
|
175
168
|
llama_cloud/types/extract_job_create_data_schema_override_zero_value.py,sha256=HHEYxOSQXXyBYOiUQg_qwfQtXFj-OtThMwbUDBIgZU0,223
|
|
169
|
+
llama_cloud/types/extract_job_create_priority.py,sha256=_Qdc-ScGUcsgb0pv9-Viq2JgEoDYUi0AKStlw2E4Rb4,810
|
|
176
170
|
llama_cloud/types/extract_mode.py,sha256=S7H-XcH1wvPbOPVdwG9kVnZaH1pMY-LNzAD6TjCm0mc,785
|
|
177
|
-
llama_cloud/types/extract_models.py,sha256=
|
|
171
|
+
llama_cloud/types/extract_models.py,sha256=w_B-TzIkRvZhiBKA2bzyHXgQHs0tT4dEiIZsEnwiCgE,2072
|
|
178
172
|
llama_cloud/types/extract_resultset.py,sha256=Alje0YQJUiA_aKi0hQs7TAnhDmZuQ_yL9b6HCNYBFQg,1627
|
|
179
173
|
llama_cloud/types/extract_resultset_data.py,sha256=v9Ae4SxLsvYPE9crko4N16lBjsxuZpz1yrUOhnaM_VY,427
|
|
180
174
|
llama_cloud/types/extract_resultset_data_item_value.py,sha256=JwqgDIGW0irr8QWaSTIrl24FhGxTUDOXIbxoSdIjuxs,209
|
|
@@ -225,23 +219,21 @@ llama_cloud/types/http_validation_error.py,sha256=iOSKYv0dJGjyIq8DAeLVKNJY-GiM1b
|
|
|
225
219
|
llama_cloud/types/hugging_face_inference_api_embedding.py,sha256=c-O87QJZHaBWl0RobjD4tMsmtJCeUOc_oTl6oHZHDYU,1887
|
|
226
220
|
llama_cloud/types/hugging_face_inference_api_embedding_config.py,sha256=EFHhuPCxU0g3Jcc3k-8X-uH_OLCoRfWNbOCUpZ3Ygd4,1232
|
|
227
221
|
llama_cloud/types/hugging_face_inference_api_embedding_token.py,sha256=A7-_YryBcsP4G5uRyJ9acao3XwX5-YC3NRndTeDAPj4,144
|
|
228
|
-
llama_cloud/types/image_block.py,sha256=Bccrsm1-B2hUzObP7Oy1H7IVnurixfTpL03i-yqfZp0,1112
|
|
229
222
|
llama_cloud/types/ingestion_error_response.py,sha256=8u0cyT44dnpkNeUKemTvJMUqi_WyPcYQKP_DMTqaFPY,1259
|
|
230
223
|
llama_cloud/types/input_message.py,sha256=Ym6-tX6CMWKuHfxRtyM2y16kqSS3BzHged9rFRFkX0g,1346
|
|
231
224
|
llama_cloud/types/job_name_mapping.py,sha256=2dQFQlVHoeSlkyEKSEJv0M3PzJf7hMvkuABj3vMY7ys,1617
|
|
232
|
-
llama_cloud/types/job_names.py,sha256=
|
|
225
|
+
llama_cloud/types/job_names.py,sha256=CR7Bc8ViDuoF-Uk57ca2166hZTRedu5kIWQ2ZqB9t54,3647
|
|
233
226
|
llama_cloud/types/job_record.py,sha256=Z6sF9AruZJo-kTRgNufAWS3WK1yaEqop6kox1GpBYy4,2219
|
|
234
227
|
llama_cloud/types/job_record_parameters.py,sha256=Oqxp5y0owPfjLc_NR7AYE8P3zM2PJo36N9olbyNl7AA,3425
|
|
235
228
|
llama_cloud/types/job_record_with_usage_metrics.py,sha256=iNV2do5TB_0e3PoOz_DJyAaM6Cn9G8KG-dGPGgEs5SY,1198
|
|
236
229
|
llama_cloud/types/l_lama_parse_transform_config.py,sha256=YQRJZvKh1Ee2FUyW_N0nqYJoW599qBgH3JCH9SH6YLo,1249
|
|
237
230
|
llama_cloud/types/legacy_parse_job_config.py,sha256=eEPExbkUi9J7lQoY0Fuc2HK_RlhPmO30cMkfjtmmizs,12832
|
|
238
231
|
llama_cloud/types/license_info_response.py,sha256=fE9vcWO8k92SBqb_wOyBu_16C61s72utA-SifEi9iBc,1192
|
|
239
|
-
llama_cloud/types/
|
|
232
|
+
llama_cloud/types/llama_extract_feature_availability.py,sha256=oHJ3OyHf2rXmZhBSQfxVNnCFOp8IMKx_28EffCIEbLU,1228
|
|
233
|
+
llama_cloud/types/llama_extract_mode_availability.py,sha256=UtpYxpdZ29u3UarhGzH89H5rurvZQtOqO6a44gMm9DM,1379
|
|
240
234
|
llama_cloud/types/llama_extract_mode_availability_status.py,sha256=_ildgVCsBdqOLD__qdEjcYxqgKunXhJ_VHUeqjZJX8c,566
|
|
241
|
-
llama_cloud/types/llama_extract_settings.py,sha256=
|
|
242
|
-
llama_cloud/types/
|
|
243
|
-
llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py,sha256=-aL8fh-w2Xf4uQs_LHzb3q6LL_onLAcVzCR5yMI4qJw,1571
|
|
244
|
-
llama_cloud/types/llama_parse_parameters.py,sha256=5RxQ9pJ4kyVKwuJJbhWxnE0TwtMpH9i7AcuP3dMGIAw,6512
|
|
235
|
+
llama_cloud/types/llama_extract_settings.py,sha256=mWMjXL9t7d-J051Y3iSMgT-qa1h8VvCKrpFFvqv3FHM,2779
|
|
236
|
+
llama_cloud/types/llama_parse_parameters.py,sha256=kNpKnxuClDUYy8wO09DNKszJ_kbjx_T-s82GSdLBcNw,6552
|
|
245
237
|
llama_cloud/types/llama_parse_parameters_priority.py,sha256=EFRudtaID_s8rLKlfW8O8O9TDbpZdniIidK-xchhfRI,830
|
|
246
238
|
llama_cloud/types/llama_parse_supported_file_extensions.py,sha256=B_0N3f8Aq59W9FbsH50mGBUiyWTIXQjHFl739uAyaQw,11207
|
|
247
239
|
llama_cloud/types/llm_model_data.py,sha256=6rrycqGwlK3LZ2S-WtgmeomithdLhDCgwBBZQ5KLaso,1300
|
|
@@ -275,12 +267,15 @@ llama_cloud/types/paginated_extract_runs_response.py,sha256=NNeVcgBm0mYTAsumwQBO
|
|
|
275
267
|
llama_cloud/types/paginated_jobs_history_with_metrics.py,sha256=Bxy6N0x0FARJhgwNKKPkNpXx8YLRHvth23G14f5Fuk4,1136
|
|
276
268
|
llama_cloud/types/paginated_list_cloud_documents_response.py,sha256=MsjS0SWlT0syELDck4x2sxxR3_NC1e6QTdepgVmK9aY,1341
|
|
277
269
|
llama_cloud/types/paginated_list_pipeline_files_response.py,sha256=2TKR2oHSQRyLMqWz1qQBSIvz-ZJb8U_94367lwOJ2S4,1317
|
|
278
|
-
llama_cloud/types/paginated_report_response.py,sha256=o79QhQi9r0HZZrhvRlA6WGjxtyPuxN0xONhwXSwxtcs,1104
|
|
279
270
|
llama_cloud/types/paginated_response_agent_data.py,sha256=u6Y-Cq9qjGF5tskMOQChUNqyI91Tk-uQ6vQdi69cs80,1159
|
|
280
271
|
llama_cloud/types/paginated_response_aggregate_group.py,sha256=1ajZLZJLU6-GuQ_PPsEVRFZ6bm9he807F_F_DmB2HlQ,1179
|
|
281
272
|
llama_cloud/types/paginated_response_classify_job.py,sha256=ABpHn-ryRS8erj02ncxshAFe2Enw5JvSZqqbZuy0nWA,1167
|
|
282
273
|
llama_cloud/types/paginated_response_quota_configuration.py,sha256=S-miK621O7V6hBB05xcFBKCwa-gBK17iTHh29Saebz8,1123
|
|
283
|
-
llama_cloud/types/
|
|
274
|
+
llama_cloud/types/parse_configuration.py,sha256=mXjChoWseMnj-OEUUwkV-B5bUjPZ0SGfHr8lX4dAlRY,1762
|
|
275
|
+
llama_cloud/types/parse_configuration_create.py,sha256=3tnlIgHH_UgFFYP2OdVKyfIpa9mAzIzQxN4hDeazf3M,1467
|
|
276
|
+
llama_cloud/types/parse_configuration_filter.py,sha256=pObpFpnMq9CXfzZteY0S-2Lmj55mIdpQU4fZrEvgiZE,1260
|
|
277
|
+
llama_cloud/types/parse_configuration_query_response.py,sha256=h5L_E4NxPll8Nt47_PvfF93SMAuOlWl9q4_J9JZq6BM,1271
|
|
278
|
+
llama_cloud/types/parse_job_config.py,sha256=ISGxvIIujO-51ksx_lfVJLxze-Bq_yaC8uh8KnEt2GQ,7015
|
|
284
279
|
llama_cloud/types/parse_job_config_priority.py,sha256=__-gVv1GzktVCYZVyl6zeDt0pAZwYl-mxM0xkIHPEro,800
|
|
285
280
|
llama_cloud/types/parse_plan_level.py,sha256=GBkDS19qfHseBa17EXfuTPNT4GNv5alyPrWEvWji3GY,528
|
|
286
281
|
llama_cloud/types/parser_languages.py,sha256=Ps3IlaSt6tyxEI657N3-vZL96r2puk8wsf31cWnO-SI,10840
|
|
@@ -296,9 +291,9 @@ llama_cloud/types/permission.py,sha256=LjhZdo0oLvk7ZVIF1d6Qja--AKH5Ri0naUhuJvZS6
|
|
|
296
291
|
llama_cloud/types/pg_vector_distance_method.py,sha256=U81o0ARjPR-HuFcVspHiJUrjIDJo3jLhB46vkITDu7M,1203
|
|
297
292
|
llama_cloud/types/pg_vector_hnsw_settings.py,sha256=-RE59xUgHwNEyAwRYmOQ8SHeAqkSYBfCAROw7QomxUU,1758
|
|
298
293
|
llama_cloud/types/pg_vector_vector_type.py,sha256=VwOohN566zw42UMlnuKTJopYJypsSnzWjCFmKRoU-bo,952
|
|
299
|
-
llama_cloud/types/pipeline.py,sha256=
|
|
294
|
+
llama_cloud/types/pipeline.py,sha256=p2jZnDDDmBpkawjIYltnlKPlawLAJdKisEp0Bqqr_4s,2962
|
|
300
295
|
llama_cloud/types/pipeline_configuration_hashes.py,sha256=7_MbOcPWV6iyMflJeXoo9vLzD04E5WM7YxYp4ls0jQs,1169
|
|
301
|
-
llama_cloud/types/pipeline_create.py,sha256=
|
|
296
|
+
llama_cloud/types/pipeline_create.py,sha256=dNreffP5ia2k1965vL77YuBSUhPMARFtlCyJ3eWe40Q,2607
|
|
302
297
|
llama_cloud/types/pipeline_create_embedding_config.py,sha256=PQqmVBFUyZXYKKBmVQF2zPsGp1L6rje6g3RtXEcdfc8,2811
|
|
303
298
|
llama_cloud/types/pipeline_create_transform_config.py,sha256=HP6tzLsw_pomK1Ye2PYCS_XDZK_TMgg22mz17_zYKFg,303
|
|
304
299
|
llama_cloud/types/pipeline_data_source.py,sha256=iKB2NgpWQTl_rNDCvnXjNyd0gzohqwfCnupzWYT_CTE,2465
|
|
@@ -324,17 +319,16 @@ llama_cloud/types/pipeline_status.py,sha256=aC340nhfuPSrFVZOH_DhgYHWe985J3WNHrwv
|
|
|
324
319
|
llama_cloud/types/pipeline_transform_config.py,sha256=zMr-ePLKGjbaScxbAHaSwYBL7rrNibVlnn0cbgElDfU,824
|
|
325
320
|
llama_cloud/types/pipeline_type.py,sha256=tTqrhxHP5xd7W2dQGD0e5FOv886nwJssyaVlXpWrtRo,551
|
|
326
321
|
llama_cloud/types/plan_limits.py,sha256=WAbDbRl8gsQxvhmuVB0YT8mry-0uKg6c66uivyppdQU,2056
|
|
327
|
-
llama_cloud/types/playground_session.py,sha256=
|
|
322
|
+
llama_cloud/types/playground_session.py,sha256=vDebFzSvw0TyhOFMa5VFY8S7rZvedl6GNVmOFJd5kZo,1816
|
|
328
323
|
llama_cloud/types/pooling.py,sha256=5Fr6c8rx9SDWwWzEvD78suob2d79ktodUtLUAUHMbP8,651
|
|
329
324
|
llama_cloud/types/preset_composite_retrieval_params.py,sha256=yEf1pk4Wz5J6SxgB8elklwuyVDCRSZqfWC6x3hJUS4Q,1366
|
|
330
325
|
llama_cloud/types/preset_retrieval_params.py,sha256=TcyljefpspJSveMR9L5DQHlqW4jZeexBsXus_LkHkJA,2365
|
|
331
326
|
llama_cloud/types/preset_retrieval_params_search_filters_inference_schema_value.py,sha256=BOp-oJMIc3KVU89mmKIhVcwwsO0XBRnuErfsPqpUjSs,234
|
|
332
327
|
llama_cloud/types/presigned_url.py,sha256=-DOQo7XKvUsl-9Gz7fX6VOHdQLzGH2XRau24ASvG92E,1275
|
|
333
|
-
llama_cloud/types/progress_event.py,sha256=Bk73A8geTVaq0ze5pMnbkAmx7FSOHQIixYCpCas_dcY,1684
|
|
334
|
-
llama_cloud/types/progress_event_status.py,sha256=yb4RAXwOKU6Bi7iyYy-3lwhF6_mLz0ZFyGjxIdaByoE,893
|
|
335
328
|
llama_cloud/types/project.py,sha256=4NNh_ZAjEkoWl5st6b1jsPVf_SYKtUTB6rS1701G4IQ,1441
|
|
336
329
|
llama_cloud/types/project_create.py,sha256=GxGmsXGJM-cHrvPFLktEkj9JtNsSdFae7-HPZFB4er0,1014
|
|
337
330
|
llama_cloud/types/prompt_conf.py,sha256=hh8I3jxk3K6e5QZoBCLqszohMYtk73PERYoL36lLmTk,1660
|
|
331
|
+
llama_cloud/types/public_model_name.py,sha256=sKnedLmz7dS6U1PyT7qSjvoUuFpB2Q9HkXYjh16-EVw,4405
|
|
338
332
|
llama_cloud/types/quota_configuration.py,sha256=gTt2pLHhh9PpWxs1gtH1sTxM3Z6BBOAgSBI8AHCRoFI,2178
|
|
339
333
|
llama_cloud/types/quota_configuration_configuration_type.py,sha256=tg7owI77nZSHaMhTYMiXO5V-_bwjlK0Ao3TP7s0TNRI,1645
|
|
340
334
|
llama_cloud/types/quota_configuration_status.py,sha256=Lcmu1Ek9GAcj7LP8ciMzHrDcXvQ6eEFXEXOzG8v_HmE,580
|
|
@@ -345,37 +339,24 @@ llama_cloud/types/re_ranker_type.py,sha256=qYItMEHrf80ePBp7gNGBSL67mkTIsqco92WJa
|
|
|
345
339
|
llama_cloud/types/recurring_credit_grant.py,sha256=19qI3p5k1mQ1Qoo-gCQU02Aa42XpEsmwxPF1F88F-Yg,1517
|
|
346
340
|
llama_cloud/types/related_node_info.py,sha256=frQg_RqrSBc62ooJ4QOF5QRKymHcNot5WVFAB_g1sMg,1216
|
|
347
341
|
llama_cloud/types/related_node_info_node_type.py,sha256=lH95d8G-EnKCllV_igJsBfYt49y162PoNxWtrCo_Kgk,173
|
|
348
|
-
llama_cloud/types/report.py,sha256=9M_WkIxi5ilmtXrLKo5XxWzJ_qV8FFf5j8bAlQRmaks,1155
|
|
349
|
-
llama_cloud/types/report_block.py,sha256=y5n5z0JxZNH9kzN0rTqIdZPRLA9XHdYvQlHTcPSraKk,1381
|
|
350
|
-
llama_cloud/types/report_block_dependency.py,sha256=TGtLpcJG2xwTKr3GU8Err53T0BR_zNTiT-2JILvPbSg,785
|
|
351
|
-
llama_cloud/types/report_create_response.py,sha256=tmnVkyAMVf0HNQy186DFVV1oZQzYGY9wxNk84cwQLKA,1020
|
|
352
|
-
llama_cloud/types/report_event_item.py,sha256=_-0wgI96Ama2qKqUODTmI_fEcrnW5eAAjL1AoFEr4cQ,1451
|
|
353
|
-
llama_cloud/types/report_event_item_event_data.py,sha256=_v_2wZVGuNgXpitYNcKlA9hJVMLECOKf8A-pUuLron8,1171
|
|
354
|
-
llama_cloud/types/report_event_type.py,sha256=cPqKDVI8STX5BLndiGEovV4baa2it5fbfvcbiKyxAY8,1230
|
|
355
|
-
llama_cloud/types/report_metadata.py,sha256=cKB8wfToixuy8QEBNKzVTBznES9x4PU42DGnyiym5lc,1551
|
|
356
|
-
llama_cloud/types/report_plan.py,sha256=UvtYQaSNUTWbmC-rP0c57rbGpDRPUQgou0c2r96FVUo,1332
|
|
357
|
-
llama_cloud/types/report_plan_block.py,sha256=YlZ4fp4J3rduNKUknm0LfpHES_pgtQGFA9ZzErHoR40,1320
|
|
358
|
-
llama_cloud/types/report_query.py,sha256=IwZNM37fgwD2CrHkQ3LtdKwUCyL2r4SrZc0xwfaTa_I,1216
|
|
359
|
-
llama_cloud/types/report_response.py,sha256=20jVA79m3DBNHp3W6zbmD_yq9-64pNh4lk427bfCnqI,1252
|
|
360
|
-
llama_cloud/types/report_state.py,sha256=gjexexoT8GaCamGKvfwivKrfRtvdhEtwSLkAt-j9EMw,1127
|
|
361
|
-
llama_cloud/types/report_state_event.py,sha256=_wf-Cl_skJdrag-7h11tz-HIy1jed_GIG3c-ksuAjT4,1270
|
|
362
|
-
llama_cloud/types/report_update_event.py,sha256=uLRC79U3pvZ5-kY6pOseQyX1MNH-0m80GUtzpjd6mkI,1270
|
|
363
342
|
llama_cloud/types/retrieval_mode.py,sha256=wV9q3OdHTuyDWbJCGdxq9Hw6U95WFlJcaMq6KWSTzyw,910
|
|
364
343
|
llama_cloud/types/retrieve_results.py,sha256=rHArmu05K3NvIQepHX5nsVOfcMsZj3MaIcPkTC6mD_8,2375
|
|
365
344
|
llama_cloud/types/retriever.py,sha256=ZItPsorL8x1XjtJT49ZodaMqU8h2GfwlB4U4cgnfZkM,1626
|
|
366
345
|
llama_cloud/types/retriever_create.py,sha256=WyUR9DRzu3Q9tzKEeXCdQuzCY6WKi9ADJkZea9rqvxU,1286
|
|
367
346
|
llama_cloud/types/retriever_pipeline.py,sha256=F1pZDxg8JdQXRHE6ciFezd7a-Wv5bHplPcGDED-J4b0,1330
|
|
368
347
|
llama_cloud/types/role.py,sha256=4pbyLVNPleDd624cDcOhu9y1WvqC0J0gmNirTOW97iA,1342
|
|
348
|
+
llama_cloud/types/schema_generation_availability.py,sha256=42x9DCjLVRH27ZQC8bB4Atxd2rKoHoX2EZTT5S3LIlU,1111
|
|
349
|
+
llama_cloud/types/schema_generation_availability_status.py,sha256=bRU9bKidO01Zh3qZLH7tTJQSMImeqOlFDzF30Rhff7o,566
|
|
369
350
|
llama_cloud/types/schema_relax_mode.py,sha256=v4or6dYTvWvBBNtEd2ZSaUAb1706I0Zuh-Xztm-zx_0,635
|
|
370
351
|
llama_cloud/types/semantic_chunking_config.py,sha256=dFDniTVWpRc7UcmVFvljUoyL5Ztd-l-YrHII7U-yM-k,1053
|
|
371
352
|
llama_cloud/types/sentence_chunking_config.py,sha256=NA9xidK5ICxJPkEMQZWNcsV0Hw9Co_bzRWeYe4uSh9I,1116
|
|
372
|
-
llama_cloud/types/
|
|
353
|
+
llama_cloud/types/sparse_model_config.py,sha256=vwt0_3ncjFCtNyWsMSYRrVuoTAWsdnQCHSTUM4HK-Lc,1529
|
|
354
|
+
llama_cloud/types/sparse_model_type.py,sha256=vmjOS3tSqopsvxWqw3keeIL4kgskJv6TJL-Gw_qQQ5s,933
|
|
373
355
|
llama_cloud/types/status_enum.py,sha256=cUBIlys89E8PUzmVqqawu7qTDF0aRqBwiijOmRDPvx0,1018
|
|
374
356
|
llama_cloud/types/struct_mode.py,sha256=ROicwjXfFmgVU8_xSVxJlnFUzRNKG5VIEF1wYg9uOPU,1020
|
|
375
357
|
llama_cloud/types/struct_parse_conf.py,sha256=3QQBy8VP9JB16d4fTGK_GiU6PUALIOWCN9GYI3in6ic,2439
|
|
376
358
|
llama_cloud/types/supported_llm_model.py,sha256=hubSopFICVNEegbJbtbpK6zRHwFPwUNtrw_NAw_3bfg,1380
|
|
377
359
|
llama_cloud/types/supported_llm_model_names.py,sha256=w2FrfffSwpJflq1EoO6Kw7ViTOZNGX4hf60k0Qf3VLA,3213
|
|
378
|
-
llama_cloud/types/text_block.py,sha256=X154sQkSyposXuRcEWNp_tWcDQ-AI6q_-MfJUN5exP8,958
|
|
379
360
|
llama_cloud/types/text_node.py,sha256=Tq3QmuKC5cIHvC9wAtvhsXl1g2sACs2yJwQ0Uko8GSU,2846
|
|
380
361
|
llama_cloud/types/text_node_relationships_value.py,sha256=qmXURTk1Xg7ZDzRSSV1uDEel0AXRLohND5ioezibHY0,217
|
|
381
362
|
llama_cloud/types/text_node_with_score.py,sha256=k-KYWO_mgJBvO6xUfOD5W6v1Ku9E586_HsvDoQbLfuQ,1229
|
|
@@ -396,9 +377,9 @@ llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPX
|
|
|
396
377
|
llama_cloud/types/vertex_ai_embedding_config.py,sha256=DvQk2xMJFmo54MEXTzoM4KSADyhGm_ygmFyx6wIcQdw,1159
|
|
397
378
|
llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2BaIMltDqGnIowU,1217
|
|
398
379
|
llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
|
|
399
|
-
llama_cloud/types/webhook_configuration.py,sha256=
|
|
380
|
+
llama_cloud/types/webhook_configuration.py,sha256=E0QIuApBLlFGgdsy5VjGIkodclJvAxSO8y8n3DsGHrg,1398
|
|
400
381
|
llama_cloud/types/webhook_configuration_webhook_events_item.py,sha256=OL3moFO_6hsKZYSBQBsSHmWA0NgLcLJgBPZfABwT60c,2544
|
|
401
|
-
llama_cloud-0.1.
|
|
402
|
-
llama_cloud-0.1.
|
|
403
|
-
llama_cloud-0.1.
|
|
404
|
-
llama_cloud-0.1.
|
|
382
|
+
llama_cloud-0.1.41.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
|
|
383
|
+
llama_cloud-0.1.41.dist-info/METADATA,sha256=pEx6MUg3UzTGrnSiK2S_g5Q06ftbdq4yVDkftJPsRrU,2706
|
|
384
|
+
llama_cloud-0.1.41.dist-info/WHEEL,sha256=b4K_helf-jlQoXBBETfwnf4B04YC67LOev0jo4fX5m8,88
|
|
385
|
+
llama_cloud-0.1.41.dist-info/RECORD,,
|